converting a sed / grep / awk / . . . bash pipe line into python
roy at panix.com
Wed Sep 3 15:41:27 CEST 2008
<7f2d4b4a-bc97-4b46-a31e-63f98e9fee73 at 34g2000hsh.googlegroups.com>,
bearophileHUGS at lycos.com wrote:
> Roy Smith:
> > No reason to limit how many splits get done if you're
> > explicitly going to slice the first two.
> You are probably right for this problem, because most lines are 2
> items long, but in scripts that have to process lines potentially
> composed of many parts, setting a max number of parts speeds up your
> script and reduces memory used, because you have less parts at the
Sounds like premature optimization to me. Make it work and be easy to
understand first. Then worry about how fast it is.
But, along those lines, I've often thought that split() needed a way to not
just limit the number of splits, but to also throw away the extra stuff.
Getting the first N fields of a string is something I've done often enough
that refactoring the slicing operation right into the split() code seems
worthwhile. And, it would be even faster :-)
More information about the Python-list