Matt Garman wrote:
> I'm trying to use Python to work with large pipe ('|') delimited data
> files.  The files range in size from 25 MB to 200 MB.
> 
> Since each line corresponds to a record, what I'm trying to do is
> create an object from each record.  However, it seems that doing this
> causes the memory overhead to go up two or three times.

    Why do you want all the records in memory at once?  Are you
doing some lookup on them, or what?  If you're processing files
sequentially, don't keep them all in memory.

    You're getting into the size range where it may be time to
use a database.

                                John Nagle
-- 
http://mail.python.org/mailman/listinfo/python-list

Reply via email to