Ross Reyes wrote: > When I use readlines, what happens if the number of lines is huge? I have > a very big file (4GB) I want to read in, but I'm sure there must be some > limitation to readlines and I'd like to know how it is handled by python.
readlines itself has no limitation, but it reads all the lines into memory, so you'll probably run out of memory before the call returns. Python raises a MemoryError exception when this happens. > I am using it like this: > slines = infile.readlines() # reads all lines into a list of strings called > "slines" as others have pointed out, an iterator is the best way to solve this. the iterator code will read blocks of data from the file, and return the lines one by one. if you need to support older versions of Python, xreadlines or repeated calls to readlines(N) can be useful (see the third example on this page for an example: http://effbot.org/zone/readline-performance.htm ) </F> -- http://mail.python.org/mailman/listinfo/python-list