Price, Jason wrote: > I'm trying to optimize a script used for processing large text log > files (around 45MB). I think I've got all the processing fairly well > optimized, but I'm wondering if there's anything I can do to speed up > the initial loading of the file. > > Currently, I'm performing operations on the file one line at a time, > using a "while (<FILE>)" loop. I'm pushing important lines into an > array, so further processing is done in memory, but the initial pass > on the file is rather time consuming. Is there a more efficient way > to work with large text files?
Since the file is line-oriented, what you're doing looks like the best way to me. You can play around with the stdio buffering using IO::Handle::setbuf(), but I doubt that will help much. -- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED] <http://learn.perl.org/> <http://learn.perl.org/first-response>