> > Howdt list. > > > > I've never had to work with really big files before( I'm used > > to tiny ones you can slurp in at once) and now I have to > > process a 180MB text file line by line and was wondering the > > most efficient method to do so form within a script not via > > coommand line. > > > > Would I just do > > open(FH ...) or die $!; > > > > while(<FH>)) { process line } > > > > close(FH); > > > > Or is there a better way? > > > > TIA > > > > Dan > > > > -- > > To unsubscribe, e-mail: [EMAIL PROTECTED] > > For additional commands, e-mail: [EMAIL PROTECTED] > > > That's the general idea. There could be variations depending > on what you'retrying to accomplish in the processing of the > lines, or if infact you need to process line by line. 180mb > is not really a "Really big file", I process files in the > gigabytes range with excellent performace results. If your > considering somthing like: > > while (<FH>) { > chomp; > @fields = split /$delim/; > foreach $field (@fields) { > # lots of field level processing > } > } > > you might consider other methods. But even this type of > construct works very well on files in the 180mb range. Bear > in mind I work on a large HP-UX box, certainly if your system > is a 386sx16 your mileage may vary. > > HTH > Steve
Thanks for the input steve! Good to know. Dan -- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]