On Sun, 2008-04-20 at 13:49 -0400, Richard Lee wrote: > can this be optimized in anyway? > open (my $source, '-|', "tail -100000 /server/server.log") > > is this the best way to get large portion(well file itself is over 20 > times) of the file into find handle? >
This will not optimize processing the file. The tail command would have to run through the entire file and save the last 100_000 lines. It might save on RAM memory, but with today's multi-gigabyte machines, the file would have to be gigabyte sized to overflow it. 2_000_000 lines at 80 characters each is 160_000_000 bytes (well under 1GB). There may be two problems with this: 1. If you are running a memory hog like Vista, you may run out of memory reading a measly 160 MB file. 2. If you are running on a server, it may be so busy that you run out of memory. All I can say is TITS (Try It To See). -- Just my 0.00000002 million dollars worth, Shawn When I was in college, my Karate instructor said, "The first hundred are easy." Since then, I discovered he'd lied. Life's too big to describe in hundreds. -- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED] http://learn.perl.org/