Okay, I'm still struggling here. My problem is I have a client who has a
rather large tab delimited text file I am attempting to slurp up and
place in a MySQL table. The file is almost 6 megs large but my ISP only
allows 2 megs of RAM per user. I can slurp up only about 1.5 megs before
I get an error message. I would like to read in only about 1.5 megs at a
time but if I use the following I exceed memory limits:
while (sysread (TEMP, $temp, 1_500_000))
{
# read into MySQL
}
Is there a way to step through a large file and process only what I read
in? I'm so stymied I wanna puke.
--
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]