Just as a little background, I am working on a BioInformatics program that
runs on large (about 300 meg) text files.  I am using a filehandle to open
and load it into an array.  Then I use the join command to read the array
into a scalar variable in order to be in a workable form for my
computationally intensive program.  The problem, however, is that the
machine I am working with only has 256 megs of RAM.  I have split the files
into 50-60 meg chunks, but my program still uses all the available physical
memory.  I am quite new to perl and do not know any other methods of working
with the data other than to make it a scalar variable, which requires
loading it into memory.  Does anyone have any solution to my memory woes?  I
would greatly appreciate your help.  Thanks a lot.



-- 
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to