Just as a little background, I am working on a BioInformatics program that runs on large (about 300 meg) text files. I am using a filehandle to open and load it into an array. Then I use the join command to read the array into a scalar variable in order to be in a workable form for my computationally intensive program. The problem, however, is that the machine I am working with only has 256 megs of RAM. I have split the files into 50-60 meg chunks, but my program still uses all the available physical memory. I am quite new to perl and do not know any other methods of working with the data other than to make it a scalar variable, which requires loading it into memory. Does anyone have any solution to my memory woes? I would greatly appreciate your help. Thanks a lot.
-- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]