The amount of data I read in is actually small. If you see my algorithm above it deals with 2000 nodes and each node has ot of attributes.
When I close the program my computer becomes stable and performs as usual. I check the performance in Performance monitor and using "top" and the total memory is being used and on top of that around half a gig swap memory is also being used. Please give some helpful pointers to overcome such memory errors. I revisited my code to find nothing so obvious which would let this leak happen. How to kill cross references in the program. I am kinda newbie and not completely aware of the finetuning such programming process. Thanks bruno at modulix wrote: > [EMAIL PROTECTED] wrote: > > I have a python code which is running on a huge data set. After > > starting the program the computer becomes unstable and gets very > > diffucult to even open konsole to kill that process. What I am assuming > > is that I am running out of memory. > > > > What should I do to make sure that my code runs fine without becoming > > unstable. How should I address the memory leak problem if any ? I have > > a gig of RAM. > > > > Every help is appreciated. > > Just a hint : if you're trying to load your whole "huge data set" in > memory, you're in for trouble whatever the language - for an example, > doing a 'buf = openedFile.read()' on a 100 gig file may not be a good > idea... > > > > -- > bruno desthuilliers > python -c "print '@'.join(['.'.join([w[::-1] for w in p.split('.')]) for > p in '[EMAIL PROTECTED]'.split('@')])" -- http://mail.python.org/mailman/listinfo/python-list