On Fri, Dec 11, 2009 at 5:12 AM, Wolodja Wentland <wentl...@cl.uni-heidelberg.de> wrote: > Hi all, > > I am writing a library for accessing Wikipedia data and include a module > that generates graphs from the Link structure between articles and other > pages (like categories). > > These graphs could easily contain some million nodes which are frequently > linked. The graphs I am building right now have around 300.000 nodes > with an average in/out degree of - say - 4 and already need around 1-2GB of > memory. I use networkx to model the graphs and serialise them to files on > the disk. (using adjacency list format, pickle and/or graphml).
Huh. Using graphine- which should be somewhat more memory hungry than networkx- I generated a naive million node 4-cycle graph and wound up using something under 600 meg of ram. Can you post some code? Geremy Condra -- http://mail.python.org/mailman/listinfo/python-list