Which graph library is best suited for large graphs?
debatem1 at gmail.com
Sat Dec 12 03:48:34 CET 2009
On Fri, Dec 11, 2009 at 5:12 AM, Wolodja Wentland
<wentland at cl.uni-heidelberg.de> wrote:
> Hi all,
> I am writing a library for accessing Wikipedia data and include a module
> that generates graphs from the Link structure between articles and other
> pages (like categories).
> These graphs could easily contain some million nodes which are frequently
> linked. The graphs I am building right now have around 300.000 nodes
> with an average in/out degree of - say - 4 and already need around 1-2GB of
> memory. I use networkx to model the graphs and serialise them to files on
> the disk. (using adjacency list format, pickle and/or graphml).
Huh. Using graphine- which should be somewhat more memory hungry
than networkx- I generated a naive million node 4-cycle graph and wound
up using something under 600 meg of ram. Can you post some code?
More information about the Python-list