Hi, I have an application where I need to parse a lot of files and create a graph of users from them. Each node will have some attributes and will be connected to some other nodes with some edge weights which can change depending upon the type of analysis. The number of nodes is extremely large (i.e. more than 10 million). I tried using hashes to represent these graphs once created using Storable but when I try to load them, the system goes out of memory. I wanted to know what is the best possible way of representing graphs in such applications. Should they be represented as disk files i.e. each node stored as a separate file in which case I would need a good distributing algorithm for putting them into a hierarchy in filesystem. Or they can be represented on single files without killing too much memory.
Please enlighten.
Thanks