There's definitely too much data to build a tree in memory-- the "small" version of the wiki dump (only the latest version of each article) that I'm starting from is about 30 GB, and my reduced data file is still about 380 MB.
I'll poke around CPAN for a streaming XML generator.
In reply to Re^2: Processing XML with MediaWiki::DumpFile
by bitingduck
in thread Processing XML with MediaWiki::DumpFile
by bitingduck
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |