in reply to Re: Processing XML with MediaWiki::DumpFile
in thread Processing XML with MediaWiki::DumpFile
There's definitely too much data to build a tree in memory-- the "small" version of the wiki dump (only the latest version of each article) that I'm starting from is about 30 GB, and my reduced data file is still about 380 MB.
I'll poke around CPAN for a streaming XML generator.
|
|---|
| Replies are listed 'Best First'. | |
|---|---|
|
Re^3: Processing XML with MediaWiki::DumpFile
by Anonymous Monk on Feb 12, 2012 at 16:29 UTC |