Thanks Perrin.
With multiple processes, I have tracked down two basic approaches, apart from LWP::Parallel which you didn't like.
- ThreadQueue -- Re: What is the fastest way to download a bunch of web pages?-- (thanks BrowserUK)
- Parallel::ForkManager (Suggested by jasonk above, and also mentioned on the "fastest way to download" thread)
Do you think one way has any advantages over the other? Or are these ways essentially the same under the hood?
FWIW I'm on linux now (new job -- yay! now I get perl in its native habitat :) ), since this seems to be relevant when forking comes into play. (Forking works better on linux.)
Also, to give a bit more contetx, I'll be downloading potentially 10s of thousands of websites, but no more than 100 from any one particular domain.
|