Perl: the Markov chain saw | |
PerlMonks |
Re^2: Mass downloads.by BrowserUk (Patriarch) |
on Jun 06, 2005 at 17:04 UTC ( [id://463993]=note: print w/replies, xml ) | Need Help?? |
I abondoned the idea. To do the indexing I envisioned justice, I would have had to download the greater majority of PM's nodes. At the mandated rate of 1 every 5 seconds+, it would require 500 hours. Split that into 2 hour chunks of connect time and it becomes untenable. Hence I've never bothered to extend the scripts beyond their simplist form: PMDown.pl takes a filename containing a list of PM nodeid's to download: BewareEven with 1 thread running, this will far exceed the download rate approved.
ExtractWords.pl
IndexDocs.pl
Note: G.pm is Jenda's module that does wildcard ARGV expansion. The result of processing is a file that looks like this:
But I manually filtered the intermediate words list. Examine what is said, not who speaks -- Silence betokens consent -- Love the truth but pardon error.
Lingua non convalesco, consenesco et abolesco. -- Rule 1 has a caveat! -- Who broke the cabal?
"Science is about questioning the status quo. Questioning authority".
The "good enough" maybe good enough for the now, and perfection maybe unobtainable, but that should not preclude us from striving for perfection, when time, circumstance or desire allow.
In Section
Perl Monks Discussion
|
|