Cody Pendant has asked for the wisdom of the Perl Monks concerning the following question:
I've spent Far Too Long trying to do it myself, and once again been reminded that we have modules like HTML::Linkextor for a reason.
But now I'm confused about the different Robot and Spider and UA modules.
My task ought to be simple for a Module, but can someone please get me started with a little help?
My pseudocode is this:
give the spider a URL, say www.whereiwork.com/site/ recursively { find all linked pages, but only within that directory } for each page found { print out their titles and any meta-tags we find } report any errors following the links
Thanks in advance.
--
($_='jjjuuusssttt annootthheer pppeeerrrlll haaaccckkeer')=~y/a-z//s;print;
|
|---|
| Replies are listed 'Best First'. | |
|---|---|
|
Re: A Spider Tool
by Aristotle (Chancellor) on Aug 18, 2002 at 00:39 UTC | |
|
•Re: A Spider Tool
by merlyn (Sage) on Aug 18, 2002 at 09:53 UTC | |
by Cody Pendant (Prior) on Aug 19, 2002 at 04:46 UTC | |
|
Re: A Spider Tool
by neilwatson (Priest) on Aug 18, 2002 at 03:23 UTC | |
by Cody Pendant (Prior) on Aug 18, 2002 at 03:46 UTC |