mkurtis has asked for the wisdom of the Perl Monks concerning the following question:

Why is it that when i run a perl script using
#!/usr/bin/perl -w use WWW::Mechanize; my $mech = WWW::Mechanize->new(); $mech->get("http://ebay.com/help/policies/group-guidelines.html"); print $mech->content(); print $mech->uri();
it allows me to get the content of the site, even though i went to the robots.txt file on ebay and checked to see what sites it disallowed to check its ability to obey rules. did i do something wrong so that it doesnt use LWP::UserAgent? Thanks

Replies are listed 'Best First'.
Re: WWW::mechanize not obeying rules
by Corion (Patriarch) on Mar 02, 2004 at 07:21 UTC
Re: WWW::mechanize not obeying rules
by esskar (Deacon) on Mar 01, 2004 at 23:54 UTC
    well, the robots.txt forbids to access help/policies/, so a search-engine-bot will not add those pages in its index.

    but your favorite web-browser is able to access that pages, right? that's why there is no reason that WWW::Mechanize (or similar stuff like LWP::Simple) can't poll those pages too.