in reply to Re: Conserving bandwidth with WWW::Mechanize's get()
in thread Conserving bandwidth with WWW::Mechanize's get()

It's not exactly a spider, in that is only grabs one piece of info from a single page.

I had assumed that it downloaded the complete package by monitoring the volume use in an hour and dividing by (60*6), giving me a rough stab at the volume per page. I then saved the page to disc from firefox and the values were roughly the same. 60kB or so of .html file, and 90kB of images and other frippery.

Is this assumption misguided somehow?

  • Comment on Re^2: Conserving bandwidth with WWW::Mechanize's get()

Replies are listed 'Best First'.
Re^3: Conserving bandwidth with WWW::Mechanize's get()
by pc88mxer (Vicar) on Jun 05, 2008 at 01:52 UTC
    WWW::Mechanize will automatically handle redirects, but those should be short messages.

    Using LWP::Debug you can get a trace of all the traffic that Mechanize is generating.

    use WWW::Mechanize; use LWP::Debug; my $mech = WWW::Mechanize->new(); LWP::Debug::level("+"); $mech->get("http://www.cnn.com/"); print length($mech->content), "\n";