cdherold has asked for the wisdom of the Perl Monks concerning the following question:
alas, I am stymied once again ... and have humbly come for assistance.
I am trying to pull the links off a page and store them in @links. There is standard code for this which I have used with success.
my @links = (); sub callback { my($tag, %attr) = @_; return if $tag ne 'a'; push(@links, values %attr); + } # Make the parser. $p = HTML::LinkExtor->new(\&callback); # Request document and parse it as it arrives $res = $ua->request(HTTP::Request->new(GET => $url), sub {$p->parse($_[0])});
Now, however, I am trying to get the links off a page that requires a username/password ... through the assistance of the monks I have accomplished a user/pass webpage grab...
$ua = LWP::UserAgent->new; $req = HTTP::Request->new(GET => $url); $req->authorization_basic('user', 'pass'); $res = $ua->request($req)->as_string,
Now the question is how to merge the user/pass webpage grab with the link extractor.
I have tried
$ua = LWP::UserAgent->new; $req = HTTP::Request->new(GET => $url); $req->authorization_basic('user', 'pass'); $res = $ua->request($req)->as_string, sub {$p->parse($_[0])};
but when I print out @links I get nothing. I think (but really have no clue) this has something to do with the ->as_string, but without it the webpage comes out as HTTP::Response=HASH(0x8435960).
Is there something else that I should be doing to get these links pulled out properly? Obviously there is, but do you guys know what that might be?
cdherold
|
|---|
| Replies are listed 'Best First'. | |
|---|---|
|
Re: Link Extraction when grabbing web page with USER/PASS
by tachyon (Chancellor) on Mar 04, 2003 at 04:30 UTC | |
by cdherold (Monk) on Mar 04, 2003 at 05:54 UTC | |
by tachyon (Chancellor) on Mar 04, 2003 at 07:08 UTC | |
|
Re: Link Extraction when grabbing web page with USER/PASS
by zakb (Pilgrim) on Mar 04, 2003 at 09:08 UTC |