alas, I am stymied once again ... and have humbly come for assistance.
I am trying to pull the links off a page and store them in @links. There is standard code for this which I have used with success.
my @links = (); sub callback { my($tag, %attr) = @_; return if $tag ne 'a'; push(@links, values %attr); + } # Make the parser. $p = HTML::LinkExtor->new(\&callback); # Request document and parse it as it arrives $res = $ua->request(HTTP::Request->new(GET => $url), sub {$p->parse($_[0])});
Now, however, I am trying to get the links off a page that requires a username/password ... through the assistance of the monks I have accomplished a user/pass webpage grab...
$ua = LWP::UserAgent->new; $req = HTTP::Request->new(GET => $url); $req->authorization_basic('user', 'pass'); $res = $ua->request($req)->as_string,
Now the question is how to merge the user/pass webpage grab with the link extractor.
I have tried
$ua = LWP::UserAgent->new; $req = HTTP::Request->new(GET => $url); $req->authorization_basic('user', 'pass'); $res = $ua->request($req)->as_string, sub {$p->parse($_[0])};
but when I print out @links I get nothing. I think (but really have no clue) this has something to do with the ->as_string, but without it the webpage comes out as HTTP::Response=HASH(0x8435960).
Is there something else that I should be doing to get these links pulled out properly? Obviously there is, but do you guys know what that might be?
cdherold
In reply to Link Extraction when grabbing web page with USER/PASS by cdherold
| For: | Use: | ||
| & | & | ||
| < | < | ||
| > | > | ||
| [ | [ | ||
| ] | ] |