morgon has asked for the wisdom of the Perl Monks concerning the following question:
for many years I have used WWW::Mechanize with excellent results to scrape a weekly magazin website for offline consumption on my Paĺm-pilot.
Now all of a sudden they have changed their table of content ...
Where there used to be nice html there is now an ugly mix of html-fragments with a lot of javascript mixed in that builds the html dynamically (a lot of document.write) and unfortunately completely breaks my conversion scripts...
So what I want now is a way to capture the html that the javascript generates - i.e. a tool that interprets the javascript and saves the resulting document-html in a file.
Any ideas on how to achieve this?
|
|---|
| Replies are listed 'Best First'. | |
|---|---|
|
Re: getting rid of javascript
by alexlc (Beadle) on May 02, 2009 at 04:51 UTC | |
|
Re: getting rid of javascript
by spx2 (Deacon) on May 02, 2009 at 07:04 UTC | |
by whakka (Hermit) on May 02, 2009 at 19:05 UTC | |
|
Re: getting rid of javascript
by whakka (Hermit) on May 02, 2009 at 16:56 UTC | |
|
Re: getting rid of javascript
by Anonymous Monk on May 02, 2009 at 06:28 UTC |