SergioQ has asked for the wisdom of the Perl Monks concerning the following question:
I have an earlier post about web scraping images using Perl and the WWW::Mechanize family for web pages that get built using JavaScript after they are loaded. Obviously a lot of hurdles with are beyond my capabilities.
So I have a new thought but not sure it's possible, I'll describe what I believe may be a kludgy solution, which I'll gladly take at this point.
Please remember, I have my own Ubuntu server sitting somewhere, and I will run this script from any laptop with a web browser.
•Have my Perl script generate an HTML form of my own design, and then output it to my browser
•Make sure my Perl script does not exit here
•Have my Perl script generate the proper URL for the page I want to scrape
•Have my Perl script redirect to the URL from the above line and open that in either a new browser window or tab
•Now have my Perl script end
Then my idea is that I would go to the redirected window, manually "View Source", copy it, and post it to the form generated by the above Perl script.
If this kludgy idea was an entry in a cheese contest, it would not just be tagged the winner, but it would be so smelly that all the judges would run out while trying not to puke. However, is this possible?
|
|---|