peterr has asked for the wisdom of the Perl Monks concerning the following question:
We are having problems with a site where session ID's are being set, despite the fact that the PHP code is meant to exclude the session ID from the link/url. This is of course causing some grief to the website owner, as we have seen both Yahoo and MSN with links to the website, and they contain the session ID. This can potentially cause big probems.
We have reviewed what needs to be done, and need to test the new PHP code, however do not want to use any of the "Search Engine Simulators" around, I would rather we place a Perl script on the website, and do our own isolated testing.
So, does anyone know of a good Perl script that simulates a 'spider crawl' please, just to show the links and related links, so that we can thouroughly test that session ID's are not appearing. If possible, the script would allow us to enter the 'user agent', because we only want to turn sessions off, for spiders/bots,etc, not the general public.
The type of 'results' I needed are the same as produced by these simulators:
http://www.1-hit.com/all-in-one/tool.search-engine-viewer.htm
http://www.webconfs.com/search-engine-spider-simulator.php
Thanks,
Peter
|
---|
Replies are listed 'Best First'. | |
---|---|
Re: Search Engine Simulator
by Corion (Patriarch) on Dec 10, 2004 at 08:06 UTC | |
by peterr (Scribe) on Dec 10, 2004 at 10:26 UTC | |
by Corion (Patriarch) on Dec 10, 2004 at 10:32 UTC | |
by peterr (Scribe) on Dec 10, 2004 at 10:49 UTC | |
Re: Search Engine Simulator
by BUU (Prior) on Dec 10, 2004 at 12:42 UTC | |
Re: Search Engine Simulator
by rupesh (Hermit) on Dec 10, 2004 at 04:44 UTC | |
Re: Search Engine Simulator
by talexb (Chancellor) on Dec 10, 2004 at 19:54 UTC |