On Mon, May 09, 2005 at 10:31:08AM +0100, David Hugh-Jones wrote: > > On 09/05/05, Richard Lyons <[EMAIL PROTECTED]> wrote: > > On Sun, May 08, 2005 at 09:48:07AM +0200, Nacho wrote: > > > > On Mon, May 02, 2005 at 01:27:41PM +0100, Richard Lyons wrote: > > > > > I am considering how to crawl a site which is dynamically generated, > > > > > and create a static version of all generated pages (or selected [...] > > > > > > Well, I don't know an "elegant" solution... one dirty approach would be to > > > first download the site with "wget -r", then you would get lots of files > > > with [...] > > > > Yes... that is the kind of thing I was imagining. It will probably be > > quite simple once I get started. But first I need to find time :-( [...] > If you end up wanting to do something more complicated, you could look > into WWW::Mechanize: > > http://search.cpan.org/perldoc?WWW%3A%3AMechanize > That looks _very_ interesting --and potent. Thanks for that too.
-- richard -- To UNSUBSCRIBE, email to [EMAIL PROTECTED] with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]