From: "Dan Muey" <[EMAIL PROTECTED]> > All I hear are crickets on the list anybody there today?
Hey I aint no cricket ;-) > I'd like to have a simple spider that will look at a url's diretcory > and simply give me a list of files in that directory. > > IE > > my $files = ????? http://www.monkey.com/bannana/ > > And have $files be an array reference or something so I could then : > > for(@{$files}) { print "-$_-\n"; } > > Or somehtign like that. As Wiggins says it may be impossible. These days you too often see "Directory Listing Denied" :-( If you can list the directory you can 1) download the http://www.monkey.com/bannana/ page with LWP 2) extract the links from it with HTML::LinkExtor (use the base parameter when creating the object so that you get absolute URLs) 3) fetch all URLs that start with http://www.monkey.com/bannana/ it may be helpfull to allow the user of the script (yourself) to specify some filtering. Eg. that you want only .gif and .jpg. I made for myself a little different script. I give it a URL like http://www.monkey.com/bannana/pix001.jpg and it tries to download pix001.jpg, pix002.jpg, ... Some sites insist on proper HTTP_REFERER so I can set that as well. I don't think the script is worth posting here ;-) Jenda ===== [EMAIL PROTECTED] === http://Jenda.Krynicky.cz ===== When it comes to wine, women and song, wizards are allowed to get drunk and croon as much as they like. -- Terry Pratchett in Sourcery -- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]