Matthew Weier O'Phinney helpfully answered my question about how one could 
download the pages in an asp application with:

 >   wget -m -L -t 5 -w 5 http://someplace.com/some.asp&page=1

This does indeed have the desired effect.  Unfortunately, it yields an 
embarrassment of riches - I seem to be getting all the pages on the site, which 
is quite large.  While I could just download the whole site, I'm only on a 56k 
modem.  I actually may not have enough disk space on my box to store all the 
files I might download.

Is there some way I can limit the URL's wget will follow?  Suppose all of my 
pages have a URL like:

http://someplace.com/some.asp&ThisIsFixed=1&page=1

I would like to only download URLs that have "ThisIsFixed=1" in them.

What would be really cool is if one could get wget to test a URL with a regular 
expression before downloading it, but I don't see a way to do that.  (I have 
been studying the man page.)

I could hack on wget's source code if necessary.  If I have to do that, maybe 
somebody could give me a tip on where to look in the source.  Maybe I could 
contribute a useful patch.

Thanks,

Mike
-- 
Michael D. Crawford
GoingWare Inc. - Expert Software Development and Consulting
http://www.goingware.com/
[EMAIL PROTECTED]

      Tilting at Windmills for a Better Tomorrow.


-- 
To UNSUBSCRIBE, email to [EMAIL PROTECTED] 
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]

Reply via email to