On May 4, 12:33 am, "Gabriel Genellina" <[EMAIL PROTECTED]> wrote: > En Sun, 04 May 2008 01:33:45 -0300, Jetus <[EMAIL PROTECTED]> escribió: > > > Is there a good place to look to see where I can find some code that > > will help me to save webpage's links to the local drive, after I have > > used urllib2 to retrieve the page? > > Many times I have to view these pages when I do not have access to the > > internet. > > Don't reinvent the wheel and use wgethttp://en.wikipedia.org/wiki/Wget > > -- > Gabriel Genellina
A lot of the functionality is already present. import urllib urllib.urlretrieve( 'http://python.org/', 'main.htm' ) from htmllib import HTMLParser from formatter import NullFormatter parser= HTMLParser( NullFormatter( ) ) parser.feed( open( 'main.htm' ).read( ) ) import urlparse for a in parser.anchorlist: print urlparse.urljoin( 'http://python.org/', a ) Output snipped: ... http://python.org/psf/ http://python.org/dev/ http://python.org/links/ http://python.org/download/releases/2.5.2 http://docs.python.org/ http://python.org/ftp/python/2.5.2/python-2.5.2.msi ... -- http://mail.python.org/mailman/listinfo/python-list