In David Waizer <[EMAIL PROTECTED]> wrote:
> Hello..
>
> I'm looking for a script (perl, python, sh...)or program (such as wget)
> that will help me get a list of ALL the links on a website.
lynx -dump (look at the bottom)
--
William Park <[EMAIL PROTECTED]>, Toronto, Canada
ThinFlash:
In article <[EMAIL PROTECTED]>,
"David Waizer" <[EMAIL PROTECTED]> wrote:
> Hello..
>
> I'm looking for a script (perl, python, sh...)or program (such as wget)
> that will help me get a list of ALL the links on a website.
>
> For example ./magicscript.pl www.yahoo.com and outputs it to a file
4 easy steps to get the links:
1. Download BeautifulSoup and import it in your script file.
2. Use urllib2 to download the html of the url.
3. mash the html using BeautifulSoup
4.
[code]
for tag in BeautifulSoupisedHTML.findAll('a'):
print tag
[/code]
David Waizer a écrit :
> Hello..
>
>
Check out the quick start section in the documentation at Beautiful
Soup http://www.crummy.com/software/BeautifulSoup/
Wes
Jonathan Curran wrote:
> On Thursday 18 January 2007 11:57, David Waizer wrote:
> > Hello..
> >
> > I'm looking for a script (perl, python, sh...)or program (such as wget)
On Thursday 18 January 2007 11:57, David Waizer wrote:
> Hello..
>
> I'm looking for a script (perl, python, sh...)or program (such as wget)
> that will help me get a list of ALL the links on a website.
>
> For example ./magicscript.pl www.yahoo.com and outputs it to a file, it
> would be kind of
Hello..
I'm looking for a script (perl, python, sh...)or program (such as wget)
that will help me get a list of ALL the links on a website.
For example ./magicscript.pl www.yahoo.com and outputs it to a file, it
would be kind of like a spidering software..
Any suggestions would be appreciated