It depends on the website and what you want to do with it.  Some sites are
so javascript heavy and change all the time that this sort of tool is not
really appropriate (twitter for example).  Usually those web sites have
APIs for accessing their data.  Unfortunately this means rolling your own
solution or using a specialized utility (like thinkup in the case of
twitter).

If you just want to download relatively simple sites then httrack will work
reasonably well.  The flag "--stay-on-same-domain" is probably what you
want, or perhaps "--can-go-down" if you only want a certain subdirectory.
 Even still, you will not get a perfectly mirrored experience with
javascript heavy sites.

Again there is no simple one program fits all use cases.


On Sun, Mar 2, 2014 at 11:18 PM, Joe S <joes...@shaw.ca> wrote:

> I am trying to download a website, but stay inside it, not
> follow links going outside, but follow links inside the site.
> Would another program work better for this?
>
> -----Original Message-----
> From Gustin Johnson <gus...@meganerd.ca>
> To CLUG General <clug-talk@clug.ca>
> Suject Re: [clug-talk] Webhttrack
> Date Fri, 28 Feb 2014 14:58:09 -0700
>
> > I am also curious what it is that you are trying to do.  This
> > may be a case of using the wrong tool for the job.  With so
> > many choices I find the most difficult part is just knowing
> > what package or project best solves my problem(s).
>
> _______________________________________________
> clug-talk mailing list
> clug-talk@clug.ca
> http://clug.ca/mailman/listinfo/clug-talk_clug.ca
> Mailing List Guidelines (http://clug.ca/ml_guidelines.php)
> **Please remove these lines when replying
>
_______________________________________________
clug-talk mailing list
clug-talk@clug.ca
http://clug.ca/mailman/listinfo/clug-talk_clug.ca
Mailing List Guidelines (http://clug.ca/ml_guidelines.php)
**Please remove these lines when replying

Reply via email to