It depends on the website and what you want to do with it. Some sites are so javascript heavy and change all the time that this sort of tool is not really appropriate (twitter for example). Usually those web sites have APIs for accessing their data. Unfortunately this means rolling your own solution or using a specialized utility (like thinkup in the case of twitter).
If you just want to download relatively simple sites then httrack will work reasonably well. The flag "--stay-on-same-domain" is probably what you want, or perhaps "--can-go-down" if you only want a certain subdirectory. Even still, you will not get a perfectly mirrored experience with javascript heavy sites. Again there is no simple one program fits all use cases. On Sun, Mar 2, 2014 at 11:18 PM, Joe S <joes...@shaw.ca> wrote: > I am trying to download a website, but stay inside it, not > follow links going outside, but follow links inside the site. > Would another program work better for this? > > -----Original Message----- > From Gustin Johnson <gus...@meganerd.ca> > To CLUG General <clug-talk@clug.ca> > Suject Re: [clug-talk] Webhttrack > Date Fri, 28 Feb 2014 14:58:09 -0700 > > > I am also curious what it is that you are trying to do. This > > may be a case of using the wrong tool for the job. With so > > many choices I find the most difficult part is just knowing > > what package or project best solves my problem(s). > > _______________________________________________ > clug-talk mailing list > clug-talk@clug.ca > http://clug.ca/mailman/listinfo/clug-talk_clug.ca > Mailing List Guidelines (http://clug.ca/ml_guidelines.php) > **Please remove these lines when replying >
_______________________________________________ clug-talk mailing list clug-talk@clug.ca http://clug.ca/mailman/listinfo/clug-talk_clug.ca Mailing List Guidelines (http://clug.ca/ml_guidelines.php) **Please remove these lines when replying