Wednesday, August 6, 2008

Part Trois

FYI:
If anyone used HTTrack, a common problem is it will download the webpages, but on certain websites it wont download the images.

Websites are complicated thingies behind the scenes, so HTTrack has complicated options to fix that (and its got lousy help texts imo).

But a simple method Ive found is go to "Set Options" --> Spider, and disable the Robots.txt file.

Works for me, maybe it will work for you.

4 comments:

Anonymous said...

Try out wget, if you want something a bit more mature. It is definitely the gold standard, and should have plenty of documentation, although it is entirely command line...

MensaRefugee said...

Anon,
Ill check it out. Thanks!

Phone Sex said...

Let me go for apply first…

Phone Sex said...

Pretty girl gets off wonder what would have happened if she had been older and a bit plain …