You don’t just want an article or an individual image, you want the whole web site. What’s the easiest way to siphon it all?

Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-driven grouping of Q&A web sites.

Image available as wallpaper at GoodFon.

The Question

SuperUser reader Joe has a simple request:

How can I download all pages from a website?

Any platform is fine.

Every page, no exception. Joe’s on a mission.

The Answer

SuperUser contributor Axxmasterr offers an application recommendation:

http://www.httrack.com/

HTTRACK works like a champ for copying the contents of an entire site. This tool can even grab the pieces needed to make a website with active code content work offline. I am amazed at the stuff it can replicate offline.

This program will do all you require of it.

Happy hunting!

We can heartily recomment HTTRACK. It’s a mature application that gets the job done. What about archivists on non-Windows platforms? Another contributor, Jonik, suggests another mature and powerful tool:

Wget is a classic command-line tool for this kind of task. It comes with most Unix/Linux systems, and you can get it for Windows too (newer 1.13.4 available here).

You’d do something like:

wget -r --no-parent http://site.com/songs/

For more details, see Wget Manual and its examples, or take a look at these:


Have something to add to the explanation? Sound off in the the comments. Want to read more answers from other tech-savvy Stack Exchange users? Check out the full discussion thread here.