How Can I Download an Entire Web Site?

You don’t just want an article or an individual image, you want the whole web site. What’s the easiest way to siphon it all?

Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-driven grouping of Q&A web sites.

Image available as wallpaper at GoodFon.

The Question

SuperUser reader Joe has a simple request:

How can I download all pages from a website?

Any platform is fine.

Every page, no exception. Joe’s on a mission.

The Answer

SuperUser contributor Axxmasterr offers an application recommendation:

HTTRACK works like a champ for copying the contents of an entire site. This tool can even grab the pieces needed to make a website with active code content work offline. I am amazed at the stuff it can replicate offline.

This program will do all you require of it.

Happy hunting!

We can heartily recomment HTTRACK. It’s a mature application that gets the job done. What about archivists on non-Windows platforms? Another contributor, Jonik, suggests another mature and powerful tool:

Wget is a classic command-line tool for this kind of task. It comes with most Unix/Linux systems, and you can get it for Windows too (newer 1.13.4 available here).

You’d do something like:

wget -r --no-parent

For more details, see Wget Manual and its examples, or take a look at these:

Have something to add to the explanation? Sound off in the the comments. Want to read more answers from other tech-savvy Stack Exchange users? Check out the full discussion thread here.

Jason Fitzpatrick is a warranty-voiding DIYer who spends his days cracking opening cases and wrestling with code so you don't have to. If it can be modded, optimized, repurposed, or torn apart for fun he's interested (and probably already at the workbench taking it apart). You can follow him on if you'd like.