You don’t just want an article or an individual image, you want the whole web site. What’s the easiest way to siphon it all?
Today’s Question & Answer session comes to us courtesy of SuperUser—a subdivision of Stack Exchange, a community-driven grouping of Q&A web sites.
Image available as wallpaper at GoodFon.
SuperUser reader Joe has a simple request:
How can I download all pages from a website?
Any platform is fine.
Every page, no exception. Joe’s on a mission.
SuperUser contributor Axxmasterr offers an application recommendation:
HTTRACK works like a champ for copying the contents of an entire site. This tool can even grab the pieces needed to make a website with active code content work offline. I am amazed at the stuff it can replicate offline.
This program will do all you require of it.
We can heartily recomment HTTRACK. It’s a mature application that gets the job done. What about archivists on non-Windows platforms? Another contributor, Jonik, suggests another mature and powerful tool:
Wget is a classic command-line tool for this kind of task. It comes with most Unix/Linux systems, and you can get it for Windows too (newer 1.13.4 available here).
You’d do something like:
wget -r --no-parent http://site.com/songs/
For more details, see Wget Manual and its examples, or take a look at these:
Have something to add to the explanation? Sound off in the the comments. Want to read more answers from other tech-savvy Stack Exchange users? Check out the full discussion thread here.
- › How to Use wget, the Ultimate Command Line Downloading Tool
- › 6 Useful Websites to Download for Offline Access
- › The Best Command Line Tools You Can Get on Your Mac With Homebrew
- › How to Add a Shortcut to Pretty Much Anything on Android
- › Update iTunes on Windows Now to Fix a Security Flaw
- › 5 Ways to See If Your Phone Is Being Tapped
- › 10 Kodi Features You Should Be Using
- › 6 Ways Our Tech Is Better Than Star Trek’s