

#Sitesucker for windows 10 zip
Once you download a site, you can zip its folder and then back that up the way you would any of your other files. I'm still a novice at HTTrack, but from my experience so far, I've found that it captures only ~90% of a website's individual pages on average. For some websites (like the one you're reading now), HTTrack seems to capture everything, but for other sites, it misses some pages. Maybe this is because of complications with redirects? I'm not sure. Still, ~90% backup is much better than 0%. You can verify which pages got backed up by opening the domain's index.html file from HTTrack's download folder and browsing around using the files on your hard drive. It's best if you disconnect from the Internet when doing this because I found that if I was online when browsing around the downloaded file contents, some pages got loaded from the Internet, not from the local files that I was testing. Pictures don't seem to load offline, but you can check that they're still being downloaded. For example, for WordPress site downloads, look at the \wp-content\uploads folder.
#Sitesucker for windows 10 full
I won't explain the full how-to steps of using HTTrack, but below are two problems that I ran into. When I tried to use HTTrack to download a single website using the program's default settings (as of Nov. 2016), I downloaded the website but also got some other random files from other domains, presumably from links on the main domain. In some cases, the number of links that the program tried to download grew without limit, and I had to cancel. In order to download files only from the desired domain, I had to do the following. Step 1: Specify the domain(s) to download (as I had already been doing). Step 2: Add a Scan Rules pattern like this: +. This way, only links on that domain will be downloaded. Including a * before the main domain name is useful in case the site has subdomains. Troubleshooting: Error: "Forbidden" (403) For example, the site has a subdomain, which would be missed if you only used the pattern +*. Some pages gave me a "Forbidden" error, which prevented any content from being downloaded. I was able to fix this by clicking on "Set options.", choosing the "Browser ID" tab, and then changing "Browser 'Identity'" from the default of "Mozilla/4.5 (compatible: HTTrack 3.0x Windows 98)" to "Java1.1.4". I chose the Java identity because it didn't contain the substring "HTTrack", which may have been the reason I was being blocked. On Mac, I download websites using SiteSucker. This page gives configuration details that I use when downloading certain sites.

I think website downloads using the above methods don't include the redirects that a site may be using. A redirect ensures that an old link doesn't break when you move a page to a new url.
