kevin-vance
1/27/2018 - 3:28 AM

wget to download website

wget to download website

wget -p -k http://www.example.com/

Some servers will responsd with a 403 code if you use wget without 
a User Agent. For example: -U 'Mozilla/5.0 (X11; U; Linux i686; 
en-US; rv:1.8.1.6) Gecko/20070802 SeaMonkey/1.1.4'

If you're finding you're still missing images etc.. then try adding this: 
-e robots=off 
wget actually reads and respects robots.txt 

===================================
To download just a webpage: 
wget -E -H -k -p http://example.com/page

This means:
-E: Append .html to the file name if it is an HTML file but doesn't end in .html or similar
-H: Download files from other hosts, too
-k: After downloading convert any link in it so they point to the downloaded files
-p: Download anything the page needs for proper offline viewing