wget -p -k http://www.example.com/
Some servers will responsd with a 403 code if you use wget without
a User Agent. For example: -U 'Mozilla/5.0 (X11; U; Linux i686;
en-US; rv:1.8.1.6) Gecko/20070802 SeaMonkey/1.1.4'
If you're finding you're still missing images etc.. then try adding this:
-e robots=off
wget actually reads and respects robots.txt
===================================
To download just a webpage:
wget -E -H -k -p http://example.com/page
This means:
-E: Append .html to the file name if it is an HTML file but doesn't end in .html or similar
-H: Download files from other hosts, too
-k: After downloading convert any link in it so they point to the downloaded files
-p: Download anything the page needs for proper offline viewing