使用wget或curl下载网站进行存档

Posted

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了使用wget或curl下载网站进行存档相关的知识,希望对你有一定的参考价值。

This will start at the specified URL and recursively download pages up to 3 links away from the original page, but only pages which are in the directory of the URL you specified (emacstips/) or one of its subdirectories.

wget will also rewrite the links in the pages it downloaded to make your downloaded copy a useful local copy, and it will download all page prerequisites (e.g. images, stylesheets, and the like).

The last two options -nH --cut-dirs=1 control where wget places the output. If you omitted those two options, wget would, for example, download http://web.psung.name/emacstips/index.html and place it under a subdirectory web.psung.name/emacstips of the current directory. With only -nH ("no host directory") wget would write that same file to a subdirectory emacstips. And with both options wget would write that same file to the current directory. In general, if you want to reduce the number of extraneous directories created, change cut-dirs to be the number of leading directories in your URL.
  1. wget -rkp -l3 -np -nH --cut-dirs=1 http://web.psung.name/emacstips/

以上是关于使用wget或curl下载网站进行存档的主要内容,如果未能解决你的问题,请参考以下文章

如何使用 curl 或 wget 从 GitLab 下载 maven 包?

无法使用 wget 或 curl 身份验证下载

在 cloudflare 上下载/镜像网站以进行存档

在 AWS EC2 上,我无法进行 wget 或 curl 或任何出站连接

curl和wget的区别和用法介绍

现在如何通过 curl 或 wget 下载 WWDC 2010 视频?