Shell: Download a website with wget
Caution
This article was published more than a year ago, there may have been developments.
Please take this into account.
In today's article we see how you download an entire website with the command “wget
“.
Wget is a download manager open source, included in major Linux distributions. It is a tool command line, of which there is a gui for Gnome (“wget“) and a gui for Windows (“wgetGUI“). Today we see some simple instruction, but for those who want to delve into the manual is always available from the terminal by typing “man wget
“.
Using wget to download a file, also large:
$ wget -c http://dominio.ext / file.iso
The option “-c
” is not essential, but especially for large files it makes the download can break at any point, launching in the future the same command from the same directory as the download will be resumed from the same point where we left.
Using wget to download an entire website with all internal links:
$ wget -r http://dominio.ext /
Without the topic “-r
” would only download the index.html file located in the root of the site. In this case, instead, “-r
” causes are downloaded even all internal links and images recursively. Be careful with this command, however,, tossing it on a very large download site is likely to never end.
Using wget to download an entire site, but limit 3 the depth level:
$ wget -r -c -l 3 -o wget.log http://dominio.ext /
This other command downloads the site and its internal links, you can resume the download if the interrupt, the link will be downloaded to the third (hierarchically), keeps a log file called “wget.log” in this directory.
Unfortunately, the gui does not have all these options to configure, turns out to be a simple download manager, the actual program, instead, is chock full of configurations for all your needs.
0 Comments