Archive a website

On occasion I find a small website that is an amazing resource, these sites are often negelected, and can occasionally disappear from the web. I find that sometimes a site is so valuable, that I'll check that it is in the wayback machine. If it is not, then I will archive the site myself, for personal reference.

This post serves as a reference for creating that archive using wget.

wget \
    --recursive \
    --level 10 \
    --page-requisites \
    --adjust-extension \
    --convert-links \
    --domains \
    --no-parent \

--recursive Turn on recursive retrieving.

--level 10 Increase the default recursive depth from 5 to 10.

--page-requisites Download all the files that are necessary to properly display a given HTML page.

--adjust-extension Add a suitable file extension depending on what type of file is being downloaded.

--convert-links After the download is complete, convert the links in the document to make them suitable for local viewing.

--domains Set domains to be followed.

--no-parent Do not ascent to the parent directory when retrieving recursively. Very useful if you only wish to archive a single page, or a small section of a site.