GNU Wget is a free software package for retrieving files using HTTP(S) and FTP, the most widely-used Internet protocols. It is a non-interactive command line tool, so it may easily be called from scripts, cron jobs, terminals without X-Windows support, etc.
Wget vs SiteSucker opinions
HTTrack is a free (GPL, libre/free software) and easy-to-use offline browser utility. It allows you to download a World Wide Web site from the Internet to a local directory, building recursively all directories, getting HTML, images, and other files from the server to your...
The last version (3.49-2) was released in May 2017.
HTTrack vs SiteSucker opinions
It's fairly easy to set up and use. I recommend WAIL instead, though, that uses HTTrack, but has a better GUI
Because modern websites are complicated and often rely on dynamic content, ArchiveBox archives the sites in several different formats beyond what public archiving services like Archive.org and Archive.is are capable of saving.
ScrapBook is a Firefox extension, which helps you to save Web pages and manage the collection. Key features are lightness, speed, accuracy and multi-language support. Major features are:
The extension is not compatible with Firefox 57. It still works on XUL-based Firefox forks like Waterfox
WebScrapBook is a browser extension that captures the web page faithfully with various archive formats and customizable configurations, for future retrieval, organization, annotation, and editing. This project inherits from legacy Firefox add-on ScrapBook X.
ScrapBook X is a Firefox addon based on ScrapBook Plus and also integrating several features from the latest versions of ScrapBook.Discontinued
PageArchiver (previously called "Scrapbook for SingleFile") is a Chrome extension that helps you to archive web pages for offline reading.
Latest commit on May 5, 2013. See https://github.com/gildas-lormeau/Scrapbook-for-SingleFile
grab-site is a crawler for archiving websites to WARC files. It includes a dashboard for monitoring multiple crawls, and supports changing URL ignore patterns during the crawl.