This is a question I get quite often. So, I will jump in and make a few suggestions to get the ball rolling
Forensic Acquisition of Websites (FAW)
FAW is a commercial tool from Italy that goes into great depth in this area. I recommend reviewing their features overview and user manual to learn more about the tool’s capabilities such as Wireshark integration
HTTrack is an excellent tool for mirroring websites with countless options. You can find Fred Cohen’s comprehensive guide on its usage here:
Cohen offers the following snippet to achieve a “forensic dump” of a website:
Hunchly has an OSINT focus and allows you to build a case as you visit web pages. I would recommend taking a look if you plan to preserve web pages as you visit them, rather than batch-preserve an entire website.
Trusted timestamping is typically an important part of forensic capture of websites as the goal is often to memorialize what a website contained—or did not contain—at a specific point in time. We have some information on how to accomplish this with open-source tools here:
I hope this helps get you started. Looking forward to hearing suggestions from others as well
Nice topic, I’m particularly interested in this field: to whom it may concern, I just held a presentation for OSDFCON Conference in December 2021 about " Forensic Acquisition of Websites, Webpages and Online Services with Open Source Tools" where I explain how to perform a forensic acquisition of websites and webpages for free, with open source tools (comparing also some state of the art tools and services, which by the way I currently use).