Maybe it's obvious, but remember first of all to make an XML dump. Everything else can be regenerated from it. https://www.mediawiki.org/wiki/API:Parsing_wikitext#API_documentation
Arlo Breault, 08/02/20 19:13:
I was suggesting you scrape those pages using wget, Scrapy, HTTrack, or some other tool.
It's also possible this extension works for you, https://www.mediawiki.org/wiki/Extension:DumpHTML
The main issue with archiving the "usual" HTML is that it's hard to tell whether you're including the resources you actually need, for instance CSS for templates. https://phabricator.wikimedia.org/T50295 https://phabricator.wikimedia.org/T40259
I don't recommend using typical scrapers for MediaWiki, it's a can of worms. If you want something that simple, you can get the HTML from the API: https://www.mediawiki.org/wiki/API:Parsing_wikitext#API_documentation
Depending on your installation, using dumpHTML might actually be easier. It was a pain for Wikimedia wikis, but mostly because they're huge and very complicated, plus Kiwix had to import the XML first.
The only advantage of using wget is that you can generate a WARC file with it. WARC can be fed into a warc-proxy which could then serve your website statically: at the moment this is the closest we have to a general purpose static website generator from any CMS. However, you'll still need to train wget (or whatever you're using) how to handle recursion. https://www.archiveteam.org/index.php?title=The_WARC_Ecosystem
Federico