Not exactly the same thing, but there is a set of grabber scripts designed to get and import an entire wiki via the api, but these are currently mostly just mw maintenance and a few python scripts that only expect to be used with an actual second mediawiki instance.
That being said, it probably would be a good idea to rearchitect them to be a bit less redundant with each other, and in so doing also make their usage a tad more flexible anyway, such as for creating dumps and backups instead. Whatever the case, at very least they might make a useful reference point.
See: https://www.mediawiki.org/wiki/Manual:Grabbers
-I
On 15/03/2019 18:43, John wrote:
Unfortunately that doesn’t really work unless it’s a fairly small wiki. If it’s bigger the export request times out.
What I was hoping for was a tool that utilizes either special export or the API to build a dump file using multiple requests.
I can probably write something to do the same thing, but was hoping for an existing solution
On Fri, Mar 15, 2019 at 1:12 PM Sterkebak sterkebak@gmail.com wrote:
XML export if enabled should do the trick.
On Fri, 15 Mar 2019 at 18:04, John phoenixoverride@gmail.com wrote:
Are there any tools that allow you to create a dump of a mediawiki
install
that doesn’t require direct database access? My primary focus is on creating a backup of the wiki contents. Thanks _______________________________________________ MediaWiki-l mailing list To unsubscribe, go to: https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
-- Met vriendelijke groet,
Huib Laurens _______________________________________________ MediaWiki-l mailing list To unsubscribe, go to: https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
MediaWiki-l mailing list To unsubscribe, go to: https://lists.wikimedia.org/mailman/listinfo/mediawiki-l