Not exactly the same thing, but there is a set of grabber scripts
designed to get and import an entire wiki via the api, but these are
currently mostly just mw maintenance and a few python scripts that only
expect to be used with an actual second mediawiki instance.
That being said, it probably would be a good idea to rearchitect them to
be a bit less redundant with each other, and in so doing also make their
usage a tad more flexible anyway, such as for creating dumps and backups
instead. Whatever the case, at very least they might make a useful
reference point.
See:
https://www.mediawiki.org/wiki/Manual:Grabbers
-I
On 15/03/2019 18:43, John wrote:
Unfortunately that doesn’t really work unless it’s a
fairly small wiki. If
it’s bigger the export request times out.
What I was hoping for was a tool that utilizes either special export or the
API to build a dump file using multiple requests.
I can probably write something to do the same thing, but was hoping for an
existing solution
On Fri, Mar 15, 2019 at 1:12 PM Sterkebak <sterkebak(a)gmail.com> wrote:
XML export if enabled should do the trick.
On Fri, 15 Mar 2019 at 18:04, John <phoenixoverride(a)gmail.com> wrote:
Are there any tools that allow you to create a
dump of a mediawiki
install
that doesn’t require direct database access? My
primary focus is on
creating a backup of the wiki contents.
Thanks
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
--
Met vriendelijke groet,
Huib Laurens
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l