Thanks Arlo for replying.
Can you please give me some referenc elink to the native parser pf
Mediawiki that you have suggested? A native parser will always be the
easiest way to cater our need. We want to convert the pages pf our
Mediawiki (1.17.5) to either pdf or html pages . All the
attachments (images), comments should also come as part of the output file.
Thanks
Ayaskant
On Fri, Feb 7, 2020 at 9:55 PM Arlo Breault <abreault(a)wikimedia.org> wrote:
On Feb 6, 2020, at 4:35 AM, Ayaskant Swain
<ayaskant.swain(a)gmail.com>
wrote:
Hi Parsoid Support Team,
I am reaching out to you to know about the usage of this tool. We have a
very old
version 1.17.5 of Mediawiki in our organization and want to
convert the pages of it to html pages and store it on disk for archiving.
As you know internally Mediawiki stores pages as WikiText.
Maybe? It's very likely that Parsoid will have some compatibility
issues that you'll need to hack around.
I also saw the documentation of VisualEditor
extension (
https://www.mediawiki.org/wiki/VisualEditor) which uses parsoid
internally to convert wikitext pages. Which tool among these 2 should we
use to do my job? Can you please suggest?
Parsoid is not included in VE, it just offers an API for VE to query.
Adding VE to the mix is an unnecessary complication.
can parsoid be used as a standalone application
or tool instead of VE?
Yes
If we use any of them do we need to just provide
the url of our
Mediawiki page (example - https://<our_dns_host>/wiki/TestPage
or do we
need to extract the content from DB which is in WikiText format and feed it
to parsoid for converting it to html page?
Parsoid has traditionally interacted with MediaWiki's action API
(the thing at /api.php). You would not need to do any manual extraction.
There seems to be an active project similar to what you're describing
at,
https://github.com/openzim/mwoffliner
However, it might be less complicated to just use the parser
that ships with the MediaWiki version you're running. In other
words, screen scrape the pages MediaWiki is already serving you.
Thanks
_______________________________________________
Wikitext-l mailing list
Wikitext-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitext-l