Oh, I should also mention that we are looking into publishing full HTML
dumps once RESTBase is online.
Gabriel
On Fri, Feb 13, 2015 at 10:12 PM, Gabriel Wicke <gwicke(a)wikimedia.org>
wrote:
Nitin,
Parsoid actually has some support for basic template expansion as well,
but it does not re-implement all extensions and parser functions from
MediaWiki. Some of those like Lua scripting are used on many pages.
However, there is a production API for Parsoid HTML at
http://parsoid-lb.eqiad.wikimedia.org/. See
https://www.mediawiki.org/wiki/Parsoid/API for the API docs. This API can
sustain moderate traffic only right now (please don't send more than 10
req/s), but a faster API using RESTBase will come online in the next weeks.
Gabriel
On Fri, Feb 13, 2015 at 9:49 PM, Nitin Gupta <nitingupta910(a)gmail.com>
wrote:
Hi,
I created a nodejs service to convert wikitext to HTML with a frontent
(written in Golang) which reads wikipedia dump and feed wikitext to this
service[1]. However, after doing all this I discovered that Parsoid needs
to contact wikimedia server for template expansion. Since I want to convert
the entire wikipedia dump and HTML, I do not want to keep hitting wikimedia
servers for template expansion requests.
So, are there any plans to add support to Parsoid to do this expansion
offline. Once the wikipedia dump is downloaded, I want the entire process
of converting to HTML to be offline (of course, don't need images).
[1]
https://github.com/nitingupta910/wikiparser
Thanks,
Nitin
_______________________________________________
Wikitext-l mailing list
Wikitext-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitext-l