This was helpfull. Do you know of any java framwework which transforms the raw wiki format into a document tree? Would that be usefull?
Thanks for the feedback, Stan
-----Original Message----- From: wikitech-l-bounces@wikimedia.org on behalf of Rowan Collins Sent: Tue 1/11/2005 1:30 AM To: Wikimedia developers Subject: Re: [Wikitech-l] Wiki Webservices
On Mon, 10 Jan 2005 12:00:07 -0500, Stan Wiechers Stan@rga.com wrote:
I have a quick question: What is the best way to integrate wikipedia into other applications without using the html interface? I would like to integrate the search functionality and the content docs into my app. Is there an xml interface to the wikipedia? I would like to bypass screen scraping. Ideal would be a REST style interface.
Well, there's "&action=raw" to get the raw wikitext of a page. Or look at http://pywikipediabot.sf.net for a Python bot framework for interacting with the site. Or download the database from http://download.wikimedia.org and do really crazy stuff to it without loading the wikimedia servers any.
But no, no XML interface. There's been discussion of code to make bot access easier, but I can't find any of it, and I don't think it resulted in much in the way of new code, I'm afraid.
There's a wiki-to-XML parser under development. It is mostly written in Bison and will compile to C/C++. Try the CVS module "flexbisonparse".
Magnus
Stan Wiechers schrieb:
This was helpfull. Do you know of any java framwework which transforms the raw wiki format into a document tree? Would that be usefull?
Thanks for the feedback, Stan
-----Original Message----- From: wikitech-l-bounces@wikimedia.org on behalf of Rowan Collins Sent: Tue 1/11/2005 1:30 AM To: Wikimedia developers Subject: Re: [Wikitech-l] Wiki Webservices
On Mon, 10 Jan 2005 12:00:07 -0500, Stan Wiechers Stan@rga.com wrote:
I have a quick question: What is the best way to integrate wikipedia into other applications without using the html interface? I would like to integrate the search functionality and the content docs into my app. Is there an xml interface to the wikipedia? I would like to bypass screen scraping. Ideal would be a REST style interface.
Well, there's "&action=raw" to get the raw wikitext of a page. Or look at http://pywikipediabot.sf.net for a Python bot framework for interacting with the site. Or download the database from http://download.wikimedia.org and do really crazy stuff to it without loading the wikimedia servers any.
But no, no XML interface. There's been discussion of code to make bot access easier, but I can't find any of it, and I don't think it resulted in much in the way of new code, I'm afraid.
Wikitech-l mailing list Wikitech-l@wikimedia.org http://mail.wikipedia.org/mailman/listinfo/wikitech-l
-----Original Message-----
Date: Tue, 11 Jan 2005 09:14:21 +0100 Subject: Re: [Wikitech-l] Wiki Webservices From: Magnus Manske magnus.manske@web.de To: Wikimedia developers wikitech-l@wikimedia.org
There's a wiki-to-XML parser under development. It is mostly written in Bison and will compile to C/C++. Try the CVS module "flexbisonparse".
Hi
I started to write a renderer in Java for the previewer of the Wikipedia Eclipse plugin: http://phpeclipse.de/tiki-index.php?page=Wikipedia+Editor+Plugin But the renderer is incomplete at the moment. Does it make sense to use your C/C++ parser for preview rendering (i.e. generate a HTML page without SQL access)? If yes, is it possible to use templates for the generated HTML pages? Are there binaries for Linux, Windows available?
Axel
axelclk@t-online.de schrieb:
I started to write a renderer in Java for the previewer of the Wikipedia Eclipse plugin: http://phpeclipse.de/tiki-index.php?page=Wikipedia+Editor+Plugin But the renderer is incomplete at the moment. Does it make sense to use your C/C++ parser for preview rendering (i.e. generate a HTML page without SQL access)?
At the moment, the parser is (slightly) incomplete as well, but I'm working to change that. It will be most useful for many purposes to have a single, standardized parser (which will then be the "reference implementation", I hope). No need to write a new renderer for our quite complicated syntax each time.
So yes, I think it would make a lot of sense.
If yes, is it possible to use templates for the generated HTML pages?
First, the parser generates *XML*, not *HTML*. The parser knows nothing about the database or even the language of the underlying wikipedia. The transition from XML to HTML has to be done in a second step, which is relatively easy, as there are lots o' tools that can generate an XML tree, in every programming language. I have done a XML-to-(X)HTML conversion in ParserXML.php, which is part of the mediawiki CVS HEAD.
Second, if by "templates" you mean wiki {{templates}}, this is done in the post-parser part described above. The parser generates a XML structure for each {{template}} or {{{variable}}}.
Are there binaries for Linux, Windows available?
On Linux, you can checkout the flexbisonparse module from the mediawiki CVS, then just "make" it. The resulting executable is "wikiparse", into which you pipe the wiki text and get the XML.
Magnus
wikitech-l@lists.wikimedia.org