This has just been brought to my attention by some friends in India, who are considering to use it to create a DVD version of Wikipedia:
http://wikifilter.sourceforge.net/
From the description: "WikiFilter is both a wiki text parser and a web
filter. It converts wiki text stored in a wiki data-base (dump) file into html text, and sends the html text to the web browser." The program is written in C. There's also an indexing component. I haven't tested it, but it looks like it could be a useful module for a DVD edition; the parser component might also be interesting for other purposes.
Erik
Might be worth mentioning that the code is pretty unportable looking Win32 C++. It's setup to serve out pages, acting like a superlight weight read only mediawiki install.. Doesn't appear to be really useful for generating html dumps. :-(
On 7/13/06, Erik Moeller eloquence@gmail.com wrote:
This has just been brought to my attention by some friends in India, who are considering to use it to create a DVD version of Wikipedia:
http://wikifilter.sourceforge.net/
From the description: "WikiFilter is both a wiki text parser and a web
filter. It converts wiki text stored in a wiki data-base (dump) file into html text, and sends the html text to the web browser." The program is written in C. There's also an indexing component. I haven't tested it, but it looks like it could be a useful module for a DVD edition; the parser component might also be interesting for other purposes.
Erik _______________________________________________ Wikitech-l mailing list Wikitech-l@wikimedia.org http://mail.wikipedia.org/mailman/listinfo/wikitech-l
Gregory Maxwell wrote:
Might be worth mentioning that the code is pretty unportable looking Win32 C++. It's setup to serve out pages, acting like a superlight weight read only mediawiki install.. Doesn't appear to be really useful for generating html dumps. :-(
That would probably be a killer, because we'd want any DVD version to work on Windows, Mac and Linux. That being said, I can't really see a DVD installing a *web-server* too in order to be read (I mean, it might not be too bad with something like Lighttpd, but still, that's a lot of baggage).
(Doesn't have a database dump lying around to do testing)
Edward Z. Yang wrote:
That being said, I can't really see a DVD installing a *web-server* too in order to be read
This is how the early Britannica CD sets worked, along with a bunch of other encyclopedias that I've used ages ago. It's not an uncommon approach. You don't have to actually install anything -- simply run the webserver from the DVD/CD directly. The task is made especially easy if you code your webserver in, say, Python, as you can make things reasonably cross-platform without too much effort.
I have a hacker trying to see if the parser component can be easily extracted and used separately.
-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1
Ivan Krstic wrote:
This is how the early Britannica CD sets worked, along with a bunch of other encyclopedias that I've used ages ago. It's not an uncommon approach. You don't have to actually install anything -- simply run the webserver from the DVD/CD directly. The task is made especially easy if you code your webserver in, say, Python, as you can make things reasonably cross-platform without too much effort.
It might be a reasonable idea, then.
I have a hacker trying to see if the parser component can be easily extracted and used separately.
Never underestimate the hideous complexity of Parser.php! Make sure you refer the guy to previous efforts:
http://meta.wikimedia.org/wiki/Alternative_parsers
Never underestimate the hideous complexity of Parser.php! Make sure you refer the guy to previous efforts:
Yes ! and you can be sure in terms of features to be always behind the PHP parser. I'm maybe dumb, but I simply never understood why people invest so much energy for an offline-reader to build a wiki parser, and whereas it's a lot easier to work on a HTML dump.
Emmanuel
Edward Z. Yang wrote:
Never underestimate the hideous complexity of Parser.php!
Actually, what I'm curios about is simply how much work it currently takes to extract the parser that the wikifilter folk already wrote, and make it a standalone library or utility. I'm not looking to add features to the parser itself.
Ivan Krstic wrote:
Edward Z. Yang wrote:
That being said, I can't really see a DVD installing a *web-server* too in order to be read
This is how the early Britannica CD sets worked, along with a bunch of other encyclopedias that I've used ages ago. It's not an uncommon approach. You don't have to actually install anything -- simply run the webserver from the DVD/CD directly. The task is made especially easy if you code your webserver in, say, Python, as you can make things reasonably cross-platform without too much effort.
I have a hacker trying to see if the parser component can be easily extracted and used separately.
Full MediaWiki and zero-install web server with a few demo pages (Windows-only, 23 MB): http://tools.wikimedia.de/~magnus/history05.zip
All I did was add a SpecialDatabase file to read raw text files as articles, which can be generated by my wiki2xml script collection (included).
Magnus
P.S.: I'm not the hacker mentioned above ;-)
wikitech-l@lists.wikimedia.org