Hello,
I recently discover kiwix and the zim file. It seems a really great
project for offline use, even if there is still a lot of challenges.
For now, as a traveller, I would want to have a recent & usable dump
of some wiki (wikipedia, wikitravel, wiki.couchsurfing, wikivoyage,
...) and find really hard to achieve it.
As a reference, I found
http://en.wikipedia.org/wiki/Wikipedia:Database_download
http://www.kiwix.org/index.php/Main_Page
http://wikitravel.org/en/Wikitravel:Offline_Reader_Expedition
http://www.wikivoyage.org/tech/Database_dumps
But in most case, the best I have is
* a mediawiki xml archive
* last, need to mirror website with httrack or wget (and ensuring to
copy only one language release: only en, fr, de, ...)
In this case, is there any available scripts to build a zim file ?
I have checked
http://www.kiwix.org/index.php/Tools but as I'm in
travel, I don't have time to make each atomic operations, would really
prefer a batch.
Else, thanks a lot for your work.
Can't wait to have it available on different devices (computer,
smartphone, ...) with all wiki :)
Cheers
Julien