Hello to all,
i've a problem with making an offline copy of my own local wiki.
The following szenario: All the admin-related stuff and also the whole documentation of our infrastructure is stored on a webserver. So the document to restore the webserver if it crashes is located on the same webserver.... so therefore i need an offline copy of my local wiki. (I know this is not the perfect solution but there is a backup from the wiki - but i need a simple way to have the important documents stored on my local computer).
I've searched the web and found some solutions:
o Script 'dumpHtml.php in the 'maintenance'-Folder: The script works and stores all the Wiki-Sites as static HTML-pages in a given folder ('php dumpHtml.php -d /my/folder/') BUT the images are lost if you copy the directory to your local pc (in apache) and the relative links inside the pages are broken (why is there '../../' in front of all links). What is the intension of this script? Am i doing something wrong? Links: - http://svn.wikimedia.org/viewvc/mediawiki/trunk/phase3/maintenance/ dumpHTML.php?view=markup
o Serveral Alternative Parsers: See http://meta.wikimedia.org/wiki/Alternative_parsers for more Information. I tried the HTML2FPDF and Mediawiki Article (http:// meta.wikimedia.org/wiki/HTML2FPDF_and_Mediawiki) but i did not get it working. Also it's not good that you habe to change some mainfiles from the mediawiki installation. The other projects are imho not 'ready' or the intension is something different.
o wget to mirror the wiki-Site I also tried mirroring the wiki with 'wget -m http://mydomain.com/ mywikidirectory/' (i also tried the url http://mydomain.com/ mywikidirectory/index.php/MainSite) but it is not only mirroring the wiki. It's also mirroring the whole site at 'http://mydomain.com/'. Why? Can i customize the wget command that it is only mirroring sites from 'http://mydomain.com/mywikidirectory/%27?
Are there any other solutions or suggestions? I think there is a need for such a offline copy because i found many questions for a tool like this. Is there any chance to customize the script dumpHTML.php?
Best regards.
Carsten Marx
-Links: - http://meta.wikimedia.org/wiki/Talk:MediaWiki_FAQ#HTML_export
My Wiki installtion: - MediaWiki: 1.7.1 - PHP: 5.1.4-Debian-0.1~sarge1 (apache2handler) - MySQL: 4.1.11-Debian_4sarge5-log
On 18/08/06, Carsten Marx carsten@merkste-was.de wrote:
o Script 'dumpHtml.php in the 'maintenance'-Folder: The script works and stores all the Wiki-Sites as static HTML-pages in a given folder ('php dumpHtml.php -d /my/folder/') BUT the images are lost if you copy the directory to your local pc (in apache) and the relative links inside the pages are broken (why is there '../../' in front of all links). What is the intension of this script? Am i doing something wrong?
The purpose of the script is to dump static HTML versions of all pages. Maintain the relative relationship between "images" and the dump directories to ensure that images continue to be accessible, i.e. dump to /path/to/wiki/html, and copy the ../wiki/images and ../wiki/html directories to wherever you're keeping the documentation locally.
The other projects are imho not 'ready' or the intension is something different.
Both.
Why? Can i customize the wget command that it is only mirroring sites from 'http://mydomain.com/mywikidirectory/%27?
man wget
Rob Church
On Friday 18 August 2006 10:03, Carsten Marx wrote:
Hello to all,
i've a problem with making an offline copy of my own local wiki.
The following szenario: All the admin-related stuff and also the whole documentation of our infrastructure is stored on a webserver. So the document to restore the webserver if it crashes is located on the same webserver.... so therefore i need an offline copy of my local wiki. (I know this is not the perfect solution but there is a backup from the wiki - but i need a simple way to have the important documents stored on my local computer).
What about a mirror of the wiki? Dump the database to another machine, upload it at regular intervals (once a day)?
On the other hand, if your entire datacenter goes down, then a hard copy would be a happy thing.
<snip>
o Serveral Alternative Parsers: See http://meta.wikimedia.org/wiki/Alternative_parsers for more Information. I tried the HTML2FPDF and Mediawiki Article (http:// meta.wikimedia.org/wiki/HTML2FPDF_and_Mediawiki) but i did not get it working. Also it's not good that you habe to change some mainfiles from the mediawiki installation. The other projects are imho not 'ready' or the intension is something different.
We are working on something with a table of content file (Stückliste) that contains the articles we want to load and in the order we want/need them printed. Adding printable=yes to the URL we strip out all of the junk, parse the file a little more and then create a PDF.
o wget to mirror the wiki-Site I also tried mirroring the wiki with 'wget -m http://mydomain.com/ mywikidirectory/' (i also tried the url http://mydomain.com/ mywikidirectory/index.php/MainSite) but it is not only mirroring the wiki. It's also mirroring the whole site at 'http://mydomain.com/'. Why? Can i customize the wget command that it is only mirroring sites from 'http://mydomain.com/mywikidirectory/%27?
That'll just get you the articles as files on your local system. Although it is probably useful in an emergency, there are better alternatives.
mediawiki-l@lists.wikimedia.org