Martina Greiner wrote:
I am trying to upload the english Wikipedia xml dump
from 20051105 xml. I have
Mysql 5, Wikimedia 1.6 phase 3, and a mac with osx 10.3.
I started the loading process six days ago using importDump.php and the process
list shows me that php, bz, and mysql are still working. However, when I go into
Mysql the row counts of the tables do not go up.
1) Checked open transaction?
2) Are you using read-only mode on the wiki? This will cause all changes to the
database to silently fail to run.
3) For large imports like this use mwdumper if you want decent speed. Please see
http://meta.wikimedia.org/wiki/
For example, the user table has
one row, and the text table has approx. 3500 rows. If I call show processes in
Mysql, then I can see that queries are running (admin tells me that approx 200
queries/second are executed) but they seem not to be executed.
They may all be in a transaction then, in which case the modifications won't be
visible to your other process until committed.
My questions: is it normal to take six days to upload
the Wikipedia dump? is it
normal that the row count does not go up?
importDump.php is relatively inefficient, and is generally expected to be used
with smallish data sets being copied in from another wiki. For bulk imports
you'll get much much better performance with mwdumper.
If not, I would highly appreciated any help.
Unfortunately I cannot use
mwdumper, because OSX 10.3 only allows java 1.4 and with java 1.4 I get
exception errors.
Java 1.5 for Mac OS X can be downloaded from
www.apple.com.
1.4 will still be the default JVM, but you can run 1.5 specifically with:
/System/Library/Frameworks/JavaVM.framework/Versions/1.5/Commands/java
-- brion vibber (brion @
pobox.com)