Martina Greiner wrote:
I am trying to upload the english Wikipedia xml dump from 20051105 xml. I have Mysql 5, Wikimedia 1.6 phase 3, and a mac with osx 10.3.
I started the loading process six days ago using importDump.php and the process list shows me that php, bz, and mysql are still working. However, when I go into Mysql the row counts of the tables do not go up.
1) Checked open transaction? 2) Are you using read-only mode on the wiki? This will cause all changes to the database to silently fail to run. 3) For large imports like this use mwdumper if you want decent speed. Please see http://meta.wikimedia.org/wiki/
For example, the user table has one row, and the text table has approx. 3500 rows. If I call show processes in Mysql, then I can see that queries are running (admin tells me that approx 200 queries/second are executed) but they seem not to be executed.
They may all be in a transaction then, in which case the modifications won't be visible to your other process until committed.
My questions: is it normal to take six days to upload the Wikipedia dump? is it normal that the row count does not go up?
importDump.php is relatively inefficient, and is generally expected to be used with smallish data sets being copied in from another wiki. For bulk imports you'll get much much better performance with mwdumper.
If not, I would highly appreciated any help. Unfortunately I cannot use mwdumper, because OSX 10.3 only allows java 1.4 and with java 1.4 I get exception errors.
Java 1.5 for Mac OS X can be downloaded from www.apple.com.
1.4 will still be the default JVM, but you can run 1.5 specifically with: /System/Library/Frameworks/JavaVM.framework/Versions/1.5/Commands/java
-- brion vibber (brion @ pobox.com)