[Mediawiki-l] Problems using mwdumper.jar

Ulisse Savi ulisavi at tin.it
Mon Jul 3 03:49:38 UTC 2006


I would like to use a wikipedia dump in my installation of Mediawiki 1.6.7 here http://encyclopedia.meta99.com/, using the proceduce found here :

http://en.wikipedia.org/wiki/Wikipedia:Database_download
http://meta.wikimedia.org/wiki/Data_dumps

I have used mwdumper, because tutorial ask that importDump is slow.
mwdumper
mwdumper is a standalone program for filtering and converting XML dumps. It can produce output as another XML dump as well as SQL statements for inserting data directly into a database in MediaWiki's 1.4 or 1.5 schema.

I have lauched the command in the server :

[root at myserver maintenance]# java -jar mwdumper.jar --format=sql:1.5 enwiki-latest-pages-articles.xml.bz2 | mysql -u databaseuser -p databasename
Enter password: "i have used my database password"

1,000 pages (19.529/sec), 1,000 revs (19.529/sec)
ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1
2,000 pages (21.128/sec), 2,000 revs (21.128/sec)
3,000 pages (21.702/sec), 3,000 revs (21.702/sec)
4,000 pages (21.048/sec), 4,000 revs (21.048/sec)
5,000 pages (20.578/sec), 5,000 revs (20.578/sec)
6,000 pages (20.369/sec), 6,000 revs (20.369/sec)
7,000 pages (20.078/sec), 7,000 revs (20.078/sec)
8,000 pages (20.085/sec), 8,000 revs (20.085/sec)
9,000 pages (20.048/sec), 9,000 revs (20.048/sec)
10,000 pages (20.223/sec), 10,000 revs (20.223/sec)
11,000 pages (20.159/sec), 11,000 revs (20.159/sec)
12,000 pages (20.117/sec), 12,000 revs (20.117/sec)
13,000 pages (19.962/sec), 13,000 revs (19.962/sec)
14,000 pages (20.244/sec), 14,000 revs (20.244/sec)
15,000 pages (20.204/sec), 15,000 revs (20.204/sec)
16,000 pages (20.167/sec), 16,000 revs (20.167/sec)
17,000 pages (20.259/sec), 17,000 revs (20.259/sec)
18,000 pages (20.18/sec), 18,000 revs (20.18/sec)
.....
.....
.....

The problem is that it is the first time in my life that i use MediaWiki and i'm not an expert.

1) I continue to have not any page in the site where i have installed wikimedia and in the database i can not see directly any new entry/page in the site. Exist a system to verify that the dump is inserted really in the database = In the site i use with wikimedia i have not any new page here :
http://encyclopedia.meta99.com/
Exist a system to verify if the dump pages are in the database ?
In a message of your list i have found about a similar problem 
http://64.233.183.104/search?q=cache:3LLMxLN9gPcJ:mail.wikipedia.org/pipermail/wikitech-l/2005-November/032747.html+importDump.php&hl=en&ct=clnk&cd=3
"2) Are you using read-only mode on the wiki? "

I have installed mediawiki using the standard procedure and at the moment i have not did any changing in the configuration/setting. To use the dump i have to make some particular setting/configuration ???

2) Why this error, after launching the script ?
[root at myserver maintenance]# java -jar mwdumper.jar --format=sql:1.5 enwiki-latest-pages-articles.xml.bz2 | mysql -u databaseuser -p databasename
Enter password: "i have used my database password"

1,000 pages (19.529/sec), 1,000 revs (19.529/sec)
ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1
Have you an idea ?

3) When mwdumper.jar has finished the work what i have to do, to use the dump in my system ? I have not found instructions about that in the wikimedia site. When msdumper has finished i need to make some particular setting/configuration in my installation of mediawiki to use the pages in my site/system ? 





More information about the MediaWiki-l mailing list