I would like to use a wikipedia dump in my installation of Mediawiki 1.6.7 here http://encyclopedia.meta99.com/, using the proceduce found here :
http://en.wikipedia.org/wiki/Wikipedia:Database_download http://meta.wikimedia.org/wiki/Data_dumps
I have used mwdumper, because tutorial ask that importDump is slow. mwdumper mwdumper is a standalone program for filtering and converting XML dumps. It can produce output as another XML dump as well as SQL statements for inserting data directly into a database in MediaWiki's 1.4 or 1.5 schema.
I have lauched the command in the server :
[root@myserver maintenance]# java -jar mwdumper.jar --format=sql:1.5 enwiki-latest-pages-articles.xml.bz2 | mysql -u databaseuser -p databasename Enter password: "i have used my database password"
1,000 pages (19.529/sec), 1,000 revs (19.529/sec) ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1 2,000 pages (21.128/sec), 2,000 revs (21.128/sec) 3,000 pages (21.702/sec), 3,000 revs (21.702/sec) 4,000 pages (21.048/sec), 4,000 revs (21.048/sec) 5,000 pages (20.578/sec), 5,000 revs (20.578/sec) 6,000 pages (20.369/sec), 6,000 revs (20.369/sec) 7,000 pages (20.078/sec), 7,000 revs (20.078/sec) 8,000 pages (20.085/sec), 8,000 revs (20.085/sec) 9,000 pages (20.048/sec), 9,000 revs (20.048/sec) 10,000 pages (20.223/sec), 10,000 revs (20.223/sec) 11,000 pages (20.159/sec), 11,000 revs (20.159/sec) 12,000 pages (20.117/sec), 12,000 revs (20.117/sec) 13,000 pages (19.962/sec), 13,000 revs (19.962/sec) 14,000 pages (20.244/sec), 14,000 revs (20.244/sec) 15,000 pages (20.204/sec), 15,000 revs (20.204/sec) 16,000 pages (20.167/sec), 16,000 revs (20.167/sec) 17,000 pages (20.259/sec), 17,000 revs (20.259/sec) 18,000 pages (20.18/sec), 18,000 revs (20.18/sec) ..... ..... .....
The problem is that it is the first time in my life that i use MediaWiki and i'm not an expert.
1) I continue to have not any page in the site where i have installed wikimedia and in the database i can not see directly any new entry/page in the site. Exist a system to verify that the dump is inserted really in the database = In the site i use with wikimedia i have not any new page here : http://encyclopedia.meta99.com/ Exist a system to verify if the dump pages are in the database ? In a message of your list i have found about a similar problem http://64.233.183.104/search?q=cache:3LLMxLN9gPcJ:mail.wikipedia.org/piperma... "2) Are you using read-only mode on the wiki? "
I have installed mediawiki using the standard procedure and at the moment i have not did any changing in the configuration/setting. To use the dump i have to make some particular setting/configuration ???
2) Why this error, after launching the script ? [root@myserver maintenance]# java -jar mwdumper.jar --format=sql:1.5 enwiki-latest-pages-articles.xml.bz2 | mysql -u databaseuser -p databasename Enter password: "i have used my database password"
1,000 pages (19.529/sec), 1,000 revs (19.529/sec) ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1 Have you an idea ?
3) When mwdumper.jar has finished the work what i have to do, to use the dump in my system ? I have not found instructions about that in the wikimedia site. When msdumper has finished i need to make some particular setting/configuration in my installation of mediawiki to use the pages in my site/system ?