I would like to use a wikipedia dump in my installation of Mediawiki 1.6.7 here http://encyclopedia.meta99.com/, using the proceduce found here :
http://en.wikipedia.org/wiki/Wikipedia:Database_download http://meta.wikimedia.org/wiki/Data_dumps
I have used mwdumper, because tutorial ask that importDump is slow. mwdumper mwdumper is a standalone program for filtering and converting XML dumps. It can produce output as another XML dump as well as SQL statements for inserting data directly into a database in MediaWiki's 1.4 or 1.5 schema.
I have lauched the command in the server :
[root@myserver maintenance]# java -jar mwdumper.jar --format=sql:1.5 enwiki-latest-pages-articles.xml.bz2 | mysql -u databaseuser -p databasename Enter password: "i have used my database password"
1,000 pages (19.529/sec), 1,000 revs (19.529/sec) ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1 2,000 pages (21.128/sec), 2,000 revs (21.128/sec) 3,000 pages (21.702/sec), 3,000 revs (21.702/sec) 4,000 pages (21.048/sec), 4,000 revs (21.048/sec) 5,000 pages (20.578/sec), 5,000 revs (20.578/sec) 6,000 pages (20.369/sec), 6,000 revs (20.369/sec) 7,000 pages (20.078/sec), 7,000 revs (20.078/sec) 8,000 pages (20.085/sec), 8,000 revs (20.085/sec) 9,000 pages (20.048/sec), 9,000 revs (20.048/sec) 10,000 pages (20.223/sec), 10,000 revs (20.223/sec) 11,000 pages (20.159/sec), 11,000 revs (20.159/sec) 12,000 pages (20.117/sec), 12,000 revs (20.117/sec) 13,000 pages (19.962/sec), 13,000 revs (19.962/sec) 14,000 pages (20.244/sec), 14,000 revs (20.244/sec) 15,000 pages (20.204/sec), 15,000 revs (20.204/sec) 16,000 pages (20.167/sec), 16,000 revs (20.167/sec) 17,000 pages (20.259/sec), 17,000 revs (20.259/sec) 18,000 pages (20.18/sec), 18,000 revs (20.18/sec) ..... ..... .....
The problem is that it is the first time in my life that i use MediaWiki and i'm not an expert.
1) I continue to have not any page in the site where i have installed wikimedia and in the database i can not see directly any new entry/page in the site. Exist a system to verify that the dump is inserted really in the database = In the site i use with wikimedia i have not any new page here : http://encyclopedia.meta99.com/ Exist a system to verify if the dump pages are in the database ? In a message of your list i have found about a similar problem http://64.233.183.104/search?q=cache:3LLMxLN9gPcJ:mail.wikipedia.org/piperma... "2) Are you using read-only mode on the wiki? "
I have installed mediawiki using the standard procedure and at the moment i have not did any changing in the configuration/setting. To use the dump i have to make some particular setting/configuration ???
2) Why this error, after launching the script ? [root@myserver maintenance]# java -jar mwdumper.jar --format=sql:1.5 enwiki-latest-pages-articles.xml.bz2 | mysql -u databaseuser -p databasename Enter password: "i have used my database password"
1,000 pages (19.529/sec), 1,000 revs (19.529/sec) ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1 Have you an idea ?
3) When mwdumper.jar has finished the work what i have to do, to use the dump in my system ? I have not found instructions about that in the wikimedia site. When msdumper has finished i need to make some particular setting/configuration in my installation of mediawiki to use the pages in my site/system ?
If you must cross-post, please send the mail with all headers properly so we have a chance of sending the reply correctly.
-- brion vibber (brion @ pobox.com)
Thank you very much.
query SQL: TRUNCATE `page` ;
TRUNCATE `revision` ;
TRUNCATE `text` ;
Is it sufficient ? Or i have to delete the entire 3 tables ? I need to launch other queries ? ------------ I will restart the procedure, after the deletion of the 3 tables you meant. At the moment this is the structure in the database : : archive categorylinks externallinks hitcounter image imagelinks interwiki ipblocks job logging math objectcache oldimage page (DELETED) pagelinks querycache recentchanges revision (DELETED) searchindex site_stats templatelinks text (DELETED) trackbacks transcache user user_groups user_newtalk validate watchlist
----- Original Message ----- From: "Brion Vibber" brion@pobox.com To: "Wikimedia developers" wikitech-l@wikimedia.org Sent: Monday, July 03, 2006 7:27 AM Subject: Re: [Wikitech-l] Problems using mwdumper.jar
Ulisse Savi wrote:
1,000 pages (19.529/sec), 1,000 revs (19.529/sec) ERROR 1062 (23000) at line 46: Duplicate entry '1' for key 1
You need to import into an empty database, or it will fail like this.
Before you start, try clearing the tables that mwdumper works in:
DELETE FROM page; DELETE FROM revision; DELETE FROM text;
-- brion vibber (brion @ pobox.com)
Wikitech-l mailing list Wikitech-l@wikimedia.org http://mail.wikipedia.org/mailman/listinfo/wikitech-l
What would be required to have a MediWiki page of regex expressions that could be used to prevent the creation of new user accounts matching the pattern?
On 04/07/06, xaosflux xaosflux@gmail.com wrote:
What would be required to have a MediWiki page of regex expressions that could be used to prevent the creation of new user accounts matching the pattern?
Someone to install the username blacklist extension.
Rob Church
wikitech-l@lists.wikimedia.org