I use mwdumper to import the latest current xml
dump enwiki-20101011-pages-meta-current.xml.bz2 to my mediawiki. Everything
seems fine, however, i found that only 6,669,091 pages in the database,
while the mwdumper stops working and exit at the number 21,894,705.
I am not sure if i have successfully imported all the current pages into
mediawiki. Is there any method for me to verify that? Is there any data on
pages for each dumps for cross referencing purpose? Any method for me to
track what error has encountered (other than viewing the huge log file)?
On the other hand, i found that the parsing efficiency drops from time to
time during the import process. It drops from (345.12/sec) to (79.125/sec).
Is it a normal phenomenon? Any method for me to boost this performance? The
strange part is this figure rise again to around (200/sec) after the 6mil
something page is imported (maybe due to nothing is inserted to the DB
Any sharing of thoughts would be appreciated. Thank you.
Show replies by date