On 4/24/07, Ivan Lanin ivanlanin@gmail.com wrote:
On 4/25/07, Mohamed Magdy mohamed.m.k@gmail.com wrote:
Guys, did you get that email? or you got it but don't know? or you got but just ignoring me because my questions are so ridicules and I should go RTFM.. if so..where?
You could start from http://meta.wikimedia.org/wiki/Data_dumps
HTH, IRL.
I knew that it existed from this page ;)
I knew a couple of things..
1-if you used it with the uncompressed xmls, it will be faster (obviously ) than the bziped... 2-it doesn't matter if it ends before it finishes because it don't reinsert the stuff again but I think it just checks if there is a difference between the records in the db and the xml file...because it goes through the previously inserted records fast then slows when it starts to enter completely new records...and that is great IMO but wouldn't it be nice if we can specify from which record it starts? 3-I didn't try it yet under linux but with windows, it takes a lot (almost) all computer resources, it even makes another programs to close..just ends them...weird.. 4-I think it don't need db details because it uses Special:Import that uses mediawiki and also that is why it is slower than mysqlimport....
Can someone add this feature? the ability for it to sleep for x number of seconds after importing y number of pages? and we can set them in it...
What is "HTH, IRL." anyway?