On 4/24/07, Ivan Lanin <ivanlanin(a)gmail.com> wrote:
On 4/25/07, Mohamed Magdy <mohamed.m.k(a)gmail.com> wrote:
Guys, did you get that email? or you got it but
don't know? or you got
but just ignoring me because my questions are so ridicules and I should
go RTFM.. if so..where?
You could start from
http://meta.wikimedia.org/wiki/Data_dumps
HTH, IRL.
I knew that it existed from this page ;)
I knew a couple of things..
1-if you used it with the uncompressed xmls, it will be faster (obviously
) than the bziped...
2-it doesn't matter if it ends before it finishes because it don't reinsert
the stuff again but I think it just checks if there is a difference between
the records in the db and the xml file...because it goes through the
previously inserted records fast then slows when it starts to enter
completely new records...and that is great IMO but wouldn't it be nice if we
can specify from which record it starts?
3-I didn't try it yet under linux but with windows, it takes a lot (almost)
all computer resources, it even makes another programs to close..just ends
them...weird..
4-I think it don't need db details because it uses Special:Import that uses
mediawiki and also that is why it is slower than mysqlimport....
Can someone add this feature? the ability for it to sleep for x number of
seconds after importing y number of pages? and we can set them in it...
What is "HTH, IRL." anyway?