The three processes we had going for "largish" wikis had been restarted from a particilar step, since I had to interrupt them earlier for kernel upgrade and reboot. These stop at the end of the run. Three regular jobs are now running; these cycle through the list of the ten largish wikis in the usual way.
While we're on the subject of de wiki, I have been considering starting to produce smaller output files much as we do for en wikipedia. 100GB is pretty large for someone to download and process, and it takes a while to produce as well. Any thoughts? CC-ing wikitech-l since some people on that list may be users of the dump that don't subscribe to xmldatadumps-l (but they should!)
Ariel
Στις 19-11-2011, ημέρα Σαβ, και ώρα 18:03 +0100, ο/η Andreas Meier έγραψε:
Hello, today we ja-dump was finished, so a new de-dump should start.
Best regards Andreas
wikitech-l@lists.wikimedia.org