Scott,

Nemo is referring to the dumpgenerator.py being broken on MediaWiki versions above 1.20, and it should not actually affect older MediaWiki versions.

You can safely continue with your grab. :)

On Sat, Nov 10, 2012 at 12:45 PM, Scott Boyd <scottdb56@gmail.com> wrote:
At this link: https://code.google.com/p/wikiteam/issues/detail?id=56 , at the bottom, there is an entry by project member nemowiki that states:

     Comment 7 by project member nemowiki, Today (9 hours ago)
    Fixed by emijrp in r806. :-)
      Status: Fixed

So does that mean this problem that "It's completely broken" is now fixed? I'm running a huge download of 64K+ page titles, and am now using the "r806" version of dumpgenerator.py. The first 35K+ page titles were downloaded with an older version). Both versions sure seem to be downloading MORE than 500 pages per namespace, but I'm not sure, since I don't know how you can tell if you are getting them all...

So is it fixed or not?


On Fri, Nov 9, 2012 at 4:27 AM, Federico Leva (Nemo) <nemowiki@gmail.com> wrote:
It's completely broken: https://code.google.com/p/wikiteam/issues/detail?id=56
It will download only a fraction of the wiki, 500 pages at most per namespace.







--
Regards,
Hydriz

We've created the greatest collection of shared knowledge in history. Help protect Wikipedia. Donate now: http://donate.wikimedia.org