Scott,
Nemo is referring to the dumpgenerator.py being broken on MediaWiki
versions above 1.20, and it should not actually affect older MediaWiki
versions.
You can safely continue with your grab. :)
On Sat, Nov 10, 2012 at 12:45 PM, Scott Boyd <scottdb56(a)gmail.com> wrote:
At this link:
https://code.google.com/p/wikiteam/issues/detail?id=56 , at
the bottom, there is an entry by project member nemowiki that states:
Comment 7 <https://code.google.com/p/wikiteam/issues/detail?id=56#c7>by
project member
nemowiki <https://code.google.com/u/101255742639286016490/>, Today (9
hours ago)
Fixed by emijrp in r806
<https://code.google.com/p/wikiteam/source/detail?r=806>. :-)
*Status:* Fixed
So does that mean this problem that "It's completely broken" is now fixed?
I'm running a huge download of 64K+ page titles, and am now using the
"r806" version of dumpgenerator.py. The first 35K+ page titles were
downloaded with an older version). Both versions sure seem to be
downloading MORE than 500 pages per namespace, but I'm not sure, since I
don't know how you can tell if you are getting them all...
So is it fixed or not?
On Fri, Nov 9, 2012 at 4:27 AM, Federico Leva (Nemo) <nemowiki(a)gmail.com>wrote;wrote:
It's completely broken:
https://code.google.com/p/**
wikiteam/issues/detail?id=56<https://code.google.com/p/wikiteam/issues/d…
It will download only a fraction of the wiki, 500 pages at most per
namespace.
--
Regards,
Hydriz
We've created the greatest collection of shared knowledge in history. Help
protect Wikipedia. Donate now:
http://donate.wikimedia.org