On Mon, May 11, 2009 at 3:07 PM, Platonides <Platonides(a)gmail.com> wrote:
Robert Rohde wrote:
Since people are doing dump redesign right now,
might I suggest that
providing better integration / information on Commons-hosted images
would actually be useful. As far as I know the current system has no
way to distinguish between Commons images and missing images except by
downloading the Commons dump files. That can be frustrating since the
Commons dumps are larger (and hence more trouble to work with) than
all but a handful of other wikis.
-Robert Rohde
You only need the image.sql dump from commons to determine if the image
exists there (it will also include other useful and not-so-useful data
like filetype, image size, metadata...).
http://download.wikimedia.org/commonswiki/20090510/
That wouldn't get you file descriptions or copyright status, etc. If
your goal is something like mirroring a wiki, you really need access
to page descriptions as well.
At present, the main solution is to copy all of Commons, which is
overkill for many applications. It would be nice if the dump
generator had a way of parsing out only the relevant Commons content.
-Robert Rohde