On Mon, May 11, 2009 at 3:07 PM, Platonides Platonides@gmail.com wrote:
Robert Rohde wrote:
Since people are doing dump redesign right now, might I suggest that providing better integration / information on Commons-hosted images would actually be useful. As far as I know the current system has no way to distinguish between Commons images and missing images except by downloading the Commons dump files. That can be frustrating since the Commons dumps are larger (and hence more trouble to work with) than all but a handful of other wikis.
-Robert Rohde
You only need the image.sql dump from commons to determine if the image exists there (it will also include other useful and not-so-useful data like filetype, image size, metadata...). http://download.wikimedia.org/commonswiki/20090510/
That wouldn't get you file descriptions or copyright status, etc. If your goal is something like mirroring a wiki, you really need access to page descriptions as well.
At present, the main solution is to copy all of Commons, which is overkill for many applications. It would be nice if the dump generator had a way of parsing out only the relevant Commons content.
-Robert Rohde