I can run a database report Monday. But keep in mind that the wiki isn't static and what you want changes on a very rapid rate


On Friday, June 19, 2015, D. Hansen <sammelaccount@tageskurier.de> wrote:
Hi!
I have tried to get a list of all .svg-files on commons.wikipedia.

Of course I could just parse through commons, but
* if there would be any way to provide a dump with the names of the really existing .svg-files, that would be a tremendeous help for me, *
and in my estimation it would reduce the download  size and cpu-burden and most importantly the HD-burden about 70% to 80% compared to browse and parse through commons. (Wheres the cpu-usage and heat problems because of the HD-burden on my notebook would be much more adverse than the burden on wikipedias server, I assume. I lost already two HDs over the years when downloading larger amounts of files in one go.)

Though I have asked at various places so far I haven't found a good solution.
One suggestion was to downloadcommonswiki-20150417-all-titles, which I did.

But this file does contain deleted names and renamed names, and names the partly have "File:" and some that don't have "File:" or a similiar indicator at the start.
Doing just a small sample resulted in 5 correct names, and around 7 deleted and 7 renamed names. I have asked at various places, and especially one person tried to help me, but this even he couldn't solve. Beside this I didn't much feedback.

Is it possible to get such a dump? Or to get another dump that I could use to update and crosscheck the all-titles file?

Greetings
D. Hansen

_______________________________________________
Xmldatadumps-l mailing list
Xmldatadumps-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l