Just like the scripts to preserve wikis, I'm working in a new script to
download all Wikimedia Commons images packed by day. But I have limited
spare time. Sad that volunteers have to do this without any help from
I started too an effort in meta: (with low activity) to mirror XML dumps.
If you know about universities or research groups which works with
Wiki[pm]edia XML dumps, they would be a possible successful target to mirror
If you want to download the texts into your PC, you only need 100GB free and
to run this Python script.
I heard that Internet Archive saves XML dumps quarterly or so, but no
official announcement. Also, I heard about Library of Congress wanting to
mirror the dumps, but not news since a long time.
L'Encyclopédie has an "uptime" of 260 years and growing. Will
Wiki[pm]edia projects reach that?
2011/6/2 Fae <faenwp(a)gmail.com>
I'm taking part in an images discussion workshop with a number of
academics tomorrow and could do with a statement about the WMF's long
term commitment to supporting Wikimedia Commons (and other projects)
in terms of the public availability of media. Is there an official
published policy I can point to that includes, say, a 10 year or 100
If it exists, this would be a key factor for researchers choosing
where to share their images with the public.
Guide to email tags: http://j.mp/faetags
foundation-l mailing list