Hi Richard;

Yes, a distributed project would be probably the best solution, but it is not easy to develop, unless you use a library like bittorrent, or similar and you have many peers. Althought most of the people don't seed the files long time, so sometimes is better to depend on a few committed persons than a big but ephemeral crowd.

Regards,
emijrp

2011/6/26 Richard Farmbrough <richard@farmbrough.co.uk>
It would be useful to have  an archive of archives.  I have to delete my old data dumps as time passes, for space reasons, however a team could, between them, maintain multiple copies of every data dump. This would make a nice distributed project.

On 26/06/2011 13:53, emijrp wrote:
Hi all;

Can you imagine a day when Wikipedia is added to this list?[1]

WikiTeam have developed a script[2] to download all the Wikipedia dumps (and her sister projects) from dumps.wikimedia.org. It sorts in folders and checks md5sum. It only works on Linux (it uses wget).

You will need about 100GB to download all the 7z files.

Save our memory.

Regards,
emijrp

[1] http://en.wikipedia.org/wiki/Destruction_of_libraries
[2] http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py
_______________________________________________ Xmldatadumps-l mailing list Xmldatadumps-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l