Hi Richard;
Yes, a distributed project would be probably the best solution, but it is
not easy to develop, unless you use a library like bittorrent, or similar
and you have many peers. Althought most of the people don't seed the files
long time, so sometimes is better to depend on a few committed persons than
a big but ephemeral crowd.
Regards,
emijrp
2011/6/26 Richard Farmbrough <richard(a)farmbrough.co.uk>
**
It would be useful to have an archive of archives. I have to delete my
old data dumps as time passes, for space reasons, however a team could,
between them, maintain multiple copies of every data dump. This would make a
nice distributed project.
On 26/06/2011 13:53, emijrp wrote:
Hi all;
Can you imagine a day when Wikipedia is added to this list?[1]
WikiTeam have developed a script[2] to download all the Wikipedia dumps
(and her sister projects) from
dumps.wikimedia.org. It sorts in folders
and checks md5sum. It only works on Linux (it uses wget).
You will need about 100GB to download all the 7z files.
Save our memory.
Regards,
emijrp
[1]
http://en.wikipedia.org/wiki/Destruction_of_libraries
[2]
http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py
_______________________________________________
Xmldatadumps-l mailing
listXmldatadumps-l@lists.wikimedia.orghttps://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l