[Foundation-l] [Xmldatadumps-l] Wikipedia dumps downloader

emijrp emijrp at gmail.com
Mon Jun 27 11:07:51 UTC 2011

Hi Richard;

Yes, a distributed project would be probably the best solution, but it is
not easy to develop, unless you use a library like bittorrent, or similar
and you have many peers. Althought most of the people don't seed the files
long time, so sometimes is better to depend on a few committed persons than
a big but ephemeral crowd.


2011/6/26 Richard Farmbrough <richard at farmbrough.co.uk>

> **
> It would be useful to have  an archive of archives.  I have to delete my
> old data dumps as time passes, for space reasons, however a team could,
> between them, maintain multiple copies of every data dump. This would make a
> nice distributed project.
> On 26/06/2011 13:53, emijrp wrote:
> Hi all;
> Can you imagine a day when Wikipedia is added to this list?[1]
> WikiTeam have developed a script[2] to download all the Wikipedia dumps
> (and her sister projects) from dumps.wikimedia.org. It sorts in folders
> and checks md5sum. It only works on Linux (it uses wget).
> You will need about 100GB to download all the 7z files.
> Save our memory.
> Regards,
> emijrp
> [1] http://en.wikipedia.org/wiki/Destruction_of_libraries
> [2]
> http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py
> _______________________________________________
> Xmldatadumps-l mailing listXmldatadumps-l at lists.wikimedia.orghttps://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l

More information about the foundation-l mailing list