[Foundation-l] [Xmldatadumps-l] Wikipedia dumps downloader

emijrp emijrp at gmail.com
Mon Jun 27 11:07:51 UTC 2011


Hi Richard;

Yes, a distributed project would be probably the best solution, but it is
not easy to develop, unless you use a library like bittorrent, or similar
and you have many peers. Althought most of the people don't seed the files
long time, so sometimes is better to depend on a few committed persons than
a big but ephemeral crowd.

Regards,
emijrp

2011/6/26 Richard Farmbrough <richard at farmbrough.co.uk>

> **
> It would be useful to have  an archive of archives.  I have to delete my
> old data dumps as time passes, for space reasons, however a team could,
> between them, maintain multiple copies of every data dump. This would make a
> nice distributed project.
>
> On 26/06/2011 13:53, emijrp wrote:
>
> Hi all;
>
> Can you imagine a day when Wikipedia is added to this list?[1]
>
> WikiTeam have developed a script[2] to download all the Wikipedia dumps
> (and her sister projects) from dumps.wikimedia.org. It sorts in folders
> and checks md5sum. It only works on Linux (it uses wget).
>
> You will need about 100GB to download all the 7z files.
>
> Save our memory.
>
> Regards,
> emijrp
>
> [1] http://en.wikipedia.org/wiki/Destruction_of_libraries
> [2]
> http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py
>
>
> _______________________________________________
> Xmldatadumps-l mailing listXmldatadumps-l at lists.wikimedia.orghttps://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l
>
>
>



More information about the wikimedia-l mailing list