[Foundation-l] [Wiki-research-l] Wikipedia dumps downloader
emijrp
emijrp at gmail.com
Mon Jun 27 11:10:40 UTC 2011
Hi SJ;
You know that that is an old item in our TODO list ; )
I heard that Platonides developed a script for that task long time ago.
Platonides, are you there?
Regards,
emijrp
2011/6/27 Samuel Klein <sjklein at hcs.harvard.edu>
> Thank you, Emijrp!
>
> What about the dump of Commons images? [for those with 10TB to spare]
>
> SJ
>
> On Sun, Jun 26, 2011 at 8:53 AM, emijrp <emijrp at gmail.com> wrote:
> > Hi all;
> >
> > Can you imagine a day when Wikipedia is added to this list?[1]
> >
> > WikiTeam have developed a script[2] to download all the Wikipedia dumps
> (and
> > her sister projects) from dumps.wikimedia.org. It sorts in folders and
> > checks md5sum. It only works on Linux (it uses wget).
> >
> > You will need about 100GB to download all the 7z files.
> >
> > Save our memory.
> >
> > Regards,
> > emijrp
> >
> > [1] http://en.wikipedia.org/wiki/Destruction_of_libraries
> > [2]
> >
> http://code.google.com/p/wikiteam/source/browse/trunk/wikipediadownloader.py
> >
> > _______________________________________________
> > Wiki-research-l mailing list
> > Wiki-research-l at lists.wikimedia.org
> > https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
> >
> >
>
>
>
> --
> Samuel Klein identi.ca:sj w:user:sj +1 617 529
> 4266
>
> _______________________________________________
> Wiki-research-l mailing list
> Wiki-research-l at lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
>
More information about the wikimedia-l
mailing list