Sorry, I forgot to mention that I have in mind the English wikipedia dump.
- sam -
wiki writes:
> Hello.
>
> I'm a newbie who wants to start playing with the xml dumps. I've found
> instructions here and there on how to import these. I'd like to seek
> guidance though as to how much free disk space one is required to have for
> the MySql import to succeed? i.e. after I have already installed LAMP +
> Mediawiki, and already allocated space for the bzip file and the converted
> import statements file, roughly how much more space is needed?
>
> Thank you!
>
> - sam -
>
> _______________________________________________
> Xmldatadumps-l mailing list
> Xmldatadumps-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l
Hello.
I'm a newbie who wants to start playing with the xml dumps. I've found
instructions here and there on how to import these. I'd like to seek
guidance though as to how much free disk space one is required to have for
the MySql import to succeed? i.e. after I have already installed LAMP +
Mediawiki, and already allocated space for the bzip file and the converted
import statements file, roughly how much more space is needed?
Thank you!
- sam -
We try to build a snapshot of wikidata.org. First we want to get a
snapshot of wikidata.org to play with.
Thus we're depending on its dump now.
I have checked the sitehttp://dumps.wikimedia.org/wikidatawiki/.
<http://dumps.wikimedia.org/wikidatawiki/> The latest wikidata dump is
for 20130228.
But it failed.
I'm wondering if it is possible to fix this failure, or we have to
wait for the next dump.
*From http://meta.wikimedia.org/wiki/Data_dumps, it mentions:*
Failures in the dump process are generally dealt with by rerunning the
portion of the dump that failed.
Isn't it the case for that wikidata failure? Will someone take care of it?
Thank all people to spend time to make dump as a great
data source :-)
François Bonzon, 04/03/2013 18:22:
> I confirm I now see interwiki language links originating from Wikidata
> in <language>wiki-<date>-langlinks.sql.gz dumps, with the format
> described in the 2nd link you sent. However, this is a MySQL dump, not a
> XML dump.
>
> Language links are then no more available in XML data dumps?
I guess not, except – probably – in the XML data dumps for Wikidata
itself, in whatever weird format ContentHandler makes them into.
Nemo
Hi,
I understand from http://www.wikidata.org/wiki/Wikidata:News that
- enwiki since February 13, 2013
- hewiki and itwiki since January 30, 2013
- huwiki January 14, 2013
have migrated to the Wikidata project. And more wikis will follow shortly.
One consequence is that wiki markup for interwiki links (cross-language
links) are being gradually removed from articles, because the MediaWiki
software can now read them from the centralized Wikidata repository.
I verified in the latest huwiki dump that some articles indeed no more have
interwiki links. Do you confirm my above statements?
How can I now extract interwiki links from dumps? Is there a separate
Wikidata dump I should download? What attributes for look for to join
Wikidata and separate language wiki dumps? Thanks for your help.
-François