Hi all -
I note that there is no RDF dump at https://dumps.wikimedia.org/wikidatawiki/entities/20160606/ this week, just the json version. Does anyone know why?
Thanks,
On Wed, Jun 8, 2016 at 1:20 PM, Eric Scott eric.d.scott@att.net wrote:
Hi all -
I note that there is no RDF dump at https://dumps.wikimedia.org/wikidatawiki/entities/20160606/ this week, just the json version. Does anyone know why?
The dump script had a database connection error on one of the shards. To prevent incomplete dumps, I think we have the script abort.
I think we can restart the rdf dump manually for this week, and look into making the script more resilient to this sort of error.
Cheers, Katie
Thanks,
Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata
On Wed, Jun 8, 2016 at 6:39 PM, Katie Filbert katie.filbert@wikimedia.de wrote:
On Wed, Jun 8, 2016 at 1:20 PM, Eric Scott eric.d.scott@att.net wrote:
Hi all -
I note that there is no RDF dump at https://dumps.wikimedia.org/wikidatawiki/entities/20160606/ this week, just the json version. Does anyone know why?
The dump script had a database connection error on one of the shards. To prevent incomplete dumps, I think we have the script abort.
I think we can restart the rdf dump manually for this week, and look into making the script more resilient to this sort of error.
We regenerated the rdf dump:
https://dumps.wikimedia.org/other/wikibase/wikidatawiki/20160608/
Cheers, Katie
Cheers, Katie
Thanks,
Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata
-- Katie Filbert Wikidata Developer
Wikimedia Germany e.V. | Tempelhofer Ufer 23-24, 10963 Berlin Phone (030) 219 158 26-0
Wikimedia Germany - Society for the Promotion of free knowledge eV Entered in the register of Amtsgericht Berlin-Charlottenburg under the number 23 855 as recognized as charitable by the Inland Revenue for corporations I Berlin, tax number 27/681/51985.
Would it also be interesting or possible, not to do a plain ttl dump but store the blazegraph indexed datafile? Such that it does not need to be loaded again. This will save a lot of time on loading...
Op wo 8 jun. 2016 om 23:29 schreef Eric Scott eric.d.scott@att.net:
Hi all -
I note that there is no RDF dump at https://dumps.wikimedia.org/wikidatawiki/entities/20160606/ this week, just the json version. Does anyone know why?
Thanks,
Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata
On 09.06.2016 07:05, Jasper Koehorst wrote:
Would it also be interesting or possible, not to do a plain ttl dump but store the blazegraph indexed datafile? Such that it does not need to be loaded again. This will save a lot of time on loading...
A ttl dump is much more useful for external users than a BlazeGraph-specific file. For the internal use (WDQS), dumps are created in a special way already, so they do not depend on the external ttl files. Moreover, they mostly update from the live edits rather than re-importing a full dump once a week.
Markus