Maarten Dammers, 14/07/19 15:04:
Several, I think. The most significant I remember were from Sweden and Finland.
Any pointers?
Maybe http://www.ksamsok.se/in-english/ , but I know more about http://data.nationallibrary.fi/bib/sparql and related.
Semi-open? According to https://www.wikidata.org/wiki/Wikidata:SPARQL_federation_input/Archive#dati.... this is cc-by 2.5. I assume they did a separate release?
They changed the license a few times, I think. When we imported it, it was CC-BY-3.0-it which is fine because it waives the sui generis rights completely.
Anyway, their platform (Lodview) is quite nice. We should also add links to things like http://dati.beniculturali.it/iccd/schede/resource/GeographicalFeature/Comune...
I guess it wouldn't harm. Matching municipalities is often a major pain. The amount of "open data" which is released with usable references to municipalities is negligible, usually you end up manually matching names or codes in free text form in some CSV.
and http://dati.beniculturali.it/iccd/schede/resource/uod/S010537 .
Importing the entire of the ontology itself can be trickier. More work on this side has been done by ICCU and ICCD (the ministry): usually it takes them a few years of manual work to connect an ontology.
For our import it was more important to handle the objects which had very little (structured) information. The more detailed descriptions are usually sparsely used (in this case you linked, only by one province which was cataloguing first world war damages? I don't know).
With the federation in place, it's possible to set up automated reports to find mismatches between the data. See for example the report on https://www.wikidata.org/wiki/Property_talk:P1006/Mismatches . Obvious report for this domain would be monuments in the beniculturali database, but not on Wikidata. Or do you already have something in place?
We used the SPARQL queries listed at the end of the page: https://www.wikidata.org/?curid=30576438#Reports_for_cleanup_and_data_improvement. I don't remember if federated queries were fast enough at the time to be usable, I only remember using them for small subsets of the data.
As far as I can see the bot is coded to get all the data at once and then see what needs doing. It doesn't attempt to get incremental updates with federated queries. https://github.com/synapta/wikidata-mibact-luoghi-cultura/blob/master/bot-mibact-to-wikidata/queries.js
Federico