Maarten Dammers, 14/07/19 15:04:
Several, I
think. The most significant I remember were from Sweden and
Finland.
Any pointers?
Maybe
http://www.ksamsok.se/in-english/ , but I know more about
http://data.nationallibrary.fi/bib/sparql and related.
Semi-open? According to
https://www.wikidata.org/wiki/Wikidata:SPARQL_federation_input/Archive#dati…
this is cc-by 2.5. I assume they did a separate release?
They changed the license a few times, I think. When we imported it, it
was CC-BY-3.0-it which is fine because it waives the sui generis rights
completely.
Anyway, their platform (Lodview) is quite nice. We should also add links
to things like
http://dati.beniculturali.it/iccd/schede/resource/GeographicalFeature/Comun…
I guess it wouldn't harm. Matching municipalities is often a major pain.
The amount of "open data" which is released with usable references to
municipalities is negligible, usually you end up manually matching names
or codes in free text form in some CSV.
Importing the entire of the ontology itself can be trickier. More work
on this side has been done by ICCU and ICCD (the ministry): usually it
takes them a few years of manual work to connect an ontology.
For our import it was more important to handle the objects which had
very little (structured) information. The more detailed descriptions are
usually sparsely used (in this case you linked, only by one province
which was cataloguing first world war damages? I don't know).
With
the federation in place, it's possible to set up automated reports to
find mismatches between the data. See for example the report on
https://www.wikidata.org/wiki/Property_talk:P1006/Mismatches . Obvious
report for this domain would be monuments in the beniculturali database,
but not on Wikidata. Or do you already have something in place?
We used the SPARQL queries listed at the end of the page:
<https://www.wikidata.org/?curid=30576438#Reports_for_cleanup_and_data_improvement>.
I don't remember if federated queries were fast enough at the time to be
usable, I only remember using them for small subsets of the data.
As far as I can see the bot is coded to get all the data at once and
then see what needs doing. It doesn't attempt to get incremental updates
with federated queries.
<https://github.com/synapta/wikidata-mibact-luoghi-cultura/blob/master/bot-mibact-to-wikidata/queries.js>
Federico