Dear all,
last year, we applied for a Wikimedia grant to feed qualified data from Wikipedia infoboxes (i.e. missing statements with references) via the DBpedia software into Wikidata. The evaluation was already quite good, but some parts were still missing and we would like to ask for your help and feedback for the next round. The new application is here: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSync
The main purpose of the grant is:
- Wikipedia infoboxes are quite rich, are manually curated and have references. DBpedia is already extracting that data quite well (i.e. there is no other software that does it better). However, extracting references is not a priority on our agenda. They would be very useful to Wikidata, but there are no user requests for this from DBpedia users.
- DBpedia also has all the infos of all infoboxes of all Wikipedia editions (>10k pages), so we also know quite well, where Wikidata is used already and where information is available in Wikidata or one language version and missing in another.
- side-goal: bring the Wikidata, Wikipedia and DBpedia communities closer together
Here is a diff between the old an new proposal:
- extraction of infobox references will still be a goal of the reworked proposal
- we have been working on the fusion and data comparison engine (the part of the budget that came from us) for a while now and there are first results:
6823 birthDate_gain_wiki.nt 3549 deathDate_gain_wiki.nt 362541 populationTotal_gain_wiki.nt 372913 total
We only took three properties for now and showed the gain where no Wikidata statement was available. birthDate/deathDate is already quite good. Details here: https://drive.google.com/file/d/1j5GojhzFJxLYTXerLJYz3Ih-K6UtpnG_/view?usp=s...
Our plan here is to map all Wikidata properties to the DBpedia Ontology and then have the info to compare coverage of Wikidata with all infoboxes across languages.
- we will remove the text extraction part from the old proposal (which is here for you reference: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/CrossWikiFact). This will still be a focus during our work in 2018, together with Diffbot and the new DBpedia NLP department, but we think that it distracted from the core of the proposal. Results from the Wikipedia article text extraction can be added later once they are available and discussed separately.
- We proposed to make an extra website that helps to synchronize all Wikipedias and Wikidata with DBpedia as its backend. While the external website is not an ideal solution, we are lacking alternatives. The Primary Sources Tool is mainly for importing data into Wikidata, not so much synchronization. The MediaWiki instances of the Wikipedias do not seem to have any good interfaces to provide suggestions and pinpoint missing info. Especially to this part, we would like to ask for your help and suggestions, either per mail to the list or on the talk page: https://meta.wikimedia.org/wiki/Grants_talk:Project/DBpedia/GlobalFactSync
We are looking forward to a fruitful collaboration with you and we thank you for your feedback!
All the best Magnus
Dear Wikidatans,
last year, we applied for a Wikimedia grant to feed qualified data from Wikipedia infoboxes (i.e. missing statements with references) via the DBpedia software into Wikidata. The evaluation was already quite good, but some parts were still missing. Now we apply again with a modified though similar proposal and we would like to ask for your support and feedback.
GlobalFactSync will produce a tool that identifies and visualises differences in infobox facts across available language versions, it suggests changes to Wikipedia and Wikidata editors when redacting infoboxes and Wikidata items according to data available from other Wikimedia projects. The project grant proposal can be found here: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSync
Of course, any comment is more than welcome on the discussion page. The proposal is still in a draft form (deadline January 31st), we would be very happy to hear your comments.
All the best, Magnus, Sebastian and Julia
Hi,
Our plan here is to map all Wikidata properties to the DBpedia Ontology and
then have the info to compare coverage of Wikidata with all infoboxes across languages.
This is a really exciting project that would improve both Wikidata and DBpedia. I would be interested to know more, especially on what has already been done in terms of mapping and what remains to do.
I see, for example, that DBpedia has a list of missing properties and classes https://wizyapp.appspot.com/k/u/e/5c6a68564e46200b83b303c7e7c5f246?a=aHR0cDovL21hcHBpbmdzLmRicGVkaWEub3JnL3NlcnZlci9vbnRvbG9neS93aWtpZGF0YS9taXNzaW5nLw%3D%3D, but I don't know if it's up to date.
Best regards,
Ettore Rizza
2018-01-15 19:57 GMT+01:00 Magnus Knuth knuth@informatik.uni-leipzig.de:
Dear all,
last year, we applied for a Wikimedia grant to feed qualified data from Wikipedia infoboxes (i.e. missing statements with references) via the DBpedia software into Wikidata. The evaluation was already quite good, but some parts were still missing and we would like to ask for your help and feedback for the next round. The new application is here: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSync
The main purpose of the grant is:
- Wikipedia infoboxes are quite rich, are manually curated and have
references. DBpedia is already extracting that data quite well (i.e. there is no other software that does it better). However, extracting references is not a priority on our agenda. They would be very useful to Wikidata, but there are no user requests for this from DBpedia users.
- DBpedia also has all the infos of all infoboxes of all Wikipedia
editions (>10k pages), so we also know quite well, where Wikidata is used already and where information is available in Wikidata or one language version and missing in another.
- side-goal: bring the Wikidata, Wikipedia and DBpedia communities closer
together
Here is a diff between the old an new proposal:
- extraction of infobox references will still be a goal of the reworked
proposal
- we have been working on the fusion and data comparison engine (the part
of the budget that came from us) for a while now and there are first results:
6823 birthDate_gain_wiki.nt 3549 deathDate_gain_wiki.nt
362541 populationTotal_gain_wiki.nt 372913 total
We only took three properties for now and showed the gain where no Wikidata statement was available. birthDate/deathDate is already quite good. Details here: https://drive.google.com/file/ d/1j5GojhzFJxLYTXerLJYz3Ih-K6UtpnG_/view?usp=sharing
Our plan here is to map all Wikidata properties to the DBpedia Ontology and then have the info to compare coverage of Wikidata with all infoboxes across languages.
- we will remove the text extraction part from the old proposal (which is
here for you reference: https://meta.wikimedia.org/ wiki/Grants:Project/DBpedia/CrossWikiFact). This will still be a focus during our work in 2018, together with Diffbot and the new DBpedia NLP department, but we think that it distracted from the core of the proposal. Results from the Wikipedia article text extraction can be added later once they are available and discussed separately.
- We proposed to make an extra website that helps to synchronize all
Wikipedias and Wikidata with DBpedia as its backend. While the external website is not an ideal solution, we are lacking alternatives. The Primary Sources Tool is mainly for importing data into Wikidata, not so much synchronization. The MediaWiki instances of the Wikipedias do not seem to have any good interfaces to provide suggestions and pinpoint missing info. Especially to this part, we would like to ask for your help and suggestions, either per mail to the list or on the talk page: https://meta.wikimedia.org/wiki/Grants_talk:Project/DBpedia/GlobalFactSync
We are looking forward to a fruitful collaboration with you and we thank you for your feedback!
All the best Magnus
-- Magnus Knuth
Universität Leipzig Institut für Informatik Abt. Betriebliche Informationssysteme, AKSW/KILT Augustusplatz 10 04109 Leipzig DE
mail: knuth@informatik.uni-leipzig.de tel: +49 177 3277537 webID: http://magnus.13mm.de/
Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata
Hello,
yes these classes and properties [1] are currently missing in the mappings. Though there is currently some ongoing work which is not pushed to the mappings wiki yet. Nevertheless, the more challenging part of the project will be mapping the template parameters and Wikidata properties.
Best regards Magnus
[1] http://mappings.dbpedia.org/server/ontology/wikidata/missing/
Am 05.02.2018 um 21:49 schrieb Ettore RIZZA ettorerizza@gmail.com:
Hi,
Our plan here is to map all Wikidata properties to the DBpedia Ontology and then have the info to compare coverage of Wikidata with all infoboxes across languages.
This is a really exciting project that would improve both Wikidata and DBpedia. I would be interested to know more, especially on what has already been done in terms of mapping and what remains to do.
I see, for example, that DBpedia has a list of missing properties and classes, but I don't know if it's up to date.
Best regards,
Ettore Rizza
2018-01-15 19:57 GMT+01:00 Magnus Knuth knuth@informatik.uni-leipzig.de: Dear all,
last year, we applied for a Wikimedia grant to feed qualified data from Wikipedia infoboxes (i.e. missing statements with references) via the DBpedia software into Wikidata. The evaluation was already quite good, but some parts were still missing and we would like to ask for your help and feedback for the next round. The new application is here: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSync
The main purpose of the grant is:
Wikipedia infoboxes are quite rich, are manually curated and have references. DBpedia is already extracting that data quite well (i.e. there is no other software that does it better). However, extracting references is not a priority on our agenda. They would be very useful to Wikidata, but there are no user requests for this from DBpedia users.
DBpedia also has all the infos of all infoboxes of all Wikipedia editions (>10k pages), so we also know quite well, where Wikidata is used already and where information is available in Wikidata or one language version and missing in another.
side-goal: bring the Wikidata, Wikipedia and DBpedia communities closer together
Here is a diff between the old an new proposal:
extraction of infobox references will still be a goal of the reworked proposal
we have been working on the fusion and data comparison engine (the part of the budget that came from us) for a while now and there are first results:
6823 birthDate_gain_wiki.nt 3549 deathDate_gain_wiki.nt 362541 populationTotal_gain_wiki.nt 372913 total
We only took three properties for now and showed the gain where no Wikidata statement was available. birthDate/deathDate is already quite good. Details here: https://drive.google.com/file/d/1j5GojhzFJxLYTXerLJYz3Ih-K6UtpnG_/view?usp=s...
Our plan here is to map all Wikidata properties to the DBpedia Ontology and then have the info to compare coverage of Wikidata with all infoboxes across languages.
we will remove the text extraction part from the old proposal (which is here for you reference: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/CrossWikiFact). This will still be a focus during our work in 2018, together with Diffbot and the new DBpedia NLP department, but we think that it distracted from the core of the proposal. Results from the Wikipedia article text extraction can be added later once they are available and discussed separately.
We proposed to make an extra website that helps to synchronize all Wikipedias and Wikidata with DBpedia as its backend. While the external website is not an ideal solution, we are lacking alternatives. The Primary Sources Tool is mainly for importing data into Wikidata, not so much synchronization. The MediaWiki instances of the Wikipedias do not seem to have any good interfaces to provide suggestions and pinpoint missing info. Especially to this part, we would like to ask for your help and suggestions, either per mail to the list or on the talk page: https://meta.wikimedia.org/wiki/Grants_talk:Project/DBpedia/GlobalFactSync
We are looking forward to a fruitful collaboration with you and we thank you for your feedback!
All the best Magnus