I would appreciate clarification what is proposed with regard to exposing problematic Wikidata ontology on Wikipedia. If the idea involves inserting poor-quality information onto English Wikipedia in order to spur us to fix problems with Wikidata, then I am likely to oppose it. English Wikipedia is not an endless resource for free labor, and we have too few skilled and good-faith volunteers to handle our already enormous scope of work.
Pine
( https://meta.wikimedia.org/wiki/User:Pine )
null
This recent announcement by the Structured Data team perhaps ought to be
quite a heads-up for us:
https://commons.wikimedia.org/wiki/Commons_talk:Structured_data#Searching_C…
Essentially the team has given up on the hope of using Wikidata
hierarchies to suggest generalised "depicts" values to store for images
on Commons, to match against terms in incoming search requests.
i.e. if an image is of a German Shepherd dog, and identified as such,
the team has given up on trying to infer in general from Wikidata that
'dog' is also a search term that such an image should score positively with.
Apparently the Wikidata hierarchies were simply too complicated, too
unpredictable, and too arbitrary and inconsistent in their design across
different subject areas to be readily assimilated (before one even
starts on the density of bugs and glitches that then undermine them).
Instead, if that image ought to be considered in a search for 'dog', it
looks as though an explicit 'depicts:dog' statement may be going to be
needed to be specifically present, in addition to 'depicts:German Shepherd'.
Some of the background behind this assessment can be read in
https://phabricator.wikimedia.org/T199119
in particular the first substantive comment on that ticket, by Cparle on
10 July, giving his quick initial read of some of the issues using
Wikidata would face.
SDC was considered a flagship end-application for Wikidata. If the data
in Wikidata is not usable enough to supply the dogfood that project was
expected to be going to be relying on, that should be a serious wake-up
call, a red flag we should not ignore.
If the way data is organised across different subjects is currently too
inconsistent and confusing to be usable by our own SDC project, are
there actions we can take to address that? Are there design principles
to be chosen that then need to be applied consistently? Is this
something the community can do, or is some more active direction going
to need to be applied?
Wikidata's 'ontology' has grown haphazardly, with little oversight, like
an untended bank of weeds. Is some more active gardening now required?
-- James.
---
This email has been checked for viruses by AVG.
https://www.avg.com
Sharing a little query example that I just created when looking for
interesting cases in the data. In case you are asked "what's your
relationship with your spouse" here is a list of most common answers:
http://tinyurl.com/ycm8fllo
Possibly useful and/or entertaining for other properties as well.
Finding all cases where a particular double relationship holds should be
an easy exercise.
Cheers,
Markus
--
Prof. Dr. Markus Kroetzsch
Knowledge-Based Systems Group
Center for Advancing Electronics Dresden (cfaed)
Faculty of Computer Science
TU Dresden
+49 351 463 38486
https://kbs.inf.tu-dresden.de/
Hello all,
As you may know, ORES is a tool analyzing edits to detect vandalism,
providing a score per edit. You can see the result on Recent Changes, you
can also let us know when you find something wrong
<https://www.wikidata.org/wiki/Wikidata:ORES/Report_mistakes/2018>.
But do you know that you can also directly help ORES to improve? We just
launched a new labeling campaign
<https://labels.wmflabs.org/ui/wikidatawiki/>: after authorizing your
account with OAuth, you will see some real edits, and you will be asked if
you find them damaging or not, good faith or bad faith. Completing a set
will take you around 10 minutes.
The last time we run this campaign was in 2015. Since then, the way of
editing Wikidata changed, some vandalism patterns as well (for example,
there are more vandalism on companies). So, if you're familiar with the
Wikidata rules and you would be willing to give a bit of time to help
fighting against vandalism, please participate
<https://labels.wmflabs.org/ui/wikidatawiki/> :)
If you encounter any problem or have question about the tool, feel free to
contact Ladsgroup <https://www.wikidata.org/wiki/User:Ladsgroup>.
Cheers,
--
Léa Lacroix
Project Manager Community Communication for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hello all,
As previously announced, the next big piece of Lexicographical Data on
Wikidata <https://www.wikidata.org/wiki/Wikidata:Lexicographical_data> is
now deployed: Senses.
Senses will allow you to describe, for each Lexeme, the different meanings
of the word. By using multilingual glosses, very short phrase giving an
idea of the meaning. In addition, each of these Senses can have statements
to indicate synonyms, antonyms, refers-to-concept and more. By connecting
Senses to other Senses and to Items, you will be able to describe precisely
the meaning of words with structured and linked data. But the most
important thing is that Senses will be able to do is collect translations
of words between languages.
Thanks to Senses, you will be able to organize and connect the existing
Lexemes better, and to provide a very important layer of information. With
Senses support, we now have all the basic technical building blocks to
allow structured machine-readable lexicographical data, that can be
reusable within and Wikimedia projects and by other stakeholders.
Feel free to try editing Senses. You can use the sandbox
<https://www.wikidata.org/wiki/Lexeme:L123> to make some tests. Let us know
if you have questions or find bugs.
Note: there are still issues with sorting the IDs of Senses, Forms and
sorting the glosses, that will be solved later this week. Thanks for your
understanding.
Cheers,
--
Léa Lacroix
Project Manager Community Communication for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hello all,
A few months ago, we enabled suggestions
<https://www.wikidata.org/wiki/Wikidata:Project_chat/Archive/2018/08#Better_…>
based on constraints values for the constraints section of a property. We
would like to explore further this possibility of having better suggestions
for entities, that’s why we created a beta feature.
If you’re going to your Preferences and check the beta features list,
you’ll be able to enable "Entity suggestions from constraint definitions".
At the moment there are two constraint types that are used for generating
suggestions:
- One-of constraint
<https://www.wikidata.org/wiki/Help:Property_constraints_portal/One_of>
used for suggestions in the context of a main value.
- Allowed qualifier constraint
<https://www.wikidata.org/wiki/Help:Property_constraints_portal/Qualifiers>
used for suggestions in the context of qualifiers.
You can learn more on this page
<https://www.wikidata.org/wiki/Help:Property_constraints_portal/Entity_sugge…>.
Feel free to try it, and let us know if you find it relevant and useful.
If you find any issue, feel free to report it in this ticket
<https://phabricator.wikimedia.org/T202712>.
--
Léa Lacroix
Project Manager Community Communication for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Dear all,
I am happy to report that we have just won the Best Paper Award of the
In-Use track of this year's International Semantic Web Conference
(ISWC), for our description of the SPARQL/RDF technology use on Wikidata
[1]. I keep telling people here that the general awesomeness of Wikidata
is the work of many, and in particular of this great community of editors.
Overall, the year's ISWC here in Monterey, CA has surprised Denny and me
with the huge uptake that Wikidata gets by now in industry and academia
alike, which was a huge breakthrough over last year. An amazing array of
people are doing great work based on this data, and again I would like
to pass on all the thank you's I have heard over this week to all of you
working hard to make this happen. Users range from individual students
to major tech companies, and I hope there will be many contributions
flowing back to us through these stakeholders. We also have seen an
increasing amount of research being done using Wikidata for evaluation
and testing, again both in published works and in conversations with
people in small and big organisations.
Let me also congratulate Fariz Darari, who is not a stranger to this
list either, on receiving the Best Dissertation Award of the Semantic
Web Science Association for the research that is also behind some of the
tools he has been creating for Wikidata.
I gave another talk related to Wikidata's ontological modeling, which I
hope did not represent the situation all too wrongly [2] ;-). There will
be videos of this and the best paper presentation, and most other ISWC
talks on VideoLectures in the not-so-far future.
Finally, since our best paper is about the use of BlazeGraph as a
platform for queries, let me also mention that we have had a number of
productive meetings here to discuss the future of this great software
(you may know that there were some organisational changes to the team
developing this so far). There will be opportunity to contribute to this
open source project, either as a developer or in other ways, in the
future. Stay tuned for more information on this.
So, thanks again to everyone working towards the great success of this
project -- amazing work!
Greetings from Asilomar
Markus
[1] Stanislav Malyshev, Markus Krötzsch, Larry González, Julius Gonsior,
Adrian Bielefeldt: "Getting the Most out of Wikidata: Semantic
Technology Usage in Wikipedia’s Knowledge Graph"
Talk slides and paper online:
https://iccl.inf.tu-dresden.de/web/Inproceedings3044/en
[2] Markus Krötzsch
Ontological Modelling in Wikidata
Invited keynote at the 9th Workshop on Ontology Design and Patterns (WOP'18)
Slides: https://iccl.inf.tu-dresden.de/web/Misc3058/en
--
Prof. Dr. Markus Kroetzsch
Knowledge-Based Systems Group
Center for Advancing Electronics Dresden (cfaed)
Faculty of Computer Science
TU Dresden
+49 351 463 38486
https://kbs.inf.tu-dresden.de/
Hello -
I've got a very basic wikibase up and running, using this tutorial:
https://medium.com/@thisismattmiller/wikibase-for-research-infrastructure-p…
The tutorial along with the docker images make getting up and running
pretty easy - So I've got a small base loaded - 7 properties, a couple of
"core items", and then about 2000 jazz musicians. All seems mostly fine - I
can query it via SPARQL, edit items in wikibase, and create pages in the
wiki. The sample code is a touch out of date because the wikidata client
library used for the examples has evolved a bit, but it wasn't too hard to
figure out.
However, I can't seem to create a link from the item to a page in my wiki -
I don't appear to be have a site that I can link against. When I try to use
this page:
http://localhost:8181/wiki/Special:SetSiteLink
whatever i try, it tells me that my site ID is unknown, and that I should
try a known site like 'enwiki' (which also fails). I've tried just about
every variation of siteID that I can think of.
Trying to add a site link in the wikidata item page for an item fails too -
when I try to add a link, I'm never presented with a list of sites to
potentially link to in the 'Wikipedia' box (and in the
wikinews/wikivoyage/etc sitelink boxes)
The wikibase installation page at
https://www.mediawiki.org/wiki/Wikibase/Installation#There_are_no_sites_I_c…
suggests running these commands:
php lib/maintenance/populateSitesTable.php
php repo/maintenance/rebuildItemsPerSite.php
So I docker exec'ed into the wikibase container and ran them. Now, I'm
presented with site link options - but all to wikipedia.org sites, not my
local site.
I've tried inserting a row into the 'sites' table in MySQL to something
that points to the local site, but still no dice.
Can someone point me at what I'm missing?
Related: the docker-compose system that comes up also brings up a regular
mediawiki site, and it seemingly has the wikibase-client extension running,
because I can reference my local wikibase data and pull out properties for
items - for example, I can make a page for 'Teddy Edwards', and use this
syntax and get a result:
{{#statements:P7|from=Q1932}}
P7 in the tutorial is the URL for an image from the linkedjazz site. (Teddy
Edwards is assigned Q1932 in the tutorial I'm following)
However, because I can't make a sitelink from
http://localhost:8181/wiki/Item:Q1932 to
http://localhost:8181/wiki/Teddy_Edwards.
And (I think) because of the lack of that sitelink, the wikibase client
only works with the |from syntax - if I put in my Teddy_Edwards article
just a {{#statements:P7}} line, I get no results when it's rendered.
Am I correct that this second problem is tied to my first problem?
Thanks in advance for any suggestions,
-Erik