Hello, and my first post here :)
For medium-size queries which timeout in a SPARQL endpoint, you can add
these clauses and it sometimes gives a result:
ORDER BY RAND()
LIMIT 100
You will only obtain some results with no specific order, but it is
better than nothing. It doesn’t work everytime, but it works more often
than without it or with only a limit. Given it is a hacky technique, it
should be the last thing to try, after other query optimisations.
For instance, when I search all elements with a GND ID (P227) without
P31 neither P279, I get a timeout even when I only add a limit. When I
add these two clauses, I get a result.
Cheers,
~ Seb35
With rand:
https://query.wikidata.org/#SELECT%20DISTINCT%20%3Fitem%20%3FitemLabel%20%3…
Without rand:
https://query.wikidata.org/#SELECT%20DISTINCT%20%3Fitem%20%3FitemLabel%20%3…
Oh boy. I thought I had a few things figured out with Wikidata...until
I read through the Property Talk discussions for
https://www.wikidata.org/wiki/Property_talk:P2236
(with its mentions of Freebase mapping)
So...
I've been adding a few bits of Schema.org mapping into Wikidata today, and
stumbled upon a few things that made me rethink a few things...lolol.
QUESTION:
How to state that a Wikidata Entity (not a Property) such as
place of birth https://www.wikidata.org/wiki/Q1322263
is the same concept or idea as a Schema.org property
http://schema.org/birthPlace ?
I thought I could use P2236 above... but then it seems its for WD
Properties, not Entities (subjects) ?
SOLUTION ? Perhaps we could do a best practice of treating
http://schema.org/birthPlace as an actual Identifier for the place of birth
concept
https://www.wikidata.org/wiki/Q1322263
...while reserving the WD Property place of brith
https://www.wikidata.org/wiki/Property:P19 to use equivalent property
https://www.wikidata.org/wiki/Property:P1628 ?
TPT and Denny didn't leave enough notes in there for what to do about
external mapping cases of external vocabularies that are also loosely
considered as metadata dictionaries as well for the common web and
developers, like Schema.org is.
Thoughts on the SOLUTION proposed ?
Thad
+ThadGuidry <https://www.google.com/+ThadGuidry>
Hey folks :)
Amir and other have worked hard over the past months to bring ORES to
Wikidata. The goal is to use machine learning to make it easier to spot
potentially bad edits. ORES is now available as a beta feature on Wikidata.
Once you have enabled it you can see some edits in recent changes and
watchlist will show up in a different color or have a little r in front of
them. These edits are judged as potentially bad and should probably get
more review. In your preferences you can adjust how harsh ORES should
judge. You can also filter your watchlist/recent changes to only show
potentially bad edits. Patrolled edits won't be shown as potentially bad.
This should be a huge step towards making it easier to find and fight
vandalism on Wikidata.
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hey folks :)
Jonas has written a user script to show an image on an item. The goal is to
make it easier to see what the item is about and also spot potential
vandalism/data quality issues. It'd be great if you could give it a try to
see if this is something we should explore further. More details at
https://www.wikidata.org/wiki/Wikidata:Project_chat#Header_image_on_items
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hey :)
Amir has written a very nice little user script that adds a little icon
next to a statement. Clicking on it leads you to a query for items with the
same statement. So if you are on an item about a cat you can click it on
the "instance of" statement and then find all other cats. More details
here:
https://www.wikidata.org/wiki/Wikidata:Project_chat#Do_you_want_to_see_stat…
Please
also leave feedback there. If you like it it can be turned into a gadget
later. (I don't think it is a good idea to put it in the default UI at this
point.)
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
[Feel free to blame me if you read this more than once]
To whom it may interest,
Full of delight, I would like to announce the first beta release of
*StrepHit*:
https://github.com/Wikidata/StrepHit
TL;DR: StrepHit is an intelligent reading agent that understands text
and translates it into *referenced* Wikidata statements.
It is a IEG project funded by the Wikimedia Foundation.
Key features:
-Web spiders to harvest a collection of documents (corpus) from reliable
sources
-automatic corpus analysis to understand the most meaningful verbs
-sentences and semi-structured data extraction
-train a machine learning classifier via crowdsourcing
-*supervised and rule-based fact extraction from text*
-Natural Language Processing utilities
-parallel processing
You can find all the details here:
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
If you like it, star it on GitHub!
Best,
Marco
I'm confused by this from today's Wikidata weekly summary:
- New request for comments: Semi-automatic Addition of References to
Wikidata Statements - feedback on the Primary Sources Tool
<https://www.wikidata.org/wiki/Wikidata:Requests_for_comment/Semi-automatic_…>
First of all, the title makes no sense because "semi-automatic addition of
references to Wikidata statements" is one of the main things that the tool
can't currently do. You'll almost always end up with duplicate statements
if there's an existing statement, rather than the desired behavior of just
adding the statement.
Second, I'm not sure who "Hjfocs" is (why does everyone have to make up
fake wikinames?), but why are they asking for more feedback when there's
been *ample* feedback already? There hasn't been an issue with getting
people to test the tool or provide feedback based on the testing. The issue
has been with getting anyone to *act* on the feedback. Everything is a)
"too hard," or b) "beyond our resources," or depends on something in
category a or b, or is incompatible with the arbitrary implementation
scheme chosen, or some other excuse.
We're 12-18+ months into the project, depending on how you measure, and not
only is the tool not usable yet, but it's no longer improving, so I think
it's time to take a step back and ask some fundamental questions.
- Is the current data pipeline and front end gadget the right approach and
the right technology for this task? Can they be fixed to be suitable for
users?
- If so, should Google continue to have sole responsibility for it or
should it be transferred to the Wikidata team or someone else who'll
actually work on it?
- If not, what should the data pipeline and tooling look like to make
maximum use of the Freebase data?
The whole project needs a reboot.
Tom