Hey folks,
we plan to drop the wb_entity_per_page table sometime soon[0], because
it is just not required (as we will likely always have a programmatic
mapping from entity id to page title) and it does not supported non
-numeric entity ids as it is now. Due to this removing it is a blocker
for the commons metadata.
Is anybody using that for their tools (on tool labs)? If so, please
tell us so that we can give you instructions and a longer grace period
to update your scripts.
Cheers,
Marius
[0]: https://phabricator.wikimedia.org/T95685
Hi folks!
My name is Glorian Yapinus, but you can simply call me Glorian ;) . For the
next 6 months, I will assist Lydia in supporting you all.
Regarding to my educational background, I hold a bachelor's degree in
Information Technology and currently, I am working on my Master's in
Software Engineering and Management.
I am a warm and nice person. So, please do not hesitate to reach out to me
for any queries :-)
Last but not least, I am looking forward to working with you.
Cheers,
Glorian
--
Glorian Yapinus
Product Management Intern for Wikidata
Imagine a world, in which every single human being can freely share in the
sum of all knowledge. That‘s our commitment.
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 B. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hi, how about a wikipedia about objects?
Instead of generic articles of , for example, "Ballpoint pen" or "Bic
cristal" it would be "Ballpoint pen Bic cristal 2014", using Wikidata
propreties and doing these for millions of objects would allow people to
have an open, free, universal and structured place to refer specific objects
.
Project page: https://meta.wikimedia.org/wiki/WikiObject
grettings
Hi,
I had this idea for some time now but never got to test/write it down.
DBpedia extracts detailed context information in Quads (where possible) on
where each triple came from, including the line number in the wiki text.
Although each DBpedia extractor is independent, using this context there is
a small window for combining output from different extractors, such as the
infobox statements we extract from Wikipedia and the very recent citation
extractors we announced [1]
I attach a very small sample from the article about Germany where I filter
out the related triples and order them by the line number they were
extracted from e.g.
dbr:Germany dbo:populationTotal "82175700"^^xsd:nonNegativeInteger <
http://en.wikipedia.org/wiki/Germany?oldid=736355524#*absolute-line=66*
&template=Infobox_country&property=population_estimate&split=1&
wikiTextSize=10&plainTextSize=10&valueSize=8> .
<https://www.destatis.de/DE/PresseService/Presse/Pressemitteilungen/2016/08/
PD16_295_12411pdf.pdf;jsessionid=996EC2DF0A8D510CF89FDCBC74DBAE
9F.cae2?__blob=publicationFile> dbp:isCitedBy dbr:Germany <
http://en.wikipedia.org/wiki/Germany?oldid=736355524#*absolute-line=66*> .
Looking at the wikipedia article we see:
|population_estimate = 82,175,700<ref>{{cite web|url=https://www.destatis.
de/DE/PresseService/Presse/Pressemitteilungen/2016/08/PD16_295_12411pdf.pdf;
jsessionid=996EC2DF0A8D510CF89FDCBC74DBAE9F.cae2?__blob=
publicationFile|title=Population at 82.2 million at the end of 2015 –
population increase due to high immigration|date=26 August 2016|work=
destatis.de}}</ref>
Could this approach be a good candidate reference suggestions in Wikidata?
(This particular one is already a reference but the anthem and GDP in the
attachment are not for example)
There are many things that can be done to improve the matching but before
getting into details I would like to see if this idea is worth exploring
more or not
Cheers,
Dimitris
[1] http://www.mail-archive.com/dbpedia-discussion%40lists.sourceforge.net/
msg07739.html
--
Dimitris Kontokostas
Department of Computer Science, University of Leipzig & DBpedia Association
Projects: http://dbpedia.org, http://rdfunit.aksw.org,
http://aligned-project.eu
Homepage: http://aksw.org/DimitrisKontokostas
Research Group: AKSW/KILT http://aksw.org/Groups/KILT
Hello folks,
The Wikidata development team is currently working on tools to improve *list
creation on Wikipedia*, based on Wikidata data.
In order to understand what could be useful for you and why, we suggest you*
three examples of user scenarios
<https://www.wikidata.org/wiki/Wikidata:List_generation_input>*, in which
you could recognize some of your current uses: how do you currently edit
some lists on Wikipedia, which tools or processes do you use, and what can
be improved.
You can answer some short questions and add comments on our assumptions on
each related talk page. This input is very important to help us understand
how you edit the lists on Wikipedia, and what tools could be useful for you.
Thanks to all of you who will take a few minutes to answer our questions!
Jan & Léa
--
Léa Lacroix
Community Communication Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hi,
I've written a code to scrape Wikidata dump following Wikidata Toolkit
examples.
In processItemDocument, I have extracted the target entityId for the
property 'instanceof' for the current item. However I'm unable to find a
way to get the label of the target entity given that I have the entityId,
but not the entityDocument? Help would be appreciated :)
-Sumit Asthana,
B.Tech Final Year,
Dept. of CSE,
IIT Patna
Hoi,
Jura1 created a wonderful list of people who died in Brazil in 2015 [1]. It
is a page that may update regularly from Wikidata thanks to the
ListeriaBot. Obviously, there may be a few more because I am falling ever
more behind with my quest for registering deaths in 2015.
I have copied his work and created a page for people who died in the
Netherlands in 2015 [2]. It is trivially easy to do this and, the result is
great. The result looks great, it can be used for any country in any
Wikipedia
The Dutch Wikipedia indicated that they nowadays maintain important
metadata at Wikidata. I am really happy that we can showcase their work. It
is important work because as someone reminded me at some stage, this is
part of what amounts to the policy of living people...
Thanks,
GerardM
[1] https://www.wikidata.org/wiki/User:Jura1/Recent_deaths_in_Brazil
[2]
https://www.wikidata.org/wiki/User:Jura1/Recent_deaths_in_the_Netherlands
Everyone,
The submission deadline for the below is soon, but the bar is low - a
paragraph or two describing what you're up to, and any challenges that
you're facing. Our main goal is to get a sense of who's doing what in this
space, and to discuss prospects for helping each other.
If you'd like to be involved in the conversation, but won't be able to
make it to Costa Rica, please let me know.
Best,
Joel.
----
The Use of Wikis in Biodiversity Informatics
Workshop to be held at TDWG 2016 (Dec. 5-9) in Santa Clara de San Carlos, Costa Rica
Instruction for submitting abstracts are at
https://mbgserv18.mobot.org/ocs/index.php/tdwg/tdwg2016/schedConf/cfp
Abstracts due: Sept. 6, 2016
Wiki technologies, in particular those built on top of Semantic MediaWiki
and Wikidata, are being used to store, curate, query, integrate, and
reason over a range of biodiversity-related data. This workshop will
comprise a selection of talks describing some of these uses, followed by a
discussion of gaps in the wiki ecology, and opportunities that might exist
to fill those gaps through coordinated research and development.
I know it's been mentioned on this list before, but it would be
incredibly useful to have incremental dumps of Wikidata, as downloading
the current dumps can now take several hours over a poor-bandwidth
Internet connection.
Here's my proposal:
* the incremental dumps should have exactly the same format as the
current JSON dumps, with two exceptions:
** entries which are unchanged since the previous dump (as determined by
their "modified" timestamp) should be omitted
** entries which have been deleted since the previous dump should have
stub entries of the form {"id": "Q123", "deleted": true}
I would imagine that these dumps would be vastly smaller than the
standard dumps, and would, for many re-users who only want to know about
changed data, be just as useful, with a fraction of the download time
and in many cases without significant modification of any of their
tools. Doing this would only need a small amount of processing time, and
add only an insignificant amount to the disk storage needed on the
servers, yet could save considerable amounts of Internet bandwidth.
This difference-file format should be easy to generate using slight
tweaks to the existing dump code, but, if needed, I can easily write a
simple Python script to take two existing dump files and generate the
differences between them in the format above. Please drop me an email,
or reply here, if you would like me to write this.
Kind regards,
Neil