Hi,
TL;DR: Did anybody consider using Wikidata items of Wikipedia templates to
store multilingual template parameters mapping?
Full explanation:
As in many other projects in the Wikimedia world, templates are one of the
biggest challenges in developing the ContentTranslation extension.
Translating a template between languages is tedious - many templates are
language-specific, many others have a corresponding template, but
incompatible parameters, and even if the parameters are compatible, there
is usually no comfortable mapping. Some work in that direction was done in
DBpedia, but AFAIK it's far from complete.
In ContentTranslation we have a simplistic mechanism for mapping between
template parameters in pairs of languages, with proof of concept for three
templates. We can enhance it with more templates, but the question is how
much can it scale.
Some templates shouldn't need such mapping at all - they should pull their
data from Wikidata. This is gradually being done for infoboxes in some
languages, and it's great.
But not all templates can be easily mapped to Wikidata data. For example -
reference templates, various IPA and language templates, quotation
formatting, and so on. For these, parameter mapping could be useful, but
doing this for a single language pair doesn't seem robust and reminds me of
the old ways in which interlanguage links were stored.
So, did anybody consider using Wikidata items of templates to store
multilingual template parameters mapping?
--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
“We're living in pieces,
I want to live in peace.” – T. Moore
For more than a year I am asking users to add their articles to Wikidata
when they have written it. That seems succesful, they added their articles
more and more and did understand how to do that. Until recently. Now I get
more and more complaints from users that they do not understand any more
how to add a newly written article to an item. They seem to have tried, but
fail in actual getting it managed. That is a worse development!
Romaine
Hey all,
so I see there is some work being done on mapping Wikidata data model
to RDF [1].
Just a thought: what if you actually used RDF and Wikidata's concepts
modeled in it right from the start? And used standard RDF tools, APIs,
query language (SPARQL) instead of building the whole thing from
scratch?
Is it just me or was this decision really a colossal waste of resources?
[1] http://korrekt.org/papers/Wikidata-RDF-export-2014.pdf
Martynas
http://graphityhq.com
Dear all,
I wanted to join in and give my birthday present to Wikidata (I am a
little bit late, though!)
(also, honestly, I didn't recall it was Wikidata's birthday, but it is
a nice occasion :P)
Here it is:
http://wikidataldf.com
What is LDF?
LDF stands for Linked Data Fragments, they are a new system for query
RDF datasets that stands middle way between having a SPARQL endpoint
and downloading the whole thing.
More formally LDF is «a publishing method [for RDF datasets] that
allows efficient offloading of query execution from servers to clients
through a lightweight partitioning strategy. It enables servers to
maintain availability rates as high as any regular HTTP server,
allowing querying to scale reliably to much larger numbers of
clients»[1].
This system was devised Ruben Verborgh, Miel Vander Sande and Pieter
Colpaert at Multimedia Lab (Ghent University) in Ghent, Belgium.
You can read more about it: http://linkeddatafragments.org/
What is Wikidata LDF?
Using the software by Verborgh et al. I have setup the website
http://wikidataldf.com that contains:
* an interface to navigate in the RDF data and query them using the
Triple Pattern Fragments client
* a web client where you can compose and execute SPARQL queries
This is not, strictly speaking, a SPARQL endpoint (not all the SPARQL
standard is implemented and it is slower, but it should be more
reliable, if you are interested in details, please do read more at the
link above).
The data are, for the moment, limited to the sitelinks dump but I am
working towards adding the other dump. I have taken the Wikidata RDF
dumps as of Oct, 13th 2014[2].
To use them I had to convert them in HDT format[3a][3b], using the
hdt-cpp library[3c] (devel) (which is taking quite a lot of resources
and computing time for the whole dumps, that's the reason why I
haven't published the rest yet ^_^).
DBpedia has also this[4]:
http://fragments.dbpedia.org/
All the software used is available under the MIT license on the LDF
repo on github[5a], and also the (two pages) website is available
here[5b].
I would like to thank Ruben for his feedback and his presentation
about LDF at SpazioDati in Trento, Italy (here's the slides[6]).
All this said, happy birthday Wikidata.
Cristian
[1] http://linkeddatafragments.org/publications/ldow2014.pdf
[2] https://tools.wmflabs.org/wikidata-exports/rdf/exports/
[3a] http://www.rdfhdt.org/
[3b] http://www.w3.org/Submission/HDT-Implementation/
[3c] https://github.com/rdfhdt/hdt-cpp
[4] http://sourceforge.net/p/dbpedia/mailman/message/32982329/
[5a] see the Browser.js, Server.js and Client.js repos in
https://github.com/LinkedDataFragments
[5b] https://github.com/CristianCantoro/wikidataldf
[6] http://www.slideshare.net/RubenVerborgh/querying-datasets-on-the-web-with-h…
Hello everyone,
I'm following up on Pyb's message about the Super Lachaise mobile app :
> I'm happy to announce the release of Super Lachaise on the App Store. It's
> a free mobile app that help you during the visit of the Père Lachaise
> cemetery. This is probably one of the firsts mobile apps to use Wikidata ;)
> http://www.superlachaise.fr/
> https://itunes.apple.com/fr/app/super-lachaise/id918263934
>
> Pyb
I'm the developer of the app. It's a small project I've been doing in my
spare time.
Not available in the Dutch iTunes Store...
Sorry about that. The app is currently only available in french language
and in the french app store, i expect to release an international/english
version before the end of the year.
That's a great idea!
> Just curious, for such a specific use case, why did you go for an App
> instead of a Website?
Thanks a lot ! There are a lot of tourists coming to the Père Lachaise
cemetery. The app is designed to be a companion, helping people finding
specific tombs with geolocation and photos, discovering what's around them
or accessing wikipedia/wikimedia commons data for the famous people buried
there.
How do you get the geocoordinates for the individual graves? Looing at
> http://www.superlachaise.fr/ I see Guillaume Apollinaire. His Wikidata
> https://www.wikidata.org/wiki/Q133855 has no geodata. The cemetery link
> and Findagrave seem neither to have geodata.
The geocoordinates come from OpenStreetMap. I've built an aggregate
database which links data from Wikidata and OSM :
https://github.com/MaximeLM/SuperLachaise/tree/master/PereLachaise/Resource…
I will contribute the geodata to Wikidata. Thanks for pointing that.
Maxime
ZOMFG, the tool that Denny introduced yesterday as a birthday gift is
unbelieavably useful and fun.
Here are a few thoughts I had about it:
I went over all the pages for the Hebrew-English pair. There were only 36,
and that is suspiciously low. Were all the articles in these languages
tested by this tool or only a subset?
Even though almost all of the tool's suggestions were correct It would be
problematic to fix these automatically. There were several types of article
pairs:
* Unrelated because one of the suggested pages was a disambiguation page
and the other was not. Sometimes there was a link to the correct related
page from the disambig page. If anybody makes a new version, this certainly
should be corrected.
* Related, but with explicit interlanguage links in the articles' source
code. This required old-style interwiki conflict resolution. There was a
surprisingly high number of these. I managed to resolve all the conflicts
manually, but it did take a few minutes for each case. Examples from
en.wikipedia: [[Bombe]], [[Bomba (cryptography)]], [[Diary of a Wimpy
Kid]], [[PFLAG]].
* Related, with a Wikidata item for each page, but without conflicts, so
easily mergeable. This can be done by a bot once it is identified for sure.
Adding links to a page without any language links shows a box to write a
language and a target title, and that's it. Adding a link to a new language
to a page which already has some interlanguage links opens the whole item
page in Wikidata (a whole other website!) and requires scrolling, editing
the links, and in many cases - merging the items manually. The result is
actually the same, so it would be very nice if the second case wouldn't be
so complicated.
That's it for now - I hope somebody finds it useful :)
I finished with Hebrew, and I'm going on to Russian, which has over a
thousand article pairs. IT'S INSANELY FUN.
--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
“We're living in pieces,
I want to live in peace.” – T. Moore
2014-10-29 19:56 GMT+02:00 Denny Vrandečić <vrandecic(a)google.com>:
> Folks,
>
> as you know, many Googlers are huge fans of Wikipedia. So here’s a little
> gift for Wikidata’s second birthday.
>
> Some of my smart colleagues at Google have run a few heuristics and
> algorithms in order to discover Wikipedia articles in different languages
> about the same topic which are missing language links between the articles.
> The results contain more than 35,000 missing links with a high confidence
> according to these algorithms. We estimate a precision of about 92+% (i.e.
> we assume that less than 8% of those are wrong, based on our evaluation).
> The dataset covers 60 Wikipedia language editions.
>
> Here are the missing links, available for download from the WMF labs
> servers:
>
> https://tools.wmflabs.org/yichengtry/merge_candidate.20141028.csv
>
> The data is published under CC-0.
>
> What can you do with the data? Since it is CC-0, you can do anything you
> want, obviously, but here are a few suggestions:
>
> There’s a small tool on WMF labs that you can use to verify the links (it
> displays the articles side by side from a language pair you select, and
> then you can confirm or contradict the merge):
>
> https://tools.wmflabs.org/yichengtry
>
> The tool does not do the change in Wikidata itself, though (we thought it
> would be too invasive if we did that). Instead, the results of the human
> evaluation are saved on WMF labs. You are welcome to take the tool and
> extend it with the possibility to upload the change directly on Wikidata,
> if you so wish, or, once the data is verified, to upload the results.
>
> Also, Magnus Manske is already busy uploading the data to the Wikidata
> game, so you can very soon also play the merge game on the data directly.
> He is also creating the missing items on Wikidata. Thanks Magnus for a very
> pleasant cooperation!
>
> I want to call out to my colleagues at Google who created the dataset -
> Jiang Bian and Si Li - and to Yicheng Huang, the intern who developed the
> tool on labs.
>
> I hope that this small data release can help a little with further
> improving the quality of Wikidata and Wikipedia! Thank you all, you are
> awesome!
>
> Cheers,
> Denny
>
>
>
> On Wed Oct 29 2014 at 10:52:05 AM Lydia Pintscher <
> lydia.pintscher(a)wikimedia.de> wrote:
>
> Hey folks :)
>
> Today Wikidata is turning two. It amazes me what we've achieved in
> just 2 years. We've built an incredible project that is set out to
> change the world. Thank you everyone who has been a part of this so
> far.
> We've put together some notes and opinions. And there are presents as
> well! Check them out and leave your birthday wishes:
> https://www.wikidata.org/wiki/Wikidata:Second_Birthday
>
>
> Cheers
> Lydia
>
> --
> Lydia Pintscher - http://about.me/lydia.pintscher
> Product Manager for Wikidata
>
> Wikimedia Deutschland e.V.
> Tempelhofer Ufer 23-24
> 10963 Berlin
> www.wikimedia.de
>
> Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
>
> Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
> unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
> Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
>
> _______________________________________________
> Wikidata-l mailing list
> Wikidata-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
> _______________________________________________
> Wikidata-l mailing list
> Wikidata-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
>
>
Hey folks :)
Today Wikidata is turning two. It amazes me what we've achieved in
just 2 years. We've built an incredible project that is set out to
change the world. Thank you everyone who has been a part of this so
far.
We've put together some notes and opinions. And there are presents as
well! Check them out and leave your birthday wishes:
https://www.wikidata.org/wiki/Wikidata:Second_Birthday
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
>
> Dear Pierre-Yves,
>
> On 10/28/2014 05:41 PM, Pierre-Yves Beaudouin wrote:
>
> >* I don't know because I'm not the developer of the app and my knowledge
> *>* is limited in this area. For many years now, I am collecting data
> *>* (information, photo, coordinates) about the cemetery. I've publish
> *>* everything on Commons, Wikidata and OSM, so developers can do something
> *>* smart with that ;)
> *
> How do you get the geocoordinates for the individual graves? Looing athttp://www.superlachaise.fr/ I see Guillaume Apollinaire. His Wikidata https://www.wikidata.org/wiki/Q133855 has no geodata. The cemetery link
> and Findagrave seem neither to have geodata.
>
>
> - Finn Årup Nielsen
>
>
Hi,
The geocoordinates are not on Wikidata. They are on OpenStreetMap and most
of them are also on Wikimedia Commons.
I didn't put the geocordinates on Wikidata because I'm still thinking how
to do it: put the geocordinates on the biography (with a qualifier) or on
the monument. And I didn't start the mass item creation of monuments
because I'm waiting the quantity datatype with units. That's why a lot of
my data are still on Commons
https://commons.wikimedia.org/wiki/Category:Grave_of_Guillaume_Apollinaire
Pyb
Dear Cristian,
> http://wikidataldf.com
Thanks so much for this, we have featured your dataset on our homepage:
http://linkeddatafragments.org/data/
> What is LDF?
Please allow me to make a technical correction here:
Linked Data Fragments are a uniform view
on _all_ possible interfaces to Linked Data,
not just the light-weight interface that Cristian set up:
- SPARQL endpoints offer Linked Data Fragments;
you can select a part of a dataset corresponding to a specific SPARQL query
- a server of Linked Data documents offers Linked Data Fragments;
you can select a part of a dataset corresponding to a specific subject
- …
What all those interfaces have in common,
is that they offer some part, some fragment, of a dataset;
hence the name “Linked Data Fragments”.
In addition, we have introduced a new kind of interface:
Triple Pattern Fragments, which offer access to parts of a dataset by triple pattern.
This is indeed a very lightweight system for the server,
as you can host live data without much processing resources.
SPARQL queries can be executed on the client side,
as Cristian's client instance shows: http://client.wikidataldf.com/
> «a publishing method [for RDF datasets] that
> allows efficient offloading of query execution from servers to clients
> through a lightweight partitioning strategy. It enables servers to
> maintain availability rates as high as any regular HTTP server,
> allowing querying to scale reliably to much larger numbers of
> clients»[1].
So the above definition is about triple pattern fragments,
not Linked Data Fragments interfaces (which include SPARQL endpoints) in general.
Thanks, Cristian, for setting this up!
I hope the Wikidata community finds good use for it.
Finally, live data from Wikidata can be queried with SPARQL :-)
Best,
Ruben