Please disseminate to interested colleagues as appropriate.
Last call: registration is closing for the Celtic Knot: Wikipedia Language Conference this week.
The Celtic Knot: Wikipedia Language Conference<https://media.ed.ac.uk/media/1_nr4uufq9>– This full day conference on 6 July is for colleagues interested in how technology can support language communities. Guest speakers from all across Europe will be joining us in Edinburgh to showcase and discuss a range of exciting open education, open data and open knowledge initiatives. Registration for this event closes 27th June so don’t delay in booking your place<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017>.
New satellite event: Introduction to Wikidata and the Wikidata Query Service<https://www.eventbrite.co.uk/e/introduction-to-wikidata-and-wikidata-query-…> – this event is now scheduled 11am to 1pm on Tuesday 4 July presented by Léa Lacroix (Project Manager Community Communication for Wikidata, Wikimedia Deutschland). Léa will also be joining us at the Celtic Knot as part of the Wikidata workshop on 6 July too. Just one of the many great speakers<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017#Speakers> and presentations<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017/Programme> at the Celtic Knot discussing community engagement and how to make Celtic and Indigenous languages a shared hub of online knowledge. Includes presentations from Catalan Wikipedia, Basque Wikipedia, Northern Sami Wikipedia, Welsh Wikipedia, Scots Gaelic Wikipedia, Greek Wikipedia and more.
Navino Evans from Histropedia will also be joining us for Wikidata I/O: a showcase event at the Repository Fringe 2017<http://www.repositoryfringe.org/> conference in August. Three workshops will include: (1) how to add data in bulk (2) the wide variety of ways the data can be consumed, queried and visualised and (3) how to build SPARQL queries. 4th August 2017<http://rfringe17.blogs.edina.ac.uk/registration/>
Very best regards,
Ewan McAndrew
Wikimedian in Residence
Tel: 07719 330076
Email: ewan.mcandrew(a)ed.ac.uk
Subscribe to the mailing list: wikimedia(a)mlist.is.ed.ac.uk
My working hours are 10.30am to 6.30pm Monday to Friday.
Wikipedia Project Page for the residency: https://en.wikipedia.org/wiki/Wikipedia:University_of_Edinburgh
The University of Edinburgh, Floor H (West), Argyle House, 3 Lady Lawson Street, Edinburgh, EH3 9DR.
www.ed.ac.uk
>
> Concerning keywords in the description field of properties: i think
> that is a shortcoming of the current autocomplete API that i'm using,
> which is the same one Wikidata uses for the searchbox in the top-right
> corner. Apparently they're not indexing those keywords, which makes it
> difficult for me to also include them in my search.
>
>
Where is the documentation for this feature? I have been getting a number
of interesting comments from GLAMs about the difficulty of using Authority
Controls within their in-house Content management systems. I feel like this
API might be a really tangible way of promoting Wikidata as facilitating
consistent identifiers for metadata.
Cheers,
Alex
Hey everyone,
i've made a tool that allows you to query Wikidata in a visual way
without using SPARQL. It's called VizQuery:
http://tools.wmflabs.org/hay/vizquery/
The possibilities of using Wikidata to do interesting queries are
endless, and the current query service allows for very powerful
queries indeed. However, i feel that for the general public,
especially those who are not that technical, it might be a bit
overwhelming and difficult for them to learn a complex language such
as SPARQL. To make people familiar with the concept of queries i
believe a somewhat less intimidating approach might be useful, hence
this tool.
VizQuery is only capable of doing a subset of possible queries. It's
basically simple triples, variables (prefixed with '?') and literals
(between "quotes"). You can do pretty powerful queries with only those
things though. For example, here's a query with vegetarians who are
married to a vegetarian:
http://bit.ly/2sydpmW
Under the hood VizQuery uses Ruben Verborgh's SPARQL.js library to
convert between JSON and SPARQL, so theoretically every SPARQL query
you could do in the regular query service can be done in VizQuery.
However, many queries won't work because the visual interface only
supports a subset of options: it's pretty hard to create user-friendly
GUI representations of many of the complex SPARQL features. :)
Anyway, i'd like to hear what you think. Bugs, feature request and
pull requests are also welcome on my Github page:
https://github.com/hay/wiki-tools
Kind regards,
-- Hay
Hi Wikidatans,
After several delays we are finally starting to think seriously about
mapping the General Finnish Ontology YSO [1] to Wikidata. A "YSO ID"
property (https://www.wikidata.org/wiki/Property:P2347) was added to
Wikidata some time ago, but it has been used only a few times so far.
Recently some 6000 places have been added to "YSO Places" [2], a new
extension of YSO, which was generated from place names in YSA and
Allärs, our earlier subject indexing vocabularies. It would probably
make sense to map these places to Wikidata, in addition to the general
concepts in YSO. We have already manually added a few links from YSA/YSO
places to Wikidata for newly added places, but this approach does not
scale if we want to link the thousands of existing places.
We also have some indirect sources of YSO/Wikidata mappings:
1. YSO is mapped to LCSH, and Wikidata also to LCSH (using P244, LC/NACO
Authority File ID). I digged a bit into both sets of mappings and found
that approximately 1200 YSO-Wikidata links could be generated from the
intersection of these mappings.
2. The Finnish broadcasting company Yle has also created some mappings
between KOKO (which includes YSO) and Wikidata. Last time I looked at
those, we could generate at least 5000 YSO-Wikidata links from them.
Probably more nowadays.
Of course, indirect mappings are a bit dangerous. It's possible that
there are some differences in meaning, especially with LCSH which has a
very different structure (and cultural context) than YSO. Nevertheless I
think these could be a good starting point, especially if a tool such as
Mix'n'Match could be used to verify them.
Now my question is, given that we already have or could easily generate
thousands of Wikidata-YSO mappings, but the rest would still have to be
semi-automatically linked using Mix'n'Match, what would be a good way to
approach this? Does Mix'n'Match look at existing statements (in this
case YSO ID / P2347) in Wikidata when you load a new catalog, or ignore
them?
I can think of at least these approaches:
1. First import the indirect mappings we already have to Wikidata as
P2347 statements, then create a Mix'n'Match catalog with the remaining
YSO concepts. The indirect mappings would have to be verified separately.
2. First import the indirect mappings we already have to Wikidata as
P2347 statements, then create a Mix'n'Match catalog with ALL the YSO
concepts, including the ones for which we already have imported a
mapping. Use Mix'n'Match to verify the indirect mappings.
3. Forget about the existing mappings and just create a Mix'n'Match
catalog with all the YSO concepts.
Any advice?
Thanks,
-Osma
[1] http://finto.fi/yso/
[2] http://finto.fi/yso-paikat/
--
Osma Suominen
D.Sc. (Tech), Information Systems Specialist
National Library of Finland
P.O. Box 26 (Kaikukatu 4)
00014 HELSINGIN YLIOPISTO
Tel. +358 50 3199529
osma.suominen(a)helsinki.fi
http://www.nationallibrary.fi
Hi all,
Below you can find the announcement for the upcoming Research Showcase. [1]
I'm forwarding the announcement to this list as the second speaker of this
month's showcase is
Markus Kroetzsch and he will talk about Understanding Wikidata Queries.
Please find more details below.
Best,
Leila
[1] https://www.mediawiki.org/wiki/Wikimedia_Research/Showcase
--
Leila Zia
Senior Research Scientist
Wikimedia Foundation
---------- Forwarded message ----------
From: Sarah R <srodlund(a)wikimedia.org>
Date: Mon, Jun 19, 2017 at 12:47 AM
Subject: [Wiki-research-l] Research Showcase Wednesday June 21, 2017
To: wikimedia-l(a)lists.wikimedia.org, analytics(a)lists.wikimedia.org,
wiki-research-l(a)lists.wikimedia.org
Hi Everyone,
The next Research Showcase will be live-streamed this Wednesday, June 21,
2017 at 11:30 AM (PST) 18:30 UTC.
YouTube stream: https://www.youtube.com/watch?v=i2jpKRwPT-Q
As usual, you can join the conversation on IRC at #wikimedia-research. And,
you can watch our past research showcases here
<https://www.mediawiki.org/wiki/Wikimedia_Research/Showcase#June_2017>.
This month's presentations:
Title: Problematizing and Addressing the Article-as-Concept Assumption in
Wikipedia
By *Allen Yilun Lin*
Abstract: Wikipedia-based studies and systems frequently assume that each
article describes a separate concept. However, in this paper, we show that
this article-as-concept assumption is problematic due to editors’ tendency
to split articles into parent articles and sub-articles when articles get
too long for readers (e.g. “United States” and “American literature” in the
English Wikipedia). In this paper, we present evidence that this issue can
have significant impacts on Wikipedia-based studies and systems and
introduce the subarticle matching problem. The goal of the sub-article
matching problem is to automatically connect sub-articles to parent
articles to help Wikipedia-based studies and systems retrieve complete
information about a concept. We then describe the first system to address
the sub-article matching problem. We show that, using a diverse feature set
and standard machine learning techniques, our system can achieve good
performance on most of our ground truth datasets, significantly
outperforming baseline approaches.
Title: Understanding Wikidata Queries
By *
Markus Kroetzsch*
Abstract: Wikimedia provides a public service that lets anyone answer
complex questions over the sum of all knowledge stored in Wikidata. These
questions are expressed in the query language SPARQL and range from the
most simple fact retrievals ("What is the birthday of Douglas Adams?") to
complex analytical queries ("Average lifespan of people by occupation").
The talk presents ongoing efforts to analyse the server logs of the
millions of queries that are answered each month. It is an important but
difficult challenge to draw meaningful conclusions from this dataset. One
might hope to learn relevant information about the usage of the service and
Wikidata in general, but at the same time one has to be careful not to be
misled by the data. Indeed, the dataset turned out to be highly
heterogeneous and unpredictable, with strongly varying usage patterns that
make it difficult to draw conclusions about "normal" usage. The talk will
give a status report, present preliminary results, and discuss possible
next steps.
--
Sarah R. Rodlund
Senior Project Coordinator-Product & Technology, Wikimedia Foundation
srodlund(a)wikimedia.org
_______________________________________________
Wiki-research-l mailing list
Wiki-research-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
NYPL's Photographers' Identities Catalog (P2750) has a new data view that
would be preferable for use in WD (the IDs are, of course, unchanged). The
current formatted URL is
http://pic.nypl.org/map/?DisplayName=$1
But I'd rather it point to:
http://pic.nypl.org/constituents/$1
I can't seem to edit it (perhaps because I'm on my phone, or more likely
it's restricted for security purposes). How can we get this switched?
Many thanks in advance!
David
--
*David Lowe | The New York Public Library**Specialist II, Photography
Collection*
*Photographers' Identities Catalog <http://pic.nypl.org>*
Dear colleagues,
How can technology support language communities?
Join us at the Celtic Knot: Wikipedia Language conference<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017> taking place Thursday 6 July 2017 at the University of Edinburgh Business School to find out. Booking closes 27 June so don’t delay!
The main objective for Celtic Knot 2017 is the coming together of those working to support Celtic and Indigenous Languages in the same room at same time; strengthening the bonds into a 'knot' and leading into action. We welcome diverse attendees ranging from Wikimedians, linguists, educators, researchers, information professionals, media professionals, translators, learning technologists and more coming together to share good practice and find fruitful new collaborations to support language communities as a result of the event.
New satellite event: Introduction to Wikidata and the Wikidata Query Service<https://www.eventbrite.co.uk/e/introduction-to-wikidata-and-wikidata-query-…> – this event is now scheduled 11am to 1pm on Tuesday 4 July presented by Léa Lacroix (Project Manager Community Communication for Wikidata, Wikimedia Deutschland). Léa will also be joining us at the Celtic Knot as part of the Wikidata workshop on 6 July too. Just one of the many great speakers<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017#Speakers> and presentations<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017/Programme> at the Celtic Knot:
Keynote speakers
* Professor Antonella Sorace<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017#Keynotes> - Professor of Developmental Linguistics at the University of Edinburgh<http://www.lel.ed.ac.uk/%7Eantonell/> and founding director of Bilingualism Matters<http://www.bilingualism-matters.org.uk/> will be speaking on ‘Bilingualism in minority languages: a resource and an opportunity’.
* Jason Evans<https://wikimedia.org.uk/wiki/Celtic_Knot_Conference_2017#Keynotes> - Wikimedian in Residence at the National Library of Wales<https://wikimedia.org.uk/wiki/Expert_outreach/Wikipedian_in_Residence_at_th…> will discuss his strategy for working with Wikimedia UK and the Welsh Government to develop the Welsh Wicipedia using a combination of community engagement, data manipulation and the implementation of Open Access policies.
Confirmed speakers also include:
* Susan Ross – Gaelic Wikipedian in Residence<https://blog.wikimedia.org.uk/2017/01/wikimedia-uk-and-national-library-of-…> at the National Library of Scotland.<http://www.nls.uk/news/archive/2017/01/gaelic-wikipedian-begins>
* Dr. Sharon Arbuthnot<http://pure.qub.ac.uk/portal/en/persons/sharon-arbuthnot%28519b4b40-d075-4b…> - Research Fellow, Queen's University, Belfast. Presenting on the AHRC-funded eDIL project<http://www.dil.ie/> (Irish Language dictionary) on Wednesday 5th July.
* Gareth Morlais – the Welsh Language Unit, Welsh Government. Gareth will speak about how mapping how much importance major companies (Google, Twitter, Apple) attach to creative activity on Wikipedia led to the Welsh Government helping to fund two Welsh-language Wikipedia initiatives.
* Delyth Prys – Head of the Language Technologies Unit, Bangor University, will speak on Welsh/Celtic speech technology and why text-to-speech and speech recognition are becoming increasingly important in our digital world.
* Àlex Hinojo<http://www.alexhinojo.cat/en/bio/> – Executive Director, Amical Wikimedia<https://meta.wikimedia.org/wiki/Amical_Wikimedia> on the Catalan language project.
* Iñaki Lopez de Luzuriaga – Developing the Basque Wikipedia: From corpus expansion to outreach.
* Astrid Carlsen – Executive Director, Wikimedia Norge speaking on Norwegian Bokmål, Norwegian Nynorsk and building a project to revitalize the Northern Sami Wikipedia.
* Robin Owain – Wales Manager, Wikimedia UK, speaking on recent developments supporting the Welsh language community.
* Mina Theofilatou presenting on ‘The Kefalonian Dialect in Wiktionary and how Wikitherapy addresses social equality in open-source language projects’.
* Duncan Brown - Llên Natur<http://www.llennatur.com/Drupal7/llennatur/>, presenting on ‘Y BYWIADUR: the dictionary of life’.
* Rémy Gerbet - Wikimedia France; presenting on the Lingua Libre project<https://www.lingualibre.fr/> for massive open audio recording.
* Käbi Suvi - Wikimedia Estonia on the ‘Miljon+’ project as part of Estonia’s 100th anniversary.
* Ilario Valdelli - Wikimedia Switzerland, speaking on the Digital Library in Romansch and the new initiatives to map the archeological sites connected with Celtic culture in the Alps.
* Wikipedia’s new Content Translation tool and how it has been successfully employed in Higher Education to share knowledge between different language Wikipedias and give students meaningful published practice – Ewan McAndrew, Wikimedian in Residence.
It promises to be a great event – including a panel on the Politics of Language Online, excellent papers, workshops and discussion spaces. Please feel free to forward this event to interested colleagues in your network.
If you would like to learn more then please contact me direct at ewan.mcandrew(a)ed.ac.uk<mailto:ewan.mcandrew@ed.ac.uk>.
Very best regards,
Ewan
Ewan McAndrew
Wikimedian in Residence
Tel: 07719 330076
Email: ewan.mcandrew(a)ed.ac.uk
Subscribe to the mailing list: wikimedia(a)mlist.is.ed.ac.uk
My working hours are 10.30am to 6.30pm Monday to Friday.
Wikipedia Project Page for the residency: https://en.wikipedia.org/wiki/Wikipedia:University_of_Edinburgh
The University of Edinburgh, Floor H (West), Argyle House, 3 Lady Lawson Street, Edinburgh, EH3 9DR.
www.ed.ac.uk
The University of Edinburgh is a charitable body, registered in
Scotland, with registration number SC005336.
Hi Magnus, Osma,
I suppose the scenario Osma pointed out is quite common for knowledge organization systems and in particular thesauri: Matching could take advantage of multilingual labels and also of synonyms, which are defined in the KOS.
For the populating STW Thesaurus for Economics ID (P3911), my preliminary plan was to match with all multilingual labels and synonyms as search string in a custom WD endpoint (Fuseki, with full text search support), and display in the ranked SPARQL results of the search with a column with a valid insert statement that can be copied and pasted into QuickStatements2.
Since Stas just announced an extension for WDQS with fulltext search (if I haven’t misunderstood his mail of 2017-06-12), it is perhaps now possible to do this kind of matching in WDQS.
It would be great if such an extended matching could be integrated into M’n’m.
Cheers, Joachim
Von: Wikidata [mailto:wikidata-bounces@lists.wikimedia.org] Im Auftrag von Magnus Manske
Gesendet: Dienstag, 6. Juni 2017 16:07
An: Discussion list for the Wikidata project.
Betreff: Re: [Wikidata] Mix'n'Match with existing (indirect) mappings
On Tue, Jun 6, 2017 at 2:44 PM Osma Suominen <osma.suominen(a)helsinki.fi<mailto:osma.suominen@helsinki.fi>> wrote:
By the way, we also have multilingual labels that could perhaps improve
the automatic matching. YSO generally has fi/sv/en, YSO places has
fi/sv. Can you make use of these too if I provided them in additional
columns?
Sorry, mix'n'match only does single language labels.