*The First Wikidata Workshop*
Co-located with the 19th International Conference on Semantic Web (ISWC
2020).
Date: October 29, 2020
The workshop will be held online, afternoon European time.
Website: https://wikidataworkshop.github.io/
== Important dates ==
Papers due: August 10, 2020
Notification of accepted papers: September 11, 2020
Camera-ready papers due: September 21, 2020
Workshop date: October 29, 2020
== Overview ==
Wikidata is an openly available knowledge base, hosted by the Wikimedia
Foundation. It can be accessed and edited by both humans and machines and
acts as a common structured-data repository for several Wikimedia projects,
including Wikipedia, Wiktionary, and Wikisource. It is used in a variety of
applications by researchers and practitioners alike.
In recent years, we have seen an increase in the number of publications
around Wikidata. While there are several dedicated venues for the broader
Wikidata community to meet, none of them focuses on publishing original,
peer-reviewed research. This workshop fills this gap - we hope to provide a
forum to build this fledgling scientific community and promote novel work
and resources that support it.
The workshop seeks original contributions that address the opportunities
and challenges of creating, contributing to, and using a global,
collaborative, open-domain, multilingual knowledge graph such as Wikidata.
We encourage a range of submissions, including novel research, opinion
pieces, and descriptions of systems and resources, which are naturally
linked to Wikidata and its ecosystem, or enabled by it. What we’re less
interested in are works which use Wikidata alongside or in lieu of other
resources to carry out some computational task - unless the work feeds back
into the Wikidata ecosystem, for instance by improving or commenting on
some Wikidata aspect, or suggesting new design features, tools and
practices.
We also encourage submissions on the topic of Abstract Wikipedia,
particularly around collaborative code management, natural language
generation by a community, the abstract representation of knowledge, and
the interaction between Abstract Wikipedia and Wikidata on the one, and
Abstract Wikipedia and the language Wikipedias on the other side.
We welcome interdisciplinary work, as well as interesting applications
which shed light on the benefits of Wikidata and discuss areas of
improvement.
The workshop is planned as an interactive half-day event, in which most of
the time will be dedicated to discussions and exchange rather than frontal
presentations. For this reason, all accepted papers will be presented in
short talks and accompanied by a poster. We are considering online options
in response to ongoing challenges such as travel restrictions and the
recent Covid-19 pandemic.
== Topics ==
Topics of submissions include, but are not limited to:
- Data quality and vandalism detection in Wikidata
- Referencing in Wikidata
- Anomaly, bias, or novelty detection in Wikidata
- Algorithms for aligning Wikidata with other knowledge graphs
- The Semantic Web and Wikidata
- Community interaction in Wikidata
- Multilingual aspects in Wikidata
- Machine learning approaches to improve data quality in Wikidata
- Tools, bots and datasets for improving or evaluating Wikidata
- Participation, diversity and inclusivity aspects in the Wikidata ecosystem
- Human-bot interaction
- Managing knowledge evolution in Wikidata
- Abstract Wikipedia
== Submission guidelines ==
We welcome the following types of contributions:
- Full research paper: Novel research contributions (7-12 pages)
- Short research paper: Novel research contributions of smaller scope than
full papers (3-6 pages)
- Position paper: Well-argued ideas and opinion pieces, not yet in the
scope of a research contribution (6-8 pages)
- Resource paper: New dataset or other resource directly relevant to
Wikidata, including the publication of that resource (8-12 pages)
- Demo paper: New system critically enabled by Wikidata (6-8 pages)
Submissions must be as PDF or HTML, formatted in the style of the Springer
Publications format for Lecture Notes in Computer Science (LNCS). For
details on the LNCS style, see Springer’s Author Instructions.
The papers will be peer-reviewed by at least two researchers. Accepted
papers will be published as open access papers on CEUR (we only publish to
CEUR if the authors agree to have their papers published).
Papers have to be submitted through easychair:
https://easychair.org/conferences/?conf=wikidataworkshop2020
== Proceedings ==
The complete set of papers will be published with the CEUR Workshop
Proceedings (CEUR-WS.org).
== Organizing committee ==
- Lucie-Aimée Kaffee, University of Southampton
- Oana Tifrea-Marciuska, Bloomberg
- Elena Simperl, King’s College London
- Denny Vrandečić, Wikimedia Foundation
== Programme committee ==
- Dan Brickley, Google
- Andrew D. Gordon, Microsoft Research & University of Edinburgh
- Dennis Diefenbach, University Jean Monet
- Aidan Hogan, Universidad de Chile
- Markus Krötzsch, Technische Universität Dresden
- Edgar Meij, Bloomberg
- Claudia Müller-Birn, FU Berlin
- Finn Årup Nielsen, Technical University of Denmark
- Thomas Pellissier Tanon, Télécom ParisTech
- Lydia Pintscher, Wikidata, Wikimedia Deutschland
- Alessandro Piscopo, BBC
- Marco Ponza, University of Pisa
- Simon Razniewski, Max Planck Institute for Informatics
- Miriam Redi, Wikimedia Foundation
- Cristina Sarasua, University of Zurich
- Maria-Esther Vidal, TIB Hannover
- Pavlos Vougiouklis, Huawei Technologies, Edinburgh
- Zainan Victor Zhou, Google
--
Lucie-Aimée Kaffee
Hello all,
As every quarter, the Wikidata development team will host an Office Hour on
July 21st at 16:00 UTC (18:00 CEST), on the Wikidata Telegram channel
<https://t.me/joinchat/AZriqUj5UagVMHXYzfZFvA>.
This session will be a bit special because we will have a guest: Guillaume
Lederrey from WMF's Search Team, who will present what they are working on
at the moment related to the Wikidata Query service: research that they
have been doing around the use of the WDQS, the reasons behind the issues
that we encounter for the past months with keeping the data up to date, and
different future paths for the service.
So if you're interested in the topic, you can prepare your questions until
July 21st!
As usual, notes of the discussions will be published onwiki
<https://www.wikidata.org/wiki/Category:Office_hour_notes> after the
meeting.
Cheers,
--
Léa Lacroix
Project Manager Community Communication for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hello everyone,
As part of the Small wiki toolkits
<https://meta.wikimedia.org/wiki/Small_wiki_toolkits> initiative, a Starter
kit <https://meta.wikimedia.org/wiki/Small_wiki_toolkits/Starter_kit> has
been developed for smaller language Wikimedia wikis! This Starter kit lists
resources, tools, and recommendations in technical areas (e.g., templates,
bots, gadgets, etc.) relevant to smaller wikis that are just getting
started. Small wiki contributors can use it to make their community's
workflow easier. You can now use and promote the Starter kit in your wiki
community, and start translating the landing page and its subpages in a
language you want.
If you have any questions, ideas for venues where it should be shared or
wiki pages where it should be linked, or any other suggestions for
improving it further, please share on this talk page
<https://meta.wikimedia.org/wiki/Talk:Small_wiki_toolkits/Starter_kit>.
If you are interested in helping with the Small Wiki Toolkits initiative
and can offer help with running workshops, developing toolkits, or
exchanging problems and challenges in smaller wiki communities, add
yourself as a member here:
https://meta.wikimedia.org/wiki/Small_wiki_toolkits#Members
Cheers,
Srishti
*Srishti Sethi*
Developer Advocate
Wikimedia Foundation <https://wikimediafoundation.org/>
Hi,
The upcoming domain name migration to on the Wikimedia Toolforge implies
that OpenRefine users need to update their Wikidata reconciliation
service to the new endpoint:
https://wdreconcile.toolforge.org/en/api
or by replacing "en" by any other Wikimedia language code.
The new home page of the service is at:
https://wdreconcile.toolforge.org/
This new endpoint will be available by default in the upcoming release
of OpenRefine (3.4).
For details about why an automatic migration via redirects is sadly not
possible, see this Phabricator ticket:
https://phabricator.wikimedia.org/T254172
Cheers,
Antonin
Hello,
Today the Wikimedia Foundation would like to introduce a new community
blog. It's called "Diff" (diff.wikimedia.org) and is a blog by – and
for – the Wikimedia volunteer community to connect and share
learnings, stories, and ideas from across our movement. We'd like to
encourage you to learn more about Diff and how it can help you in
sharing and learning from your fellow Wikimedians.
Everyone is invited to contribute!
https://diff.wikimedia.org/2020/07/14/welcome-to-diff-a-community-blog-for-…
The name “Diff” is in reference to the wiki interface that displays
the difference between one version and another of a Wikipedia page. It
also reflects the “difference” our communities and movement make in
the world every day.
For some background, Diff builds on lessons and experiences from the
Wikimedia Blog, the Wikimedia Foundation News, and Wikimedia Space;
previous posts from these channels are archived on Diff. The channel
is primarily intended for community-authored posts, in which
volunteers can share their stories, learnings, and ideas with each
other.
Diff offers a simple and accessible editorial process, moderated by
Foundation communications staff and open to volunteers, to encourage
participation from all — especially emerging and under-represented
communities. Additionally, content on Diff can be written and
translated into languages to reach a wide audience. Diff also has a
code of conduct and comments can be flagged and moderated.
Still curious to learn more?
https://diff.wikimedia.org/2020/07/14/welcome-to-diff-a-community-blog-for-…
Yours,
Chris Koerner (he/him)
Community Relations Specialist
Wikimedia Foundation
This is an announcement for a breaking change to the pageterms submodule of
the query API module, which only affects Wikibase repository wikis. If you
do not use that API module, or only use it on client wikis (e. g.
Wikipedias) and not on repository wikis (Wikidata, Wikimedia Commons), you
can ignore this message.
For years, the pageterms
<https://www.wikidata.org/wiki/Special:ApiHelp/query%2Bpageterms> API
module has served a double role: on client wikis, it returned the “terms”
(Label, Description, Aliases) of the Wikidata Item linked to the given
page(s), whereas on repo wikis, it would return the terms of the Item (or
other Entity) on that page itself. For example, querying for the Label of
Wikipedia:Village pump on English Wikipedia
<https://en.wikipedia.org/w/api.php?action=query&prop=pageterms&titles=Wikip…>
would return “Project:Village pump” (the Label of Q16503
<https://www.wikidata.org/wiki/Q16503>), but querying for the Label of
Wikidata:Project chat on Wikidata
<https://www.wikidata.org/w/api.php?action=query&prop=pageterms&titles=Wikid…>
would not return anything, even though that page is linked to the same Item
– you would have to query for the Label of Q16503
<https://www.wikidata.org/w/api.php?action=query&prop=pageterms&titles=Q1650…>
instead. This behavior is inconsistent and also mixes repo and client
concerns in a way that makes the Wikibase code harder to maintain.
To resolve this, we introduced a new entityterms
<https://www.wikidata.org/wiki/Special:ApiHelp/query%2Bentityterms> API
module (a submodule of the query module, just like the pageterms module)
which has the same behavior as the pageterms module currently has for Item
(or other Entity) pages, and which is only available on repo wikis. If you
want to get the terms of Q16503, you can now use
action=query&prop=entityterms&titles=Q16503
<https://www.wikidata.org/w/api.php?action=query&prop=entityterms&titles=Q16…>
instead of action=query&prop=pageterms&titles=Q16503
<https://www.wikidata.org/w/api.php?action=query&prop=pageterms&titles=Q16503>.
(You can also use wbgetentities
<https://www.wikidata.org/wiki/Special:ApiHelp/wbgetentities>, which gives
you much more control over the returned data; pageterms/entityterms may be
faster and can also be combined with other submodules of the query module.)
On or shortly after 5 August 2020, we will remove the special repo behavior
of the pageterms module, and it will then behave just like it always has on
client wikis, and return the terms of the Item linked to a page, not the
terms of the Item (or other Entity) on a page. (Because the new API module
is already available on Wikidata, and you can start using it immediately,
we are not making this pageterms behavior change available on Test Wikidata
significantly before that date.)
If you have any issue or question, feel free to leave a comment at T257658
<https://phabricator.wikimedia.org/T257658>. For more information, see also
T115117 <https://phabricator.wikimedia.org/T115117>, T255882
<https://phabricator.wikimedia.org/T255882> and T256255
<https://phabricator.wikimedia.org/T256255>.
Cheers,
Lucas
--
Lucas Werkmeister (he/er)
Full Stack Developer
Wikimedia Deutschland e. V. | Tempelhofer Ufer 23-24 | 10963 Berlin
Phone: +49 (0)30 219 158 26-0
https://wikimedia.de
Imagine a world in which every single human being can freely share in the
sum of all knowledge. Help us to achieve our vision!
https://spenden.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 B. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
------------------------------------------------------------------------------
FINAL CALL FOR CONTRIBUTIONS
THE SUBMISSION DEADLINE IS ON AUGUST 10TH, 2020
------------------------------------------------------------------------------
The Fifteenth International Workshop on
ONTOLOGY MATCHING
(OM-2020)
http://om2020.ontologymatching.org/
November 2nd or 3rd, 2020,
International Semantic Web Conference (ISWC) Workshop Program,
VIRTUAL CONFERENCE
BRIEF DESCRIPTION AND OBJECTIVES
Ontology matching is a key interoperability enabler for the Semantic Web,
as well as a useful technique in some classical data integration tasks
dealing with the semantic heterogeneity problem. It takes ontologies
as input and determines as output an alignment, that is, a set of
correspondences between the semantically related entities of those
ontologies.
These correspondences can be used for various tasks, such as ontology
merging, data interlinking, query answering or navigation over knowledge
graphs.
Thus, matching ontologies enables the knowledge and data expressed
with the matched ontologies to interoperate.
The workshop has three goals:
1.
To bring together leaders from academia, industry and user institutions
to assess how academic advances are addressing real-world requirements.
The workshop will strive to improve academic awareness of industrial
and final user needs, and therefore, direct research towards those needs.
Simultaneously, the workshop will serve to inform industry and user
representatives about existing research efforts that may meet their
requirements. The workshop will also investigate how the ontology
matching technology is going to evolve, especially with respect to
data interlinking, knowledge graph and web table matching tasks.
2.
To conduct an extensive and rigorous evaluation of ontology matching
and instance matching (link discovery) approaches through
the OAEI (Ontology Alignment Evaluation Initiative) 2020 campaign:
http://oaei.ontologymatching.org/2020/
3.
To examine similarities and differences from other, old, new
and emerging, techniques and usages, such as web table matching
or knowledge embeddings.
This year, in sync with the main conference, we encourage submissions
specifically devoted to: (i) datasets, benchmarks and replication studies,
services, software, methodologies, protocols and measures
(not necessarily related to OAEI), and (ii) application of
the matching technology in real-life scenarios and assessment
of its usefulness to the final users.
TOPICS of interest include but are not limited to:
Business and use cases for matching (e.g., big, open, closed data);
Requirements to matching from specific application scenarios (e.g.,
public sector, homeland security);
Application of matching techniques in real-world scenarios (e.g., in
cloud, with mobile apps);
Formal foundations and frameworks for matching;
Novel matching methods, including link prediction, ontology-based
access;
Matching and knowledge graphs;
Matching and deep learning;
Matching and embeddings;
Matching and big data;
Matching and linked data;
Instance matching, data interlinking and relations between them;
Privacy-aware matching;
Process model matching;
Large-scale and efficient matching techniques;
Matcher selection, combination and tuning;
User involvement (including both technical and organizational aspects);
Explanations in matching;
Social and collaborative matching;
Uncertainty in matching;
Expressive alignments;
Reasoning with alignments;
Alignment coherence and debugging;
Alignment management;
Matching for traditional applications (e.g., data science);
Matching for emerging applications (e.g., web tables, knowledge graphs).
SUBMISSIONS
Contributions to the workshop can be made in terms of technical papers and
posters/statements of interest addressing different issues of ontology
matching
as well as participating in the OAEI 2020 campaign. Long technical papers
should
be of max. 12 pages. Short technical papers should be of max. 5 pages.
Posters/statements of interest should not exceed 2 pages.
All contributions have to be prepared using the LNCS Style:
http://www.springer.com/computer/lncs?SGWID=0-164-6-793341-0
and should be submitted in PDF format (no later than August 10th, 2020)
through the workshop submission site at:
https://www.easychair.org/conferences/?conf=om2020
Contributors to the OAEI 2020 campaign have to follow the campaign
conditions
and schedule at http://oaei.ontologymatching.org/2020/.
DATES FOR TECHNICAL PAPERS AND POSTERS:
August 10th, 2020: Deadline for the submission of papers.
September 11th, 2020: Deadline for the notification of
acceptance/rejection.
September 21st, 2020: Workshop camera ready copy submission.
November 2nd or 3rd, 2020: OM-2020, Virtual Conference.
Contributions will be refereed by the Program Committee.
Accepted papers will be published in the workshop proceedings
as a volume of CEUR-WS as well as indexed on DBLP.
ORGANIZING COMMITTEE
1. Pavel Shvaiko (main contact)
Trentino Digitale, Italy
2. Jérôme Euzenat
INRIA & Univ. Grenoble Alpes, France
3. Ernesto Jiménez-Ruiz
City, University of London, UK & SIRIUS, University of Oslo, Norway
4. Oktie Hassanzadeh
IBM Research, USA
5. Cássia Trojahn
IRIT, France
PROGRAM COMMITTEE (to be completed):
Alsayed Algergawy, Jena University, Germany
Manuel Atencia, INRIA & Univ. Grenoble Alpes, France
Zohra Bellahsene, LIRMM, France
Jiaoyan Chen, University of Oxford, UK
Valerie Cross, Miami University, USA
Jérôme David, University Grenoble Alpes & INRIA, France
Gayo Diallo, University of Bordeaux, France
Daniel Faria, Instituto Gulbenkian de Ciéncia, Portugal
Alfio Ferrara, University of Milan, Italy
Marko Gulic, University of Rijeka, Croatia
Wei Hu, Nanjing University, China
Ryutaro Ichise, National Institute of Informatics, Japan
Antoine Isaac, Vrije Universiteit Amsterdam & Europeana, Netherlands
Naouel Karam, Fraunhofer, Germany
Prodromos Kolyvakis, EPFL, Switzerland
Patrick Lambrix, Linköpings Universitet, Sweden
Oliver Lehmberg, University of Mannheim, Germany
Majeed Mohammadi, TU Delft, Netherlands
Peter Mork, MITRE, USA
Andriy Nikolov, Metaphacts GmbH, Germany
George Papadakis, University of Athens, Greece
Catia Pesquita, University of Lisbon, Portugal
Henry Rosales-Méndez, University of Chile, Chile
Kavitha Srinivas, IBM, USA
Giorgos Stoilos, Huawei Technologies, Greece
Pedro Szekely, University of Southern California, USA
Ludger van Elst, DFKI, Germany
Xingsi Xue, Fujian University of Technology, China
Ondrej Zamazal, Prague University of Economics, Czech Republic
Songmao Zhang, Chinese Academy of Sciences, China
-------------------------------------------------------
More about ontology matching:
http://www.ontologymatching.org/http://book.ontologymatching.org/
Best Regards,
Pavel
-------------------------------------------------------
Pavel Shvaiko, PhD
Trentino Digitale, Italy
http://www.ontologymatching.org/https://www.trentinodigitale.it/http://www.dit.unitn.it/~pavel
-------------------------------------------------------
--
Cap. Soc. Euro 6.433.680,00 - REG. IMP. / C.F. / P.IVA 00990320228
E-mail:
tndigit(a)tndigit.it <mailto:infotn@infotn.it> - www.trentinodigitale.it
<http://www.infotn.it>
Società soggetta ad attività di direzione e
coordinamento da parte della Provincia Autonoma di Trento - C.Fisc.
00337460224.
Questo messaggio è indirizzato esclusivamente ai destinatari
in intestazione, può contenere informazioni protette e riservate ai sensi
della normativa vigente e ne è vietato qualsiasi impiego diverso da quello
per cui è stato inviato. Se lo avete ricevuto per errore siete pregati di
eliminarlo in ogni sua parte e di avvisare il mittente
Hello all!
The Search Platform team will join the WIkidata office hours on July 21st
16:00 UTC [1]. We are looking forward to discussing Wikidata Query Service
and anything else you might find of interest.
We've been hard at work on Wikimedia Commons Query Service (WCQS) [2]. This
will be a SPARL endpoint similar to WDQS, but serving the Structured Data
on Commons dataset. Our goal is to open a beta service, hosted on Wikimedia
Cloud Service (WMCS) by the end of July. The service will require an
account on Commons for authentication and will allow federation with WDQS.
We don't have a streaming update process ready yet, the data will be
reloaded from Commons dumps weekly for a start.
As part of that work, the dumps for Structured Data on Commons are now
available [3]. Note that the prefix used in the TTL dumps is "wd", which
does not make much sense. We are working with WMDE on renaming the
prefixes, but this is more complex than expected since "wd" is hardcoded in
more places than it should be. Those prefix should only be valid in the
local context of the dumps, so renaming them is technically a non breaking
change. That being said, if you start using those dumps, make sure you
don't rely on this prefix, or that you are ready for a rename [4].
We are planning to dig more into the data we have to get a better
understanding of the use cases around WDQS [5] (not much content on that
task yet, but it is coming). Some very preliminary analysis indicates that
less then 2% of the queries on WDQS generate more than 90% of the load.
This is definitely something we need to better understand. We will be
working on defining the kind of questions we need to answer, and improving
our data collection to be able to answer those questions.
We have started an internal discussion around "planning for disaster" [6].
We want to better understand the potential failure scenarios around WDQS
and have a plan if that worst case does happen. This will include some
analytics work and some testing to better understand the constraints and
what degraded mode we might still be able to provide in case of
catastrophic failure.
Thanks for reading!
Guillaume
[1] https://www.wikidata.org/wiki/Wikidata:Events#Office_hours
[2] https://phabricator.wikimedia.org/T251488
[3] https://dumps.wikimedia.org/other/wikibase/commonswiki/
[4]
https://dumps.wikimedia.org/other/wikibase/commonswiki/README_commonsrdfdum…
[5] https://phabricator.wikimedia.org/T257045
[6] https://phabricator.wikimedia.org/T257055
--
Guillaume Lederrey
Engineering Manager, Search Platform
Wikimedia Foundation
UTC+1 / CET