Dear all,
last year, we applied for a Wikimedia grant to feed qualified data from Wikipedia infoboxes (i.e. missing statements with references) via the DBpedia software into Wikidata. The evaluation was already quite good, but some parts were still missing and we would like to ask for your help and feedback for the next round. The new application is here: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/GlobalFactSync
The main purpose of the grant is:
- Wikipedia infoboxes are quite rich, are manually curated and have references. DBpedia is already extracting that data quite well (i.e. there is no other software that does it better). However, extracting references is not a priority on our agenda. They would be very useful to Wikidata, but there are no user requests for this from DBpedia users.
- DBpedia also has all the infos of all infoboxes of all Wikipedia editions (>10k pages), so we also know quite well, where Wikidata is used already and where information is available in Wikidata or one language version and missing in another.
- side-goal: bring the Wikidata, Wikipedia and DBpedia communities closer together
Here is a diff between the old an new proposal:
- extraction of infobox references will still be a goal of the reworked proposal
- we have been working on the fusion and data comparison engine (the part of the budget that came from us) for a while now and there are first results:
6823 birthDate_gain_wiki.nt
3549 deathDate_gain_wiki.nt
362541 populationTotal_gain_wiki.nt
372913 total
We only took three properties for now and showed the gain where no Wikidata statement was available. birthDate/deathDate is already quite good. Details here: https://drive.google.com/file/d/1j5GojhzFJxLYTXerLJYz3Ih-K6UtpnG_/view?usp=…
Our plan here is to map all Wikidata properties to the DBpedia Ontology and then have the info to compare coverage of Wikidata with all infoboxes across languages.
- we will remove the text extraction part from the old proposal (which is here for you reference: https://meta.wikimedia.org/wiki/Grants:Project/DBpedia/CrossWikiFact). This will still be a focus during our work in 2018, together with Diffbot and the new DBpedia NLP department, but we think that it distracted from the core of the proposal. Results from the Wikipedia article text extraction can be added later once they are available and discussed separately.
- We proposed to make an extra website that helps to synchronize all Wikipedias and Wikidata with DBpedia as its backend. While the external website is not an ideal solution, we are lacking alternatives. The Primary Sources Tool is mainly for importing data into Wikidata, not so much synchronization. The MediaWiki instances of the Wikipedias do not seem to have any good interfaces to provide suggestions and pinpoint missing info. Especially to this part, we would like to ask for your help and suggestions, either per mail to the list or on the talk page: https://meta.wikimedia.org/wiki/Grants_talk:Project/DBpedia/GlobalFactSync
We are looking forward to a fruitful collaboration with you and we thank you for your feedback!
All the best
Magnus
--
Magnus Knuth
Universität Leipzig
Institut für Informatik
Abt. Betriebliche Informationssysteme, AKSW/KILT
Augustusplatz 10
04109 Leipzig DE
mail: knuth(a)informatik.uni-leipzig.de
tel: +49 177 3277537
webID: http://magnus.13mm.de/
Hi everyone, I wanted to follow up with the results from the FAST interest survey (I posted to this list in November and December last year). The survey closed on 15 December 2017. That survey was developed by representatives of five research libraries - Brown, Columbia, Cornell, Harvard, and Yale - to gauge interest in the current features of OCLC's FAST (Faceted Application of Subject Terminology) and potential enhancements.
We received 586 responses to the survey. The "FAST Five" have prepared the attached report summarizing the results, conclusions, and next steps. It will be posted on the same lists as the original survey link was posted, so you may see copies of the report on other lists you subscribe to.
Many thanks to all of you who responded to the survey!
Posted on behalf of the FAST Five:
Brown (Boaz Nadav-Manes)
Columbia (Kate Harcourt)
Cornell (Jim LeBlanc)
Harvard (Scott Wicks)
Yale (Marty Kurth)
Merrilee
--
Merrilee Proffitt
OCLC * Senior Program Officer, OCLC Research Library Partnership
155 Bovet Rd, Suite 500, San Mateo, CA 94402
T +1-650-287-2136 M +1-510-684-4717
[OCLC]<http://www.oclc.org/home.en.html?cmpid=emailsig_logo>
OCLC.org<http://www.oclc.org/home.en.html?cmpid=emailsig_link> * Facebook<http://www.facebook.com/pages/OCLC/20530435726> * Twitter<http://twitter.com/MerrileeIam> * YouTube<http://www.youtube.com/OCLCvideo>
The OCLC Research Library Partnership<http://www.oclc.org/research/partnership/> provides a unique transnational
collaborative network of peers to address common issues as well
as the opportunity to engage directly with OCLC Research. Join with us!
From: Proffitt,Merrilee
Sent: Monday, December 11, 2017 10:27 AM
To: 'Wikimedia & GLAM collaboration [Public]' <glam(a)lists.wikimedia.org>; 'Discussion list for the Wikidata project.' <wikidata(a)lists.wikimedia.org>; 'Wikimedia & Libraries' <libraries(a)lists.wikimedia.org>
Subject: RE: FAST Interest Survey
A final reminder! This survey wraps up on Friday.
Colleagues -
Five Research Libraries - Brown, Columbia, Cornell, Harvard and Yale - along with OCLC colleagues have created a survey to gauge interest in the current features of OCLC's FAST<http://www.oclc.org/research/themes/data-science/fast.html> (Faceted Application of Subject Terminology) and potential enhancements. We'd like as many responses as possible--even from those who are not using FAST! Please share this link with anyone you know who applies subject terminology to metadata.
https://www.surveymonkey.com/r/Next_Steps_for_FAST
We've beta-tested the survey, and it takes 5-15 minutes to complete.
You will likely see the survey on other lists that you subscribe to. Kate Harcourt at Columbia has volunteered to serve as coordinator for the survey, so please direct any questions about it to her at harcourt(a)columbia.edu<mailto:harcourt@columbia.edu>
The deadline for responses is 15 December 2017.
Thank you
Posted on behalf of:
Kate Harcourt
Marty Kurth
Jim LeBlanc
Boaz Nadav Manes
Scott Wicks
--
Merrilee Proffitt
OCLC * Senior Program Officer, OCLC Research
155 Bovet Rd, Suite 500, San Mateo, CA 94402
T +1-650-287-2136
[OCLC]<http://www.oclc.org/home.en.html?cmpid=emailsig_logo>
OCLC.org<http://www.oclc.org/home.en.html?cmpid=emailsig_link> * Facebook<http://www.facebook.com/pages/OCLC/20530435726> * Twitter<http://twitter.com/oclc> * YouTube<http://www.youtube.com/OCLCvideo>
**Apologies for cross-posting*
Call for workshops and tutorials
In conjunction with EKAW 2018
https://project.inria.fr/ekaw2018/
November 12 - 16 Nancy, France
Introduction
The International Conference on Knowledge Engineering and Knowledge Management (EKAW) is concerned with all aspects of eliciting, acquiring, modelling and managing knowledge, as well its role in the construction of knowledge-intensive systems and services for the semantic web, knowledge management, e-business, natural language processing, intelligent information integration, etc. This year, EKAW will pay special attention to topics related to knowledge and artificial intelligence.
Besides the regular conference tracks, EKAW will host a number of workshops and tutorials on topics related to the theme of the conference. We hope our workshops to provide an informal setting where participants have the opportunity to discuss specific technical topics in an atmosphere that fosters the active exchange of ideas; and tutorials to enable attendees to fully appreciate current issues, main schools of thought, and possible application areas.
Topics of Interest
In order to meet these goals, workshop/tutorial proposals should address topics that satisfy the following criteria:
• the falls in the general scope of EKAW 2018;
• there is a clear focus on a specific technology, problem or application;
• there is a sufficiently large community interested in the topic.
Submission Guidelines
Proposals should be submitted via EasyChair at: https://easychair.org/conferences/?conf=ekaw2018wst Submissions should be a single PDF file of no more than 5 pages, specifying "Workshop Proposal" or "Tutorial Proposal", and should contain the following information.
Workshop proposals:
• Title.
• Abstract (200 words).
• Motivation on why the topic is of particular interest at this time and its relation to the main conference topics.
• Workshop format, discussing the mix of events such as paper presentations, invited talks, panels, and general discussion.
• Intended audience and expected number of participants.
• List of (potential) members of the program committee (at least 50% have to be confirmed at the time of the proposal, confirmed participants should be marked specifically).
• Indication of whether the workshop should be considered for a half-day or full-day.
• The tentative dates (submission, notification, camera-ready deadline, etc.)
• Past versions of the workshop, including URLs as well as number of submissions and acceptance rates.
• Details of the organisers (name, affiliation, email address, homepage) and short CV.
• We strongly advise having more than one organiser, preferably from different institutions, bringing different perspectives to the workshop topic.
• We welcome, and will prioritise, workshops with creative structures and organisations that attract various types of contributions and ensure rich interactions.
Tutorial proposals:
• Title.
• Abstract (200 words).
• Relation to the conference topics, i.e. why it will be of interest to the conference attendants.
• If the tutorial, or a very similar tutorial, has been given elsewhere, explanation of the benefit of presenting it again to the EKAW community.
• Overview of content, description of the aims, presentation style, potential/preferred prerequisite knowledge.
• Indication on whether the tutorial should be considered for a half-day or full-day.
• Intended audience and expected number of participants.
• Audio-visual or technical requirements and any special room requirements (for hands-on sessions, any software needed and download sites must be provided by the tutorial presenters).
• Details of the presenters (name, affiliation, email address, homepage) and short CV including also their expertise, experiences in teaching and in tutorial presentation.
Workshop Organiser Responsibilities
The organisers of accepted workshops are expected to:
• prepare a workshop webpage (linked to the official EKAW website) containing the call for papers and detailed information about the workshop organisation and timelines.
• be responsible for the workshop publicity.
• be responsible for their own reviewing process, decide upon the final program content and report the number of submissions and accepted papers to the workshop chair.
• be responsible for publishing electronic proceedings (e.g. on the CEUR-WS website).
• ensure workshop participants are informed they have to register to the main conference and the workshop.
• schedule, attend and coordinate their entire workshop.
Tutorial Organisers Responsibilities
The proposers of accepted tutorials are expected to
• prepare a tutorial webpage (linked to the official EKAW website) containing detailed information about the tutorial
• prepare the tutorial materials
• publicity distribute materials to participants
• schedule, attend and coordinate their tutorial.
Important Dates
• Proposals due: 9 May, 2018
• Notifications: 23 May 2018
Suggested Timeline for Workshops
• Workshop website up and calls: 1 June 2018
• Deadline to submit Papers to Workshops: 8 September 2018
• Acceptance of Papers for Workshops: 1 October 2018
• Workshop days: 12 or 13 November 2018
Chairs
• Manuel Atencia, Université Grenoble Alpes & Inria, France
• Marieke van Erp, KNAW Humanities Cluster, The Netherlands
Contact:
ekaw2018wst(a)easychair.org
--
Digital Humanities Lab
KNAW Humanities Cluster
http://www.mariekevanerp.com
Hello all,
As every year, Google Summer of Code <https://summerofcode.withgoogle.com/>
support student developers all around the world to work on projects.
Wikimedia as an organization is part of the mentors.
A list of projects
<https://www.mediawiki.org/wiki/Google_Summer_of_Code/2018> is already
available, you can also add your own. One of these projects is signed
statements for Wikidata: developing the technical feature that will allow
institutions to donate verified and sourced information and increase the
quality of the data.
If you want to participate or become a mentor, feel free to check the
information on the pages linked above.
Thanks,
--
Léa Lacroix
Project Manager Community Communication for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hello all,
We're currently encountering some problems with the mailing-list, I
apologize for this.
The Weekly Summary of this week is available here:
https://www.wikidata.org/wiki/Wikidata:Status_updates/2018_02_12
--
Léa Lacroix
Project Manager Community Communication for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Greetings,
As the subject line says, there will be a Wikimedia Foundation-hosted IRC
office hours [0] this coming Tuesday, 13 February 2018, from 18:00-19:00
UTC [1]. The topic is Structured Data on Commons, and the subjects are
mainly whatever those who attend would like to discuss. The Structured Data
hub has information about what the development team has been up this past
year as well as upcoming plans [2] for those who might like to prepare or
find interesting things to talk about.
The Structured Data team also issues a newsletter every few months. You can
subscribe to have it delivered to a talk page, receive a notification
instead delivery, and read past issues. Find out more on Meta [3].
I'll be sending out a reminder email a few hours before this occurs in one
week's time.
Thank you for your time, I hope to see you there.
0. https://meta.wikimedia.org/wiki/IRC_office_hours
1. To check your local date and time for the office hours <
https://www.timeanddate.com/worldclock/fixedtime.html?hour=18&min=00&sec=0&…
>
2. https://commons.wikimedia.org/wiki/Commons:Structured_data/Development
3.
https://meta.wikimedia.org/wiki/Global_message_delivery/Targets/Structured_…
--
Keegan Peterzell
Technical Collaboration Specialist
Wikimedia Foundation
*Apologies for cross-posting*
---------------------------------------------------------------------------------
Call For Papers
---------------------------------------------------------------------------------
1st International Workshop on Deep Learning for Knowledge Graphs
and Semantic Technologies (DL4KGs)
http://usc-isi-i2.github.io/DL4KGS/
In conjunction with ESWC 2018, 3rd-7th June 2018, Heraklion, Crete, Greece
---------------------------------------------------------------------------------
Workshop Overview
---------------------------------------------------------------------------------
Semantic Web technologies and deep learning share the goal of creating intelligent artifacts that emulate human capacities such as reasoning, validating, and predicting. There are notable examples of contributions leveraging either deep neural architectures or distributed representations learned via deep neural networks in the broad area of Semantic Web technologies. Knowledge Graphs (KG) are one of the most well-known outcomes from the Semantic Web community, with wide use in web search, text classification, entity linking etc. KGs are large networks of real-world entities described in terms of their semantic types and their relationships to each other. Most famous examples of KGs are: DBpedia, Wikidata and Yago.
A challenging but paramount task for problems ranging from entity classification to entity recommendation or entity linking is that of learning features representing entities in the knowledge graph (building knowledge graph embeddings ) that can be fed into machine learning algorithms. The feature learning process ought to be able to effectively capture the relational structure of the graph (i.e. connectivity patterns) as well as the semantics of its properties and classes, either in an unsupervised way and/or in a supervised way to optimize a downstream prediction task. In the past years, Deep Learning (DL) algorithms have been used to learn features from knowledge graphs, resulting in enhancements of the state-of-the-art in entity relatedness measures, entity recommendation systems and entity classification. DL algorithms have equally been applied to classic problems in semantic applications, such as (semi-automated) ontology learning, ontology alignment, duplicate re!
cognition, ontology prediction, relation extraction, and semantically grounded inference.
---------------------------------------------------------------------------------
Topics of Interest
---------------------------------------------------------------------------------
Topics of interest for this first workshop on Deep Learning for Knowledge Graphs and Semantic Technologies, include but are not limited to the following fields and problems:
Knowledge graph embeddings for entity linking, recommendation, relatedness
Knowledge graph embeddings for link prediction and validation
Time-aware and scalable knowledge graph embeddings
Text-based entity embeddings vs knowledge graph entity embeddings
Deep learning models for learning knowledge representations from text
Knowledge graph agnostic embeddings
Knowledge Base Completion
Type Inference
Question Answering
Domain Specific Knowledge Base Construction
Reasoning over KGs and with deep learning methods
Neural networks and logic rules for semantic compositionality
Quality checking and Data cleaning
Multilingual resources for neural representations of linguistics
Commonsense reasoning and vector space models
Deep ontology learning
Deep learning ontological annotations
Applications of knowledge graph embeddings in real business scenarios
------------------------------------------------------------------------------------
Important Dates
------------------------------------------------------------------------------------
Submission deadline: Friday March 16, 2018
Notification of Acceptance: Tuesday April 17, 2018
Camera-ready Submission: Tuesday April 24, 2018
---------------------------------------------------------------------------------
WORKSHOP CO-CHAIRS
---------------------------------------------------------------------------------
Michael Cochez, Fraunhofer Institute for Applied Information Technology, Germany
Thierry Declerck, DFKI GmbH, Germany
Gerard de Melo, Rutgers University, USA
Luis Espinosa Anke, Cardiff University, UK
Besnik Fetahu, L3S Research Center, Leibniz University of Hannover, Germany
Dagmar Gromann, Technical University Dresden, Germany
Mayank Kejriwal, Information Sciences Institute, USA
Maria Koutraki, FIZ-Karlsruhe, Karlsruhe Institute of Technology (KIT), Germany
Freddy Lecue, Accenture Technology Labs, Ireland; INRIA, France
Enrico Palumbo, ISMB, Italy; EURECOM, France; Politecnico di Torino, Italy
Harald Sack, FIZ Karlsruhe, Karlsruhe Institute of Technology (KIT), Germany
More information about DL4KGs 2018 is available at: http://usc-isi-i2.github.io/DL4KGS/
?
Kaya
I like to announce the development of a new QR code creator. The original
idea for the upgrade came from Wikimedia Australia and its work with
Toodyay WikiTown project.
Wikimedia Australia has setup out to create two new independent resources
for the Qrpedia project with the support of Wikimedia UK who first
developed the QRpedia concept back 2011 and presented it to the community
in 2012.
The two concepts are;
- a new qr code creator that uses the stable Wikidata item url.
- the reason for for this was to firstly avoid what turned out to be a
costly exercise for Toodyay when some articles were renamed in a
couple of
instances just breaking the qr link to wikipedia, in another instance
sending the reader to an unrelated article
- this also gives us the opportunity to link directly to the other
projects
- and connect all of that in a two way exchange of knowledge with the
numerous 3rd GLAM's through their unique identifiers
- the concept is now to create a qr reader thats based within the
Wikimedia Foundation family
- this first and foremost will ensure that our readers aren't being
directed to third party services that capture user data, and use
it to spam
them with usual set internet advertising, scams, and phishing
- a feature that is significant to GLAMs, Education institutions,
and ourselves
- this reader will enable user to connect with all of our information
whether be Wikipedia articles, Quotes, diaries & journals(Wikisource), or
Wikivoyage and its itineraries.
- I think there is also the potential for Education/Wikiversity to
develop learning programs connecting real life observations to
WMF content.
- For GLAMs it'll be able to provide a safe reliable way for links to
and from their collections
- all of this within a multi lingual environment free from
advertising
I'd like to thank Dave for his efforts in turning a discussion at a Perth
meetup into reality. I also acknowledge the WMF & WMAU for supporting me
on a detour post WMCONF 2017 to discuss the concept with the WMUK and the
some of the people involved in the development of the original project,
those discussion were invaluable getting to this point.
I looked forward to the next part of the project and hope that people will
come forward with suggestions, requests, and help so that as a community we
take an even bigger step in sharing the sum of all knowledge
Regards
Gideon Digby(Gnangarra)
Vice President Wikimedia Australia
Noongarpedia: https://incubator.wikimedia.org/wiki/Wp/nys/Main_Page
WMAU: http://www.wikimedia.org.au/wiki/User:Gnangarra
---------- Forwarded message ----------
From: Meta <wiki(a)wikimedia.org>
Date: 10 February 2018 at 17:43
Subject: Meta email from user "Evad37" - Free Knowledge Portal
To: Gnangarra <gnangarra(a)gmail.com>
Hi, can you please forward the following to Wikimedia-l and other places of
interest, per our discussion this afternoon.
Thanks, David (User:Evad37)
---------------
New tool "The Free Knowledge Portal"
Hi all,
I've created a new tool, The Free Knowledge Portal,[0] that is a solution to
the Community Wishlist Survey proposal "Qr codes for all items".[1]
The basic idea is to provide stable urls that showcase a Wikidata item's
sitelinks, and related items.
The tool also lets you generate QR codes that link to those urls, and so is
like a successor to QRpedia: proving easy access our projects' pages via QR
codes, but for all Wikimedia projects, not just Wikipedia.
And it is multi-lingual - it will detect the device language and serve
sitelinks for that language (i.e. enwiki if the langauge is English, frwiki
if
the language is French, etc).
Also, it designed to be backwards-compatible with existing QRpedia codes, by
using a page title and site to determine the relevant Wikidata item id. (But
of course its up to the QRpedia people to redirect the codes to these urls)
Examples:
*Boston (Q100), using your devices language:
** https://tools.wmflabs.org/portal/Q100
*Boston (Q100), using French:
** https://tools.wmflabs.org/portal/Q100/fr
*Boston (Q100), using Spanish:
** https://tools.wmflabs.org/portal/Q100/es
*Backwards-compatible url for Boston on English Wikipedia
** https://tools.wmflabs.org/portal/?title=Boston&site=enwiki
The following features already work:
*QR code generator
*Portal page with
**Item label
**Item description
**Wikimedia sitelinks
**Related items (from 'What links here')
**Neary items (if the item has coordinates)
I'm also planning to show links to external identifiers, if the item has
any.
Another feature that would be nice would be to keep a record page visits,
which could then be visualised into graphs and the like. I'm not entirely
sure how this should be done, so any advice or help people could offer would
be appreciated.
(I'm thinking of making a mysql table with just a couple of
columns for item id and date-timestamp of visit, where each visit would be
recorded in a row. Then you could get page visits by doing a query that
counts the number of rows with a matching item and a date-timestamp within
a given range.)
I could also use some help with i18n/translations (what I've got at the
moment
has come from Wikidata labels and Google translate, which is far from ideal)
Anyway, suggestion, other feedback, and code patches would be appreciated:
either directly on Github[2], or on the Meta page which I've just recently
created.[3]
Cheers, David (User:Evad37)
-- Links --
[0] https://tools.wmflabs.org/portal
[1] https://meta.wikimedia.org/wiki/2017_Community_Wishlist_
Survey/Wikidata/Qr_codes_for_all_items
[2] https://github.com/evad37/wm-portal
[3] https://meta.wikimedia.org/wiki/Free_Knowledge_Portal
--
This email was sent by Evad37 to Gnangarra by the "Email this user"
function at Meta. If you reply to this email, your email will be sent
directly to the original sender, revealing your email address to them.
Dear Wikidata Community,
funded by the OpenKnowledge Foundation's Prototype Fund[1] I am working
on a project called OpenMetroMaps[2] since last September. The project
is an open-source and open-data project. I develop free software,
libraries and data formats for storing, editing and processing data
about transit networks and for designing schematic maps. The idea is to
provide the infrastructure for local communities to collect free data
about their transit network and to enable them to make use of such data
with a current focus on schematic network maps.
The project is concerned with data collected and topics addressed in
both the Wikidata and the OpenStreetMap projects. Albeit there are
license incompatiblities involved preventing direct import of data from
one project into the other, both projects might benefit from tools that
visualize and validate their data.
Since I am currently working on the project mostly alone, I have
organized a meeting with the intention to meet other people interested
in the topic. I would like to discuss ideas about where to go with the
project and problems that should be addressed in the Wikimedia family
of projects as well as the OpenStreetMap when it comes to data about
public transport. It will take place on February 11 from 14h to
16:30h in the Berlin office of Wikmedia Germany[3]. Please let me know
if you would like to attend so that I roughly estimate how many people
are going to participate.
You can also read the invitation on the German OpenStreetMap mailing
list[4] for some more details.
I hope to see you on Sunday,
Sebastian
[1] https://prototypefund.de
[2] http://www.openmetromaps.org
[3] https://de.wikipedia.org/wiki/Wikipedia:TU23
[4] https://lists.openstreetmap.org/pipermail/talk-de/2018-February/114636.html