Apologies for cross-posting
========================================================
13th ESWC 2016
http://2016.eswc-conferences.org/call-challenges
Call for Semantic Web Challenges Entries
Open Knowledge Extraction (OKE) Challenge
Challenge on Semantic Sentiment Analysis
Conference Live app Challenge
Open Challenge on Question Answering over Linked Data
Top-K Shortest Path in Large Typed RDF Graphs Challenge
Semantic Publishing Challenge
schema.org - Bonus Challenge
========================================================
OVERVIEW
The 13th ESWC, to be held from May 29th to June 2nd in Heraklion, Crete,
features no less than seven challenges this year!
The purpose of the challenges is to showcase the maturity of state of
the art methods and tools on tasks common to the Semantic Web community
and adjacent disciplines, in a controlled setting involving rigorous
evaluation.
Semantic Web Challenges are an official track of the conference,
ensuring significant visibility for the challenges as well as
participants. Challenge participants are asked to present their
submissions as well as provide a paper describing their work. The
details of the submissions may vary per challenge and will be found in
the individual calls. These papers must undergo a peer-review by experts
relevant to the challenge task, and will be published in the official
ESWC2016 Satellite Events proceedings.
IMPORTANT DATES
Individual challenges may deviate from these dates but as a rule the
following dates apply:
* Training data ready and challenges Calls for Papers sent: Friday
January 15th, 2016
* Challenge papers submission deadline: Friday March 11th, 2016
* Challenge paper reviews due: Tuesday April 5th, 2016
* Notifications sent to participants and invitations to submit task
results: Friday April 8th, 2016
* Test data (and other participation tools) published: Friday April 8th,
2016
* Camera ready papers due: Sunday April 24th, 2016
CHALLENGES AT A GLANCE
Open Knowledge Extraction (OKE) Challenge
The OKE challenge, launched as first edition at last year Extended
Semantic Web Conference, ESWC2015, has the ambition to provide a
reference framework for researc on Knowledge Extraction from text for
the Semantic Web by re-defining a number of tasks (typically from
information and knowledge extraction), taking into account specific SW
requirements.
http://2016.eswc-conferences.org/eswc-16-open-knowledge-extraction-oke-chal…
Challenge on Semantic Sentiment Analysis
Social media evolution has given users one important opportunity for
expressing their thoughts and opinions online. The information thus
produced is related to many different areas such as commerce, tourism,
education, health and causes the size of the Social Web to expand
exponentially.
http://2016.eswc-conferences.org/eswc-16-challenge-semantic-sentiment-analy…
Conference Live app Challenge
In the past two years the Extended Semantic Web Conference (ESWC) has
provided a semantic Web application to browse conference data. The
application, called Conference Live, is a Web and mobile application
based on conference data from the Semantic Web Dog Food server, which
provides facilities to browse papers and authors at a specific conference.
http://2016.eswc-conferences.org/eswc-16-conference-live-app-challenge
6th Open Challenge on Question Answering over Linked Data (QALD-6)
The past years have seen a growing amount of research on question
answering over Semantic Web data, shaping an interaction paradigm that
allows end users to profit from the expressive power of Semantic Web
standards while at the same time hiding their complexity behind an
intuitive and easy-to-use interface. The Question Answering over Linked
Data challenge provides an up-to-date benchmark for assessing and
comparing systems that mediate between a user, expressing his or her
information need in natural language, and RDF data.
http://2016.eswc-conferences.org/6th-open-challenge-question-answering-over…
Top-K Shortest Path in Large Typed RDF Graphs Challenge
The advent of SPARQL 1.1 introduced property paths as a new graph
matching paradigm that allows the employment of Kleene star * (and it's
variant +) unary operators to build SPARQL queries that are agnostic of
the underlying RDF graph structure. The ability to express path patterns
that are agnostic of the underlying graph structure is certainly a step
forward.
http://2016.eswc-conferences.org/top-k-shortest-path-large-typed-rdf-graphs…
Semantic Publishing Challenge 2016 – Assessing the Quality of Scientific
Output in its Ecosystem
This is the next iteration of the successful Semantic Publishing
Challenge of ESWC 2014 and 2015. We continue pursuing the objective of
assessing the quality of scientific output, evolving thedataset
bootstrapped in 2014 and 2015 to take into account the wider ecosystem
of publications.
http://2016.eswc-conferences.org/assessing-quality-scientific-output-its-ec…schema.org - Bonus Challenge
Rather than create a separate schema.org challenge, we encourage where
appropriate submissions to other ESWC2016 challenges to consider also
exploring schema.org's relationship with Linked Data and Semantic Web
tools, technologies, vocabularies and datasets.
http://2016.eswc-conferences.org/bonus-challenge
CONTACT
ESWC 2016 Challenge Chairs
* Stefan Dietze, L3S Research Center, Germany (dietze(a)l3s.de)
* Anna Tordai, Elsevier, Netherlands (a.tordai(a)elsevier.com)
--
Prof. Dr. Heiko Paulheim
Data and Web Science Group
University of Mannheim
Phone: +49 621 181 2661
B6, 26, Room C1.09
D-68159 Mannheim
Mail: heiko(a)informatik.uni-mannheim.de
Web: www.heikopaulheim.com
Hi Magnus,
>> >I was aware of the Sourcerer tool: I'm concerned with those references
>> >coming from Wikipedia articles though, since they stem from inside a
>> >Wikimedia project, and I want to make sure that everything comes from
>> >the outside.
>> >
> The Sourcerer references do NOT come from Wikipedia! I am using third-party
> sites for which we already have IDs (e.g. GND) to auto-validate values, and
> add the appropriate reference if identical. Basically, what you want to do,
> on the cheap;-)
Whoops, sorry, I was probably confused by:
1. the Sourcerer **user script** [1], which claims to "get a list of all
external links from all language editions of Wikipedia for an item";
2. the **Sourcery** tool [2], which claims to load "all URLs in all
associated Wikipedia pages".
So, third-party references indeed, but still curated by Wikipedians, right?
That's what I meant, I should have been more specific in my concern.
Is it correct that the Sourcerer **bot** is a different thing, or am I
getting this completely wrong?
Cheers,
Marco
[1] https://www.wikidata.org/wiki/Wikidata:Tools/User_scripts#Sourcerer
[2] https://tools.wmflabs.org/wikidata-todo/sourcery.html
Hi everyone,
The curated list of biographical sources for StrepHit has now passed the
objective of 40 items [1].
Your help in validating the list is essential to ensure the reliability
of the corpus that will be collected upon it.
In practice, are the sources:
1. *reliable* (cf. [2])?
2. *third-party*, i.e., not created by users of Wikimedia projects?
I kindly ask you to answer those questions in the discussion page of [1].
Cheers,
Marco
[1]
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
[2] https://en.wikipedia.org/wiki/Wikipedia:Verifiability#Reliable_sources
Wikidata SPARQL aficionados,
This SPARQL query worked for several weeks, but quit working a few days
ago:
PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
PREFIX wikibase: <http://wikiba.se/ontology#>
PREFIX entity: <http://www.wikidata.org/entity/>
PREFIX p: <http://www.wikidata.org/prop/direct/>
SELECT ?propUrl ?propLabel ?valUrl ?valLabel
WHERE {
hint:Query hint:optimizer 'None' .
entity:Q42 ?propUrl ?valUrl .
?valUrl rdfs:label ?valLabel
FILTER (LANG(?valLabel) = 'en') .
?property ?ref ?propUrl .
?property a wikibase:Property .
?property rdfs:label ?propLabel
FILTER (lang(?propLabel) = 'en' )
}
ORDER BY ?propUrl ?valUrl LIMIT 100
Here is the first part of the exception:
java.util.concurrent.ExecutionException:
java.util.concurrent.ExecutionException:
org.openrdf.query.QueryEvaluationException: java.lang.RuntimeException:
java.util.concurrent.ExecutionException: java.lang.RuntimeException:
java.util.concurrent.ExecutionException: java.lang.Exception:
task=ChunkTask{query=7be8b715-1429-4348-bf97-237d06d6084c,bopId=17,partitionId=-1,sinkId=19,altSinkId=null},
cause=java.util.concurrent.ExecutionException:
java.lang.RuntimeException:
com.bigdata.rdf.internal.NotMaterializedException: Vocab(6)
Any ideas on how I can modify the query to work again?
Please advise,
James Weaver
CoDeS 2016
International Workshop on
Completing and Debugging the Semantic Web
May 29 or 30, 2016
Heraklion, Greece
http://www.ida.liu.se/~patla00/conferences/CoDeS16/
co-located with ESWC 2016 (http://2016.eswc-conferences.org/)
---------------------------------------------------------------------
Developing ontologies and Semantic Web data sets is not an easy task
and, as the ontologies and data sets grow in size, they are likely to
show a number of defects (wrong information as well as omissions). Such
ontologies and data sets, although often useful, also lead to problems
when used in semantically-enabled applications. Wrong conclusions may be
derived or valid conclusions may be missed. Further, during the recent
years, more and more mappings both between ontologies as well as
entities in the Linked Open Data cloud have been generated, e.g., using
ontology alignment and/or entity linking systems, forming a linked
network of data sets and ontologies. This has led to a new opportunity
to deal with defects, as links between datasets and ontologies may be
exploited for debugging, or the interlinks between them. On the other
hand, it also has introduced a new difficulty as the mappings may not
always be correct and need to be debugged themselves. Also the linked
data level may contain wrong information and omissions in the data as
well as the links.
Defects can take different forms. Syntactic defects are usually easy to
find and to resolve. Defects regarding style include such things as
unintended redundancy. More interesting and severe defects are the
modeling defects which require domain knowledge to detect and resolve
such as defects in the structure, and semantic defects such as
unsatisfiable concepts and inconsistent ontologies.
Topics of interest
This workshop intends to be a forum where issues related to completing
and debugging the Semantic Web are discussed. Topics of interest
include, but are not limited to:
* detecting and repairing defects in ontologies
* detecting and repairing defects in linked data
* detecting and repairing defects in mappings
* debugging and completing ontology networks and modular ontologies
* ontology alignment and repair of alignments
* justifications, belief revision, ontology patterns for debugging
* user support for ontology debugging and completion
* visualization and user interfaces for debugging and completion
* case studies
Important dates
* Submission: March 4, 2016
* Notification: April 1, 2016
* Camera-ready: April 15, 2016
* Workshop: May 29 or 30, 2016
Hi Daniel,
Thanks for getting in touch and for the useful information.
After 2 rounds of feedback from the community via StrepHit's
dissemination activities, I have opted for the biographical domain.
Cf.
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
The biomedical domain is next in line: while it is out of scope for the
project time frame (6 months), I believe it will be an excellent use
case in case of an extension.
Thanks again.
Cheers,
Marco
> Date: Sat, 23 Jan 2016 23:00:26 +0100
> From: Daniel Mietchen<daniel.mietchen(a)googlemail.com>
> To: "Discussion list for the Wikidata project."
> <wikidata(a)lists.wikimedia.org>
> Subject: Re: [Wikidata] [REMINDER] StrepHit IEG project kick-off
> seminar
> Message-ID:
> <CAN6n2b1wgiTi8quWO4fE=yBaFdkiad3kHz2hWsG-5QDMwEpERg(a)mail.gmail.com>
> Content-Type: text/plain; charset=UTF-8
>
> Thanks, Marco - I just watched the video and would be interested in
> knowing whether you have picked your focus by now from the three
> domains you suggested (biographies, companies, biomedical) or perhaps
> something else. If you go for biomedical, there would be overlap with
> https://www.wikidata.org/wiki/Wikidata:WikiProject_Source_MetaData
> and
> https://www.wikidata.org/wiki/Wikidata:WikiProject_Medicine
> and probably also
> https://www.wikidata.org/wiki/User:ProteinBoxBot ,
> all of which have a number of active people who I'd expect to be
> interested in a test run of your pipeline on biomedical topics.
>
> In any case, I would welcome it if you would include PubMed into your
> set of reliable third-party sources. We already have about 17k
> Wikidata items with a PubMed ID (cf.
> http://tools.wmflabs.org/autolist/autolist1.html?q=claim%5B698%5D ),
> which are increasingly being used as references for Wikidata
> statements (cf.http://tinyurl.com/zhauolt ), but we have barely
> scratched the surface of what needs to be done here (the English
> Wikipedia alone has ca. 30k medical articles, many of them heavily
> referenced to sources indexed by PubMed).
>
> So far, the link of those 17k items to Wikipedia is just that almost
> all of the corresponding papers have been cited on some Wikipedia
> (some come also from Wikispecies, Wikisource or other Wikimedia
> sites), albeit not always using a PubMed ID (sometimes via DOI, PubMed
> Central ID or via a link or in some other way).
>
> Daniel
>
> On Fri, Jan 15, 2016 at 10:47 AM, Marco Fossati<fossati(a)spaziodati.eu> wrote:
>> >Hi everyone,
>> >
>> >the seminar will start in a few minutes.
>> >Cheers,
>> >
>> >Marco
>> >
>> >On 1/11/16 16:52, Marco Fossati wrote:
>>> >>
>>> >>Here is the link for the online streaming:
>>> >>https://youtu.be/uvfd_HmPOrc
>>> >>
>>> >>Cheers,
>>> >>
>>> >>Marco
>>> >>
>>> >>2016-01-11 16:11 GMT+01:00 Marco Fossati <fossati(a)spaziodati.eu
>>> >><mailto:fossati@spaziodati.eu>>:
>>> >>
>>> >> Dear all,
>>> >>
>>> >> This is a kind reminder for the upcoming StrepHit IEG project
>>> >> kick-off seminar.
>>> >> Schedule: 15 January 2016, 11:00 am
>>> >>
>>> >> **Important update:** the location has moved to downtown Trento.
>>> >> **New location:** Aula Grande - Fondazione Bruno Kessler, Via
>>> >> S.Croce 77, Trento, Italy -http://www.openstreetmap.org/way/67197096
>>> >>
>>> >> The seminar will be streamed online, a link will be shared as soon
>>> >> as it is available.
>>> >>
>>> >> See you in Trento!
>>> >> Cheers,
>>> >>
>>> >> Marco
>>> >>
>>> >> 2015-12-23 17:03 GMT+01:00 Marco Fossati <fossati(a)spaziodati.eu
>>> >> <mailto:fossati@spaziodati.eu>>:
>>> >>
>>> >>
>>> >> [Begging pardon if you read this multiple times]
>>> >>
>>> >> Hi everyone,
>>> >>
>>> >> I would like to announce with great pleasure the StrepHit IEG
>>> >> project kick-off seminar.
>>> >> Of course, you are all invited to attend.
>>> >>
>>> >> The event will be held in a special day: Wikipedia's birthday!
>>> >>
>>> >> Below you can find the details.
>>> >>
>>> >> Schedule: 15 January 2016, 11:00 am, Luigi Stringa Conference Room
>>> >> Location: Fondazione Bruno Kessler, Via Sommarive 18, Povo,
>>> >> Trento, Italy -http://www.openstreetmap.org/way/28933739
>>> >>
>>> >> Abstract: We kick-off StrepHit, a project funded by the
>>> >> Wikimedia Foundation through the Individual Engagement Grants
>>> >> program.
>>> >> StrepHit is a Natural Language Processing pipeline that
>>> >> understands human language, extracts facts from text and
>>> >> produces Wikidata statements with reference URLs.
>>> >> It will enhance the data quality of Wikidata by suggesting
>>> >> references to validate statements, and will help Wikidata become
>>> >> the gold-standard hub of the Open Data landscape.
>>> >>
>>> >> Link:
>>> >>
>>> >>https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
>>> >>
>>> >> Speaker's bio: Marco Fossati is a researcher with a double
>>> >> background in Natural Languages and Information Technologies. He
>>> >> works at the Data and Knowledge Management (DKM) research unit
>>> >> at Fondazione Bruno Kessler, Trento, Italy. He is member of the
>>> >> DBpedia Association board of trustees, founder and
>>> >> representative of its Italian chapter. He has interdisciplinary
>>> >> skills both in linguistics and in programming. His research
>>> >> focuses on bridging the gap between Natural Language Processing
>>> >> techniques and Large Scale Structured Knowledge Bases in order
>>> >> to drive the Web of Data towards its full potential.
>>> >>
>>> >> See you in Trento and long live Wikipedia!
>>> >> Cheers,
>>> >>
>>> >> Marco
>>> >>
>>> >>
>>> >>
>> >
>> >_______________________________________________
>> >Wikidata mailing list
>> >Wikidata(a)lists.wikimedia.org
>> >https://lists.wikimedia.org/mailman/listinfo/wikidata
Dear all,
This is a kind reminder for the upcoming StrepHit IEG project kick-off
seminar.
Schedule: 15 January 2016, 11:00 am
**Important update:** the location has moved to downtown Trento.
**New location:** Aula Grande - Fondazione Bruno Kessler, Via S.Croce 77,
Trento, Italy - http://www.openstreetmap.org/way/67197096
The seminar will be streamed online, a link will be shared as soon as it is
available.
See you in Trento!
Cheers,
Marco
2015-12-23 17:03 GMT+01:00 Marco Fossati <fossati(a)spaziodati.eu>:
> [Begging pardon if you read this multiple times]
>
> Hi everyone,
>
> I would like to announce with great pleasure the StrepHit IEG project
> kick-off seminar.
> Of course, you are all invited to attend.
>
> The event will be held in a special day: Wikipedia's birthday!
>
> Below you can find the details.
>
> Schedule: 15 January 2016, 11:00 am, Luigi Stringa Conference Room
> Location: Fondazione Bruno Kessler, Via Sommarive 18, Povo, Trento, Italy
> - http://www.openstreetmap.org/way/28933739
>
> Abstract: We kick-off StrepHit, a project funded by the Wikimedia
> Foundation through the Individual Engagement Grants program.
> StrepHit is a Natural Language Processing pipeline that understands human
> language, extracts facts from text and produces Wikidata statements with
> reference URLs.
> It will enhance the data quality of Wikidata by suggesting references to
> validate statements, and will help Wikidata become the gold-standard hub of
> the Open Data landscape.
>
> Link:
> https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
>
> Speaker's bio: Marco Fossati is a researcher with a double background in
> Natural Languages and Information Technologies. He works at the Data and
> Knowledge Management (DKM) research unit at Fondazione Bruno Kessler,
> Trento, Italy. He is member of the DBpedia Association board of trustees,
> founder and representative of its Italian chapter. He has interdisciplinary
> skills both in linguistics and in programming. His research focuses on
> bridging the gap between Natural Language Processing techniques and Large
> Scale Structured Knowledge Bases in order to drive the Web of Data towards
> its full potential.
>
> See you in Trento and long live Wikipedia!
> Cheers,
>
> Marco
>
Wikidata is broken.
It has allsorts of kludges to link items to the items they should be in.
Lets merge all these Categories, lists, duplicate articles in a different
language etc. with the items they should have been part of from the
beginning. (Normal rules apply - If you were to describe the Subject of
each using properties could they be the same properties)
We now have the tool to fix the wikibase client problem - Badges!
If duplicates have suitable badges then wikibase client can use these to
keep the language links. wikibase client will also be able to add other
lists
"articles on the wiki in other languages" for instance
joe