Hi All,
I created a tool to extract translations from different editions of
Wiktionary. Right now it supports 39 different Wiktionaries. It only
extracts translations and ignores the rest.
Supported Wiktionaries:
Azerbaijani, Bulgarian, Catalan, Czech, Danish, Greek, English, Esperanto,
Spanish, Estonian, Basque, Finnish, French, Galician, Hebrew, Croatian,
Hungarian, Indonesian, Italian, Georgian, Latin, Lithuanian, Malagasy,
Dutch, Norwegian, Occitan, Polish, Portuguese, Romanian, Russian, Slovak,
Slovenian, Serbian, Swedish, Swahili, Turkish, Ukrainian, Vietnamese and
Chinese.
Adding a new Wiktionary is done via a configuration file.
Right now the beta version is available for download at:
https://github.com/juditacs/wikt2dict
Documentation is in progress, until then the README should be enough to get
started.
Please test it and send me your feedback and bug reports.
Thanks,
Judit Ács
To add up a couple of comments to what Denny said, from my experience with
Wikisource, reaching out to international, loosely connected communities is
already a big challenge on its own. I would like to invite Wiktionary
contributors to take a look to this Individual Engagement Grant project
that Aubrey and me are doing for Wikisource, because maybe it would make
sense that a group of involved Wiktionarians started a similar initiative
for Wiktionary. The original application can be found here:
http://meta.wikimedia.org/wiki/Grants:IEG/Elaborate_Wikisource_strategic_vi…
And the midterm report:
http://meta.wikimedia.org/wiki/Grants:IEG/Elaborate_Wikisource_strategic_vi…
If anyone from the Wiktionary community wants to step forward, I would be
more than happy to share experiences and provide advice.
Cheers,
Micru
On Sat, Aug 10, 2013 at 3:30 AM, Denny Vrandečić <
denny.vrandecic(a)wikimedia.de> wrote:
> [Sorry for cross-posting]
>
> Yes, I agree that the OmegaWiki community should be involved in the
> discussions, and I pointed GerardM to our proposals whenever and
> discussions, using him as a liaison. We also looked and keep looking at the
> OmegaWiki data model to see what we are missing.
>
> Our latest proposal is different from OmegaWiki in two major points:
>
> * our primary goal is to provide support for structured data in the
> Wiktionaries. We do not plan to be the main resource ourselves, where
> readers come to in order to look up something, we merely provide structured
> data that a Wiktionary may or may not use. This parallels the role of
> Wikidata has with regards to Wikipedia. This also highlights the difference
> between Wikidata and OmegaWiki, since OmegaWiki's goal is "to create a
> dictionary of all words of all languages, including lexical, terminological
> and ontological information."
>
> * a smaller difference is the data model. Wikidata's latest proposal to
> support Wiktionary is centered around lexemes, and we do not assume that
> there is such a things as a language-independent defined meaning. But no
> matter what model we end up with, it is important to ensure that the bulk
> of the data could freely flow between the projects, and even though we
> might disagree on this issue in the modeling, it is ensured that the
> exchange of data is widely possible.
>
> We tried to keep notes on the discussion we had today: <
> http://epl.wikimedia.org/p/WiktionaryAndWikidata>
>
> My major take home message for me is that:
> * the proposal needs more visual elements, especially a mock-up or sketch
> of how it would look like and how it could be used on the Wiktionaries
> * there is no generally accepted place for a discussion that involves all
> Wiktionary projects. Still, my initial decision to have the discussion on
> the Wikidata wiki was not a good one, and it should and will be moved to
> Meta.
>
> Having said that, the current proposal for the data model of how to support
> Wiktionary with Wikidata seems to have garnered a lot of support so far. So
> this is what I will continue building upon. Further comments are extremely
> welcomed. You can find it here:
>
> <http://www.wikidata.org/wiki/Wikidata:Wiktionary>
>
> As said, it will be moved to Meta, as soon as the requested mockups and
> extensions are done.
>
> Cheers,
> Denny
>
>
>
>
>
> 2013/8/10 Samuel Klein <meta.sj(a)gmail.com>
>
> > Hello,
> >
> > > On Fri, Aug 9, 2013 at 6:13 PM, JP Béland <lebo.beland(a)gmail.com>
> wrote:
> > >> I agree. We also need to include the Omegawiki community.
> >
> > Agreed.
> >
> > On Fri, Aug 9, 2013 at 12:22 PM, Laura Hale <laura(a)fanhistory.com>
> wrote:
> > > Why? The question of moving them into the WMF fold was pretty much no,
> > > because the project has an overlapping purpose with Wiktionary,
> >
> > This is not actually the case.
> > There was overwhelming community support for adopting Omegawiki - at
> > least simply providing hosting. It stalled because the code needed a
> > security and style review, and Kip (the lead developer) was going to
> > put some time into that. The OW editors and dev were very interested
> > in finding a way forward that involved Wikidata and led to a combined
> > project with a single repository of terms, meanings, definitions and
> > translations.
> >
> > Recap: The page describing the OmegaWiki project satisfies all of the
> > criteria for requesting WMF adoption.
> > * It is well-defined on Meta http://meta.wikimedia.org/wiki/Omegawiki
> > * It describes an interesting idea clearly aligned with expanding the
> > scope of free knowledge
> > * It is not a 'competing' project to Wiktionaries; it is an idea that
> > grew out of the Wiktionary community, has been developed for years
> > alongside it, and shares many active contributors and linguiaphiles.
> > * It started an RfC which garnered 85% support for adoption.
> > http://meta.wikimedia.org/wiki/Requests_for_comment/Adopt_OmegaWiki
> >
> > Even if the current OW code is not used at all for a future Wiktionary
> > update -- and this idea was proposed and taken seriously by the OW
> > devs -- their community of contributors should be part of discussions
> > about how to solve the Wiktionary problem that they were the first to
> > dedicate themselves to.
> >
> > Regards,
> > Sam.
> >
> > _______________________________________________
> > Wikimedia-l mailing list
> > Wikimedia-l(a)lists.wikimedia.org
> > Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/wikimedia-l,
> > <mailto:wikimedia-l-request@lists.wikimedia.org?subject=unsubscribe>
> >
>
>
>
> --
> Project director Wikidata
> Wikimedia Deutschland e.V. | Obentrautstr. 72 | 10963 Berlin
> Tel. +49-30-219 158 26-0 | http://wikimedia.de
>
> Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e.V.
> Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
> der Nummer 23855 B. Als gemeinnützig anerkannt durch das Finanzamt für
> Körperschaften I Berlin, Steuernummer 27/681/51985.
> _______________________________________________
> Wikimedia-l mailing list
> Wikimedia-l(a)lists.wikimedia.org
> Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/wikimedia-l,
> <mailto:wikimedia-l-request@lists.wikimedia.org?subject=unsubscribe>
>
--
Etiamsi omnes, ego non
<wiktionary-l(a)lists.wikimedia.org>Hi,
If there is someone in Wikimania interested in participating in the talks
about the future support of Wiktionary in Wikidata, we will having a
discussion about the several proposals.
http://wikimania2013.wikimedia.org/wiki/Support_of_Wiktionary_in_Wikidata
Date : Saturday, 10 Aug, 11:30 am - 1:00 pm
Place: Y520 (block Y, 5th floor)
See you there,
Micru
Le 2013-08-09 13:04, Romaine Wiki a écrit :
> Are there much users from Wiktionary in Hong Kong? I do not think any
> of the Dutch users is, I can't say for others.
>
> I think it would be essential that this subject is discussed inside
> the wider Wiktionary community. To me the group of users
> participating
> is too narrow. Also is a mailing list not handy as most of the users
> from Wiktionary do not read that. I think a Wikt-community wide
> discussion is needed.
I agree, and I think meta would be the most obvious channel for such
a discussion.
As said in the previous email, there's already [[Wiktionary future]]
which is waiting for contributions and discussion on meta. Anyway,
whatever the canal, it would be realy important to make aware as
much contributors as possible aware of this initiative, so they can
provide relevant feedback specific to their needs.
>
> Romaine
>
>
> --------------------------------------------
> On Fri, 8/9/13, David Cuenca <dacuetu(a)gmail.com> wrote:
>
> Subject: [Wikidata-l] Meeting about the support of Wiktionary in
> Wikidata
> To: wiktionary-l(a)lists.wikimedia.org, "Wikimania general list (open
> subscription)" <wikimania-l(a)lists.wikimedia.org>, "Discussion list
> for
> the Wikidata project." <wikidata-l(a)lists.wikimedia.org>, "Wikimedia
> Mailing List" <wikimedia-l(a)lists.wikimedia.org>
> Date: Friday, August 9, 2013, 4:43 AM
>
> Hi,
>
>
> If there is someone in Wikimania interested in participating
> in the talks about the future support of Wiktionary in
> Wikidata, we will having a discussion about the several
> proposals.
>
>
>
>
> http://wikimania2013.wikimedia.org/wiki/Support_of_Wiktionary_in_Wikidata
>
> Date : Saturday, 10 Aug, 11:30 am - 1:00 pm
>
>
> Place: Y520 (block Y, 5th floor)
>
> See you there,
> Micru
>
>
> -----Inline Attachment Follows-----
>
> _______________________________________________
> Wikidata-l mailing list
> Wikidata-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
>
>
> _______________________________________________
> Wikidata-l mailing list
> Wikidata-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
--
Association Culture-Libre
http://www.culture-libre.org/
It has been proposed on Swedish Wiktionary to use a significant amount of
material from a dictionary [1] which allegedly is out of copyright due to
only having been protected under the law of "katalogskydd" ('catalogue
protection') which is in effect for 15 years after publication. The last
edition was published in 1989, the author died in 1986, and the last
reprinting was done in 1999 -- though mere reprints are not supposed to
grant extensions of the time of protection.
Assuming that the assessment about the material falling under the
"katalogskydd" is correct, is that sufficient for WMF to be comfortable
hosting the material, or does US copyright law interfere in any way?
\Mike
(cc-ing to Wiktionary-l to see if anyone there has experience with this
particular situation.)
[1] http://runeberg.org/svaraord/
Apologies for cross-posting!
=======================
NLP & DBpedia Workshop 2013
=======================
Free, open, interoperable and multilingual NLP for DBpedia and DBpedia
for NLP:
http://nlp-dbpedia2013.blogs.aksw.org/
Collocated with the International Semantic Web Conference 2013 (ISWC 2013)
21-22 October 2013, in Sydney, Australia (*Submission deadline July 8th*)
Please email us, if you need a deadline extension!
**********************************
Recently, the DBpedia community has experienced an immense increase in
activity and we believe, that the time has come to explore the
connection between DBpedia & Natural Language Processing (NLP) in a yet
unpreceded depth. The goal of this workshop can be summarized by this
(pseudo-) formula:
NLP & DBpedia == DBpedia4NLP && NLP4DBpedia
http://db0.aksw.org/downloads/CodeCogsEqn_bold2.gif
DBpedia has a long-standing tradition to provide useful data as well as
a commitment to reliable Semantic Web technologies and living best
practices. With the rise of WikiData, DBpedia is step-by-step relieved
from the tedious extraction of data from Wikipedia's infoboxes and can
shift its focus on new challenges such as extracting information from
the unstructured article text as well as becoming a testing ground for
multilingual NLP methods.
Contribution
=========
Within the timeframe of this workshop, we hope to mobilize a community
of stakeholders from the Semantic Web area. We envision the workshop to
produce the following items:
* an open call to the DBpedia data consumer community will generate a
wish list of data, which is to be generated from Wikipedia by NLP
methods. This wish list will be broken down to tasks and benchmarks and
a GOLD standard will be created.
* the benchmarks and test data created will be collected and published
under an open license for future evaluation (inspired by OAEI and
UCI-ML). An overview of the benchmarks can be found here:
http://nlp-dbpedia2013.blogs.aksw.org/benchmarks
Please sign up to our mailing list, if you are interested in discussing
guidelines and NLP benchmarking:
http://lists.informatik.uni-leipzig.de/mailman/listinfo/nlp-dbpedia-public
Important dates
===========
8 July 2013, Paper Submission Deadline
9 August 2013, Notification of accepted papers sent to authors
Motivation
=======
The central role of Wikipedia (and therefore DBpedia) for the creation
of a Translingual Web has recently been recognized by the Strategic
Research Agenda (cf. section 3.4, page 23) and most of the contributions
of the recently held Dagstuhl seminar on the Multilingual Semantic Web
also stress the role of Wikipedia for Multilingualism. As more and more
language-specific chapters of DBpedia appear (currently 14 language
editions), DBpedia is becoming a driving factor for a Linguistic Linked
Open Data cloud as well as localized LOD clouds with specialized domains
(e.g. the Dutch windmill domain ontology created from
http://nl.dbpedia.org ).
The data contained in Wikipedia and DBpedia have ideal properties for
making them a controlled testbed for NLP. Wikipedia and DBpedia are
multilingual and multi-domain, the communities maintaining these
resource are very open and it is easy to join and contribute. The open
license allows data consumers to benefit from the content and many parts
are collaboratively editable. Especially, the data in DBpedia is widely
used and disseminated throughout the Semantic Web.
NLP4DBpedia
==========
DBpedia has been around for quite a while, infusing the Web of Data with
multi-domain data of decent quality. These triples are, however, mostly
extracted from Wikipedia infoboxes. To unlock the full potential of
Wikipedia articles for DBpedia, the information contained in the
remaining part of the articles needs to be analysed and triplified.
Here, the NLP techniques may be of favour.
DBpedia4NLP
==========
On the other hand NLP, and information extraction techniques in
particular, involve various resources while processing texts from
various domains. These resources may be used e.g. as an element of a
solution e.g. gazetteer being an important part of a rule created by an
expert or disambiguation resource, or while delivering a solution e.g.
within machine learning approaches. DBpedia easily fits in both of these
roles.
We invite papers from both these areas including:
1. Knowledge extraction from text and HTML documents (especially
unstructured and semi-structured documents) on the Web, using
information in the Linked Open Data (LOD) cloud, and especially in DBpedia.
2. Representation of NLP tool output and NLP resources as RDF/OWL, and
linking the extracted output to the LOD cloud.
3. Novel applications using the extracted knowledge, the Web of Data or
NLP DBpedia-based methods.
The specific topics are listed below.
Topics
=====
- Improving DBpedia with NLP methods
- Finding errors in DBpedia with NLP methods
- Annotation methods for Wikipedia articles
- Cross-lingual data and text mining on Wikipedia
- Pattern and semantic analysis of natural language, reading the Web,
learning by reading
- Large-scale information extraction
- Entity resolution and automatic discovery of Named Entities
- Multilingual entity recognition task of real world entities
- Frequent pattern analysis of entities
- Relationship extraction, slot filling
- Entity linking, Named Entity disambiguation, cross-document
co-reference resolution
- Disambiguation through knowledge base
- Ontology representation of natural language text
- Analysis of ontology models for natural language text
- Learning and refinement of ontologies
- Natural language taxonomies modeled to Semantic Web ontologies
- Use cases for potential data extracted from Wikipedia articles
- Use cases of entity recognition for Linked Data applications
- Impact of entity linking on information retrieval, semantic search
Furthermore, an informal list of NLP tasks can be found on this
Wikipedia page:
http://en.wikipedia.org/wiki/Natural_language_processing#Major_tasks_in_NLP
These are relevant for the workshop as long as they fit into the
DBpedia4NLP and NLP4DBpedia frame (i.e. the used data evolves around
Wikipedia and DBpedia).
Submission formats
==============
Paper submission
-----------------------
All papers must represent original and unpublished work that is not
currently under review. Papers will be evaluated according to their
significance, originality, technical content, style, clarity, and
relevance to the workshop. At least one author of each accepted paper is
expected to attend the workshop.
* Full research paper (up to 12 pages)
* Position papers (up to 6 pages)
* Use case descriptions (up to 6 pages)
* Data/benchmark paper (2-6 pages, depending on the size and complexity)
Note: data and benchmarks papers are meant to provide a citable
reference for your data and benchmarks. We kindly require, that you
upload any data you use to our benchmark repository in parallel to the
submission. We recommend to use an open license (e.g. CC-BY), but
minimum requirement is free use. Please write to the mailing list, if
you have any problems.
Full instructions are available at:
http://nlp-dbpedia2013.blogs.aksw.org/submission/
Submission of data and use cases
--------------------------------------------
This workshop also targets non-academic users and developers. If you
have any (open) data (e.g. texts or annotations) that can be used for
benchmarking NLP tools, but do not want or needd to write an academic
paper about it, please feel free to just add it to this table:
http://tinyurl.com/nlp-benchmarks or upload it to our repository:
http://github.com/dbpedia/nlp-dbpedia
Full instructions are available at:
http://nlp-dbpedia2013.blogs.aksw.org/benchmarks/
Also if you have any ideas, use cases or data requests please feel free
to just post them on our mailing list: nlp-dbpedia-public [at]
lists.informatik.uni-leipzig.de or send them directly to the chairs:
nlp-dbpedia2013 [at] easychair.org
Program committee
==============
* Guadalupe Aguado, Universidad Politécnica de Madrid, Spain
* Chris Bizer, Universität Mannheim, Germany
* Volha Bryl, Universität Mannheim, Germany
* Paul Buitelaar, DERI, National University of Ireland, Galway
* Charalampos Bratsas, OKFN, Greece, ???????????? ????????????
????????????, (Aristotle University of Thessaloniki), Greece
* Philipp Cimiano, CITEC, Universität Bielefeld, Germany
* Samhaa R. El-Beltagy, ?????_????? (Nile University), Egypt
* Daniel Gerber, AKSW, Universität Leipzig, Germany
* Jorge Gracia, Universidad Politécnica de Madrid, Spain
* Max Jakob, Neofonie GmbH, Germany
* Anja Jentzsch, Hasso-Plattner-Institut, Potsdam, Germany
* Ali Khalili, AKSW, Universität Leipzig, Germany
* Daniel Kinzler, Wikidata, Germany
* David Lewis, Trinity College Dublin, Ireland
* John McCrae, Universität Bielefeld, Germany
* Uroš Miloševic', Institut Mihajlo Pupin, Serbia
* Roberto Navigli, Sapienza, Università di Roma, Italy
* Axel Ngonga, AKSW, Universität Leipzig, Germany
* Asunción Gómez Pérez, Universidad Politécnica de Madrid, Spain
* Lydia Pintscher, Wikidata, Germany
* Elena Montiel Ponsoda, Universidad Politécnica de Madrid, Spain
* Giuseppe Rizzo, Eurecom, France
* Harald Sack, Hasso-Plattner-Institut, Potsdam, Germany
* Felix Sasaki, Deutsches Forschungszentrum für künstliche Intelligenz,
Germany
* Mladen Stanojevic', Institut Mihajlo Pupin, Serbia
* Hans Uszkoreit, Deutsches Forschungszentrum für künstliche
Intelligenz, Germany
* Rupert Westenthaler, Salzburg Research, Austria
* Feiyu Xu, Deutsches Forschungszentrum für künstliche Intelligenz, Germany
Contact
=====
Of course we would prefer that you will post any questions and comments
regarding NLP and DBpedia to our public mailing list at:
nlp-dbpedia-public [at] lists.informatik.uni-leipzig.de
If you want to contact the chairs of the workshop directly, please write
to:
nlp-dbpedia2013 [at] easychair.org
Kind regards,
Sebastian Hellmann, Agata Filipowska, Caroline Barrière,
Pablo N. Mendes, Dimitris Kontokostas
--
Dipl. Inf. Sebastian Hellmann
Department of Computer Science, University of Leipzig
Events:
* NLP & DBpedia 2013 (http://nlp-dbpedia2013.blogs.aksw.org, Deadline:
*July 8th*)
* LSWT 23/24 Sept, 2013 in Leipzig (http://aksw.org/lswt)
Venha para a Alemanha como PhD: http://bis.informatik.uni-leipzig.de/csf
Projects: http://nlp2rdf.org , http://linguistics.okfn.org ,
http://dbpedia.org/Wiktionary , http://dbpedia.org
Homepage: http://bis.informatik.uni-leipzig.de/SebastianHellmann
Research Group: http://aksw.org
Dear colleagues!
I added section about using Wiktionary data in natural language
processing tasks to the Wikipedia article "Wiktionary".
You are welcome to extend this section:
http://en.wikipedia.org/wiki/Wiktionary#Wiktionary_data_in_NLP
Best regards,
Andrew Krizhanovsky.
Apologies for cross-posting!
=======================
NLP & DBpedia Workshop 2013
=======================
Free, open, interoperable and multilingual NLP for DBpedia and DBpedia
for NLP:
http://nlp-dbpedia2013.blogs.aksw.org/
Collocated with the International Semantic Web Conference 2013 (ISWC 2013)
21-22 October 2013, in Sydney, Australia (*Submission deadline July 8th*)
**********************************
Recently, the DBpedia community has experienced an immense increase in
activity and we believe, that the time has come to explore the
connection between DBpedia & Natural Language Processing (NLP) in a yet
unpreceded depth. The goal of this workshop can be summarized by this
(pseudo-) formula:
NLP & DBpedia == DBpedia4NLP && NLP4DBpedia
http://db0.aksw.org/downloads/CodeCogsEqn_bold2.gif
DBpedia has a long-standing tradition to provide useful data as well as
a commitment to reliable Semantic Web technologies and living best
practices. With the rise of WikiData, DBpedia is step-by-step relieved
from the tedious extraction of data from Wikipedia's infoboxes and can
shift its focus on new challenges such as extracting information from
the unstructured article text as well as becoming a testing ground for
multilingual NLP methods.
Contribution
=========
Within the timeframe of this workshop, we hope to mobilize a community
of stakeholders from the Semantic Web area. We envision the workshop to
produce the following items:
* an open call to the DBpedia data consumer community will generate a
wish list of data, which is to be generated from Wikipedia by NLP
methods. This wish list will be broken down to tasks and benchmarks and
a GOLD standard will be created.
* the benchmarks and test data created will be collected and published
under an open license for future evaluation (inspired by OAEI and
UCI-ML). An overview of the benchmarks can be found here:
http://nlp-dbpedia2013.blogs.aksw.org/benchmarks
Please sign up to our mailing list, if you are interested in discussing
guidelines and NLP benchmarking:
http://lists.informatik.uni-leipzig.de/mailman/listinfo/nlp-dbpedia-public
Important dates
===========
8 July 2013, Paper Submission Deadline
9 August 2013, Notification of accepted papers sent to authors
Motivation
=======
The central role of Wikipedia (and therefore DBpedia) for the creation
of a Translingual Web has recently been recognized by the Strategic
Research Agenda (cf. section 3.4, page 23) and most of the contributions
of the recently held Dagstuhl seminar on the Multilingual Semantic Web
also stress the role of Wikipedia for Multilingualism. As more and more
language-specific chapters of DBpedia appear (currently 14 language
editions), DBpedia is becoming a driving factor for a Linguistic Linked
Open Data cloud as well as localized LOD clouds with specialized domains
(e.g. the Dutch windmill domain ontology created from
http://nl.dbpedia.org ).
The data contained in Wikipedia and DBpedia have ideal properties for
making them a controlled testbed for NLP. Wikipedia and DBpedia are
multilingual and multi-domain, the communities maintaining these
resource are very open and it is easy to join and contribute. The open
license allows data consumers to benefit from the content and many parts
are collaboratively editable. Especially, the data in DBpedia is widely
used and disseminated throughout the Semantic Web.
NLP4DBpedia
==========
DBpedia has been around for quite a while, infusing the Web of Data with
multi-domain data of decent quality. These triples are, however, mostly
extracted from Wikipedia infoboxes. To unlock the full potential of
Wikipedia articles for DBpedia, the information contained in the
remaining part of the articles needs to be analysed and triplified.
Here, the NLP techniques may be of favour.
DBpedia4NLP
==========
On the other hand NLP, and information extraction techniques in
particular, involve various resources while processing texts from
various domains. These resources may be used e.g. as an element of a
solution e.g. gazetteer being an important part of a rule created by an
expert or disambiguation resource, or while delivering a solution e.g.
within machine learning approaches. DBpedia easily fits in both of these
roles.
We invite papers from both these areas including:
1. Knowledge extraction from text and HTML documents (especially
unstructured and semi-structured documents) on the Web, using
information in the Linked Open Data (LOD) cloud, and especially in DBpedia.
2. Representation of NLP tool output and NLP resources as RDF/OWL, and
linking the extracted output to the LOD cloud.
3. Novel applications using the extracted knowledge, the Web of Data or
NLP DBpedia-based methods.
The specific topics are listed below.
Topics
=====
- Improving DBpedia with NLP methods
- Finding errors in DBpedia with NLP methods
- Annotation methods for Wikipedia articles
- Cross-lingual data and text mining on Wikipedia
- Pattern and semantic analysis of natural language, reading the Web,
learning by reading
- Large-scale information extraction
- Entity resolution and automatic discovery of Named Entities
- Multilingual entity recognition task of real world entities
- Frequent pattern analysis of entities
- Relationship extraction, slot filling
- Entity linking, Named Entity disambiguation, cross-document
co-reference resolution
- Disambiguation through knowledge base
- Ontology representation of natural language text
- Analysis of ontology models for natural language text
- Learning and refinement of ontologies
- Natural language taxonomies modeled to Semantic Web ontologies
- Use cases for potential data extracted from Wikipedia articles
- Use cases of entity recognition for Linked Data applications
- Impact of entity linking on information retrieval, semantic search
Furthermore, an informal list of NLP tasks can be found on this
Wikipedia page:
http://en.wikipedia.org/wiki/Natural_language_processing#Major_tasks_in_NLP
These are relevant for the workshop as long as they fit into the
DBpedia4NLP and NLP4DBpedia frame (i.e. the used data evolves around
Wikipedia and DBpedia).
Submission formats
==============
Paper submission
-----------------------
All papers must represent original and unpublished work that is not
currently under review. Papers will be evaluated according to their
significance, originality, technical content, style, clarity, and
relevance to the workshop. At least one author of each accepted paper is
expected to attend the workshop.
* Full research paper (up to 12 pages)
* Position papers (up to 6 pages)
* Use case descriptions (up to 6 pages)
* Data/benchmark paper (2-6 pages, depending on the size and complexity)
Note: data and benchmarks papers are meant to provide a citable
reference for your data and benchmarks. We kindly require, that you
upload any data you use to our benchmark repository in parallel to the
submission. We recommend to use an open license (e.g. CC-BY), but
minimum requirement is free use. Please write to the mailing list, if
you have any problems.
Full instructions are available at:
http://nlp-dbpedia2013.blogs.aksw.org/submission/
Submission of data and use cases
--------------------------------------------
This workshop also targets non-academic users and developers. If you
have any (open) data (e.g. texts or annotations) that can be used for
benchmarking NLP tools, but do not want or needd to write an academic
paper about it, please feel free to just add it to this table:
http://tinyurl.com/nlp-benchmarks or upload it to our repository:
http://github.com/dbpedia/nlp-dbpedia
Full instructions are available at:
http://nlp-dbpedia2013.blogs.aksw.org/benchmarks/
Also if you have any ideas, use cases or data requests please feel free
to just post them on our mailing list: nlp-dbpedia-public [at]
lists.informatik.uni-leipzig.de or send them directly to the chairs:
nlp-dbpedia2013 [at] easychair.org
Program committee
==============
* Guadalupe Aguado, Universidad Politécnica de Madrid, Spain
* Chris Bizer, Universität Mannheim, Germany
* Volha Bryl, Universität Mannheim, Germany
* Paul Buitelaar, DERI, National University of Ireland, Galway
* Charalampos Bratsas, OKFN, Greece, ???????????? ????????????
????????????, (Aristotle University of Thessaloniki), Greece
* Philipp Cimiano, CITEC, Universität Bielefeld, Germany
* Samhaa R. El-Beltagy, ?????_????? (Nile University), Egypt
* Daniel Gerber, AKSW, Universität Leipzig, Germany
* Jorge Gracia, Universidad Politécnica de Madrid, Spain
* Max Jakob, Neofonie GmbH, Germany
* Anja Jentzsch, Hasso-Plattner-Institut, Potsdam, Germany
* Ali Khalili, AKSW, Universität Leipzig, Germany
* Daniel Kinzler, Wikidata, Germany
* David Lewis, Trinity College Dublin, Ireland
* John McCrae, Universität Bielefeld, Germany
* Uroš Miloševic', Institut Mihajlo Pupin, Serbia
* Roberto Navigli, Sapienza, Università di Roma, Italy
* Axel Ngonga, AKSW, Universität Leipzig, Germany
* Asunción Gómez Pérez, Universidad Politécnica de Madrid, Spain
* Lydia Pintscher, Wikidata, Germany
* Elena Montiel Ponsoda, Universidad Politécnica de Madrid, Spain
* Giuseppe Rizzo, Eurecom, France
* Harald Sack, Hasso-Plattner-Institut, Potsdam, Germany
* Felix Sasaki, Deutsches Forschungszentrum für künstliche Intelligenz,
Germany
* Mladen Stanojevic', Institut Mihajlo Pupin, Serbia
* Hans Uszkoreit, Deutsches Forschungszentrum für künstliche
Intelligenz, Germany
* Rupert Westenthaler, Salzburg Research, Austria
* Feiyu Xu, Deutsches Forschungszentrum für künstliche Intelligenz, Germany
Contact
=====
Of course we would prefer that you will post any questions and comments
regarding NLP and DBpedia to our public mailing list at:
nlp-dbpedia-public [at] lists.informatik.uni-leipzig.de
If you want to contact the chairs of the workshop directly, please write
to:
nlp-dbpedia2013 [at] easychair.org
Kind regards,
Sebastian Hellmann, Agata Filipowska, Caroline Barrière,
Pablo N. Mendes, Dimitris Kontokostas
--
Dipl. Inf. Sebastian Hellmann
Department of Computer Science, University of Leipzig
Events: NLP & DBpedia 2013 (http://nlp-dbpedia2013.blogs.aksw.org,
Deadline: *July 8th*)
Venha para a Alemanha como PhD: http://bis.informatik.uni-leipzig.de/csf
Projects: http://nlp2rdf.org , http://linguistics.okfn.org ,
http://dbpedia.org/Wiktionary , http://dbpedia.org
Homepage: http://bis.informatik.uni-leipzig.de/SebastianHellmann
Research Group: http://aksw.org
Hello,
Based on the feedback I got from the community, I have made my final
proposal. Since only 3 days are left, I urge everyone to kindly reply as
early as possible.
My proposal can be found here:
https://www.mediawiki.org/wiki/User:Rahul21/Gsoc
Thanks in advance,
Rahul
I have been working on a pronunciation recording extension as a part of my
Summer Code Project.https://www.mediawiki.org/wiki/User:Rahul21/Gsoc
Could you people take a look at the UI Mockups that i have made and give me
some suggestions or feedback as to how would you want the tool to be
deployed?
Thanks in Advance
Rahul