Denny's project is a very interesting.
We already have Wikidata and Magnus Manske's autodesc which can create
paragraph-length natural language for some types of items.
"""Denny Vrandečić is a Croatia researcher, programmer, and computer
He was born on February 27, 1978 in Stuttgart.
He studied at Karlsruhe Institute of Technology from October 2004 until
June 2010, University of Stuttgart from September 1998 until February
2004, University of Stuttgart from September 1997 until February 2004,
and Geschwister-Scholl-Gymnasium. He worked for Google from October
2013, for Wikimedia Deutschland from March 2012 until September 2013,
and for Karlsruhe Institute of Technology from 2004 until 2012."""
Currently he seems to support English, French and Dutch.
I think Magnus Manske would accept pull request to other languages at
So how would we go beyond Magnus? Would the Wikidata representation
suffice? I have seen Q50827579 and Q28819478 for Wikidata to language
generation, but I am not aware of running applications and are they
better than Magnus' hard-coded approach?
I have been experimenting a bit the other way. Ordia can go from natural
language to Wikidata-lexemes (for single Danish example):
>> from ordia.base import Base
>> base = Base()
>> base.words_to_form_ids('der kom en soldat marcherende henad
[['L3064-F1'], ['L3065-F3', 'L3065-F6'], ['L2022-F1',
['L3074-F1'], ['L3075-F5'], ['L3215-F1'], ['L3216-F2']]
Writing the encyclopedic text in "Wikidata-lexemesh" could perhaps ease
translation, particularly after 18 October when senses are planned to be
On 09/29/2018 08:42 PM, Pine W wrote:
Forwarding because this (ambitious!) proposal may be
of interest to
people on other lists. I'm not endorsing the proposal at this time, but
I'm curious about it.
---------- Forwarded message ---------
From: *Denny Vrandečić* <vrandecic(a)gmail.com <mailto:firstname.lastname@example.org>>
Date: Sat, Sep 29, 2018 at 6:32 PM
Subject: [Wikimedia-l] Wikipedia in an abstract language
To: Wikimedia Mailing List <wikimedia-l(a)lists.wikimedia.org
Semantic Web languages allow to express ontologies and knowledge bases in a
way meant to be particularly amenable to the Web. Ontologies formalize the
shared understanding of a domain. But the most expressive and widespread
languages that we know of are human natural languages, and the largest
knowledge base we have is the wealth of text written in human languages.
We looks for a path to bridge the gap between knowledge representation
languages such as OWL and human natural languages such as English. We
propose a project to simultaneously expose that gap, allow to collaborate
on closing it, make progress widely visible, and is highly attractive and
valuable in its own right: a Wikipedia written in an abstract language to
be rendered into any natural language on request. This would make current
Wikipedia editors about 100x more productive, and increase the content of
Wikipedia by 10x. For billions of users this will unlock knowledge they
currently do not have access to.
My first talk on this topic will be on October 10, 2018, 16:45-17:00, at
the Asilomar in Monterey, CA during the Blue Sky track of ISWC. My second,
longer talk on the topic will be at the DL workshop in Tempe, AZ, October
27-29. Comments are very welcome as I prepare the slides and the talk.
Link to the paper: http://simia.net/download/abstractwikipedia.pdf
Wikimedia-l mailing list, guidelines at:
New messages to: Wikimedia-l(a)lists.wikimedia.org
Wikidata mailing list