Hey,
I'd like to reiterate over the importance of using @covers tags [0] for
test cases rather then writing "Test for SomeClass". These tags make
coverage reporting a lot more accurate, and thus make it easier to spot
areas lacking tests.
Today I added a bunch of @covers tags in the DataModel component. The
reported coverage dropped from 85% to 57%, quite a difference.
[0]
http://phpunit.de/manual/current/en/appendixes.annotations.html#appendixes.…
Cheers
--
Jeroen De Dauw
http://www.bn2vs.com
Don't panic. Don't be evil. ~=[,,_,,]:3
--
Forwarding to the Wikidata tech list in case this makes a future
Wiktionary collaboration easier.
-------- Original Message --------
Subject: [Wiki-research-l] Java-based Wiktionary Library (JWKTL) 1.0.0
released as open source software (Wiki-research-l Digest, Vol 96, Issue 22)
Date: Tue, 20 Aug 2013 14:20:56 +0000
From: Judith Eckle-Kohler <eckle-kohler(a)ukp.informatik.tu-darmstadt.de>
Reply-To: wiki-research-l(a)lists.wikimedia.org
To: wiki-research-l(a)lists.wikimedia.org
<wiki-research-l(a)lists.wikimedia.org>
[Apologies for X-posting]
We are pleased to announce the release of the Java-based Wiktionary
Library (JWKTL) 1.0.0 - an application programming interface for Wiktionary.
Project homepage: http://code.google.com/p/jwktl/
== Overview ==
JWKTL (Java-based Wiktionary Library) is an application programming
interface for the free multilingual online dictionary Wiktionary
(http://www.wiktionary.org). JWKTL enables efficient and structured
access to the information encoded in the English, the German, and the
Russian Wiktionary language editions, including sense definitions, part
of speech tags, etymology, example sentences, translations, semantic
relations, and many other lexical information types. The Russian JWKTL
parser is based on Wikokit (http://code.google.com/p/wikokit/).
Prior to being available as open source software, JWKTL has been a
research project at the Ubiquitous Knowledge Processing (UKP) Lab of the
Technische Universität Darmstadt, Germany. The following people have
mainly contributed to this project: Yevgen Chebotar, Iryna Gurevych,
Christian M. Meyer, Christof Müller, Lizhen Qu, Torsten Zesch.
== Publications ==
A detailed description of Wiktionary and JWKTL is available in our
scientific articles:
* Christian M. Meyer and Iryna Gurevych: Wiktionary: A new rival for
expert-built lexicons? Exploring the possibilities of collaborative
lexicography, Chapter 13 in S. Granger & M. Paquot (Eds.): Electronic
Lexicography, pp. 259291, Oxford: Oxford University Press, November
2012.
(http://www.ukp.tu-darmstadt.de/publications/details/?no_cache=1&tx_bibtex_p…)
* Christian M. Meyer and Iryna Gurevych: OntoWiktionary Constructing
an Ontology from the Collaborative Online Dictionary Wiktionary, chapter
6 in M. T. Pazienza and A. Stellato (Eds.): Semi-Automatic Ontology
Development: Processes and Resources, pp. 131161, Hershey, PA: IGI
Global, February 2012.
(http://www.ukp.tu-darmstadt.de/publications/details/?no_cache=1&tx_bibtex_p…)
* Torsten Zesch, Christof Müller, and Iryna Gurevych: Extracting Lexical
Semantic Knowledge from Wikipedia and Wiktionary, in: Proceedings of the
6th International Conference on Language Resources and Evaluation
(LREC), pp. 16461652, May 2008. Marrakech, Morocco.
(http://www.ukp.tu-darmstadt.de/publications/details/?no_cache=1&tx_bibtex_p…)
== License and Availability ==
The latest version of JWKTL is available via Maven Central. If you use
Maven as your build tool, then you can add JWKTL as a dependency in your
pom.xml file:
<dependency>
<groupId>de.tudarmstadt.ukp.jwktl</groupId>
<artifactId>jwktl</artifactId>
<version>1.0.0</version>
</dependency>
JWKTL is available as open source software under the Apache License 2.0
(ASL). The software thus comes "as is" without any warranty (see license
text for more details). JWKTL makes use of Berkeley DB Java Edition
5.0.73 (Sleepycat License), Apache Ant 1.7.1 (ASL), Xerces 2.9.1 (ASL),
JUnit 4.10 (CPL).
Some classes have been taken from the Wikokit project (available under
multiple licenses, redistributed under the ASL license). See NOTICE.txt
for further details.
== Contact ==
Please direct any questions or suggestions to
https://groups.google.com/forum/#!forum/jwktl-users
Group E-Mail: jwktl-users(a)googlegroups.com
Best wishes,
Christian M. Meyer
--
Christian M. Meyer, M.Sc.
Doctoral Researcher
Ubiquitous Knowledge Processing (UKP Lab)
FB 20 Computer Science Department
Technische Universität Darmstadt
Hochschulstr. 10, D-64289 Darmstadt, Germany
Phone [+49] (0)6151 16-5386, fax -5455, room S2/02/B113
meyer(a)ukp.informatik.tu-darmstadt.de
www.ukp.tu-darmstadt.de<http://www.ukp.tu-darmstadt.de>
Web Research at TU Darmstadt (WeRC)
www.werc.tu-darmstadt.de<http://www.werc.tu-darmstadt.de>
---------- Forwarded message ----------
From: "Sumana Harihareswara" <sumanah(a)wikimedia.org>
Date: Aug 22, 2013 7:06 PM
Subject: Re: [Wikitech-l] unexpected error info in HTML
To: "Wikimedia developers" <wikitech-l(a)lists.wikimedia.org>, "Jiang BIAN" <
bianjiang(a)google.com>
Cc:
> On 08/01/2013 03:08 AM, Jiang BIAN <bianjiang(a)google.com> wrote:
> > Hi,
> >
> > I noticed some pages we crawled containing error message like this;
> >
> > <div id="mw-content-text" lang="zh-CN" dir="ltr"
class="mw-content-ltr"><p
> > class="error">Failed to render property P373:
> > Wikibase\LanguageWithConversion::factory: given languages do not have
the
> > same parent language</p>
> >
> >
> > But when I open the url in browser, there is no such message. And using
> > index.php can also get normal content without error messages.
> >
> > Here are examples you can retry:
> >
> > bad
> > $ wget 'http://zh.wikipedia.org/zh-cn/Google'
> >
> > good
> > $ wget 'http://zh.wikipedia.org/w/index.php?title=Google'
> >
> >
> > Looks like something is wrong on Wikipedia side, anything need to fix?
> >
> >
> >
> > Thanks
>
> I checked with Jiang Bian and found out that this is still happening --
> can anyone help Google out here? :-)
>
> --
> Sumana Harihareswara
> Engineering Community Manager
> Wikimedia Foundation
>
> _______________________________________________
> Wikitech-l mailing list
> Wikitech-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
I have updated doxygen documentation for Wikibase and added it for some
related extensions. (some query-related components are still missing)
http://wikidata-docs.wikimedia.de/
If this is useful, I can add remaining components, make it look nicer and
set it up to update regularly.
Cheers,
Katie
--
Katie Filbert
Wikidata Developer
Wikimedia Germany e.V. | NEW: Obentrautstr. 72 | 10963 Berlin
Phone (030) 219 158 26-0
http://wikimedia.de
Wikimedia Germany - Society for the Promotion of free knowledge eV Entered
in the register of Amtsgericht Berlin-Charlottenburg under the number 23
855 as recognized as charitable by the Inland Revenue for corporations I
Berlin, tax number 27/681/51985.
Hey :)
For the unlikely case that some readers of this list aren't aware of
this yet: There are a lot of Wikidata bugs in bugzilla that could use
some help. They're marked with the keyword need-volunteer. You can see
them here: https://bugzilla.wikimedia.org/buglist.cgi?keywords=need-volunteer%2C%20&ke…
(Obviously you're welcome to work on other stuff too but if you don't
know what to put time into this is the perfect list.) If you're
working on any of those please leave a comment in the bug report so
others know you're on it. In case you're more into writing a bot then
http://www.wikidata.org/wiki/Wikidata:Bot_requests is the place to
look.
If you need help here and the IRC channel #wikimedia-wikidata on
freenode is a good place to ask.
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Technical Projects
Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
Hey :)
It seems that badges support is stalled on a decision about how
exactly to define the set of available badges if I read
https://bugzilla.wikimedia.org/show_bug.cgi?id=40810 correctly. Can we
make a decision and move forward? It's the most voted on bug we have.
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Technical Projects
Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
Hey,
Regarding https://gerrit.wikimedia.org/r/#/c/77862/4
The approach implemented in PS4 is different then the one I was thinking of
originally when we discussed this. It has
{
qualifiers: { snaks: {}, order: [] }
}
rather then
{
qualifiers: {},
qualifiers-order: []
}
While I agree that the former is nicer, it is a breaking change in our API.
I'm thus not merging it as long as it is not clear that we are all well
aware of this breakage and find it an acceptable price for having a nicer
format. If we decide to go forward with this, it should also be announced
to our API users and properly documented.
Cheers
--
Jeroen De Dauw
http://www.bn2vs.com
Don't panic. Don't be evil. ~=[,,_,,]:3
--
Hi everyone,
Summary: We are looking for ways of how to map individual triples from
DBpedia to something that can be consumed by the Wikidata API.
Long story:
Denis Lukovnikov is a Google summer of code student, working on new and
pretty user interface for DBpedia.
A demo of the work in progress can be found at [1].
Besides the new design, a further goal of the project is to offer a
framework for triple-level actions.
One of the planned actions is about pushing DBpedia triples to Wikidata:
For example, I notice that the Wikidata page for my home town "Berndorf
in Lower Austria" does not contain the population:
http://www.wikidata.org/wiki/Q666615
Looking at the corresponding DBpedia entry, this information actually
exists there:
http://dbpedia.org/resource/Berndorf,_Lower_Austria
The new DBpedia interface should offer a button next to the "population
8728" triple which enables transfer of this information to Wikidata.
In another GSoC project, Hady Elsahar is working on mappings between the
wikidata RDF vocabulary and the DBpedia vocabulary.
This means, we can in principle map DBpedia RDF data to Wikidata RDF.
However, looking at the Wikidata API [2] there is
action=wbcreateclaim *
with the example:
api.php?action=wbcreateclaim&entity=q42&property=p9001&snaktype=novalue&token=foobar&baserevid=7201010
So the core question is, how can we map e.g. properties such as
wikidata:population (if that existed) to their respective Wikidata
property identifier (Q12345)?
This goes for any property that may occur in an RDF dump, such as:
http://www.wikidata.org/wiki/Special:EntityData/Q666615.nt
Ideally I think we are looking for either a Wikidata service or a
dataset similar to [3] what is what we have done in the LinkedGeoData
Project:
In this project, we map OpenStreetMap (OSM) data to RDF, and we maintain
our RDF mappings in the database [3] together with the original OSM data.
Then we use an RDB2RDF mapper configured with the view definitions in
[4] to expose the relational database as Linked Data [5] and a virtual
SPARQL endpoint [6].
Cheers,
Claus
[1] WiP demo:
http://dbpv.cstadler.aksw.org/#/page/Berndorf,_Lower_Austria (currently
loading may take a while for pages with many inverse statements).
[2] WikiData API: http://www.wikidata.org/w/api.php
[3] Raw-Data to RDF mappings for OpenStreetMap:
https://github.com/GeoKnow/LinkedGeoData/blob/master/linkedgeodata-core/src…
[4] RDB2RDF view definitions:
https://github.com/GeoKnow/LinkedGeoData/blob/master/linkedgeodata-core/src…
[5] Eiffel Tower resource: http://linkedgeodata.org/page/triplify/way5013364
[6] Explanation of the unterlying SQL:
http://linkedgeodata.org/vsnorql/?query=Explain+Select+*+{+%3Chttp%3A%2F%2F…
--
Dipl. Inf. Claus Stadler
Department of Computer Science, University of Leipzig
Research Group: http://aksw.org/
Workpage & WebID: http://aksw.org/ClausStadler
Phone: +49 341 97-32260
Hi,
there seems to be a problem with the dumps or with the way in which we
interpret the dumps.
Currently, daily dumps come with a maxrevid.txt file that is supposed to
give the largest revision number in the dump. For example, daily dump of
1st Aug 2013 has max id 62860640 [1]. I guess that's true.
The wda scripts (that also create the statistics and digested dumps we
publish) use this number to figure out if a daily is still relevant or
if it is already contained in the latest full dump. For this, we need to
get the maximal revision id of the full dump. We do this by reading the
file site_stats.sql.gz, where we look for the line starting with INSERT
INTO `site_stats` and take a revision number from there (third number in
the insert tuple). For example, for the dump of 27 July 2013, this
number is 63069374 [2].
There is a problem here, since the maximal revision in the dumps of 27
July 2013 is not actually that high (the history dump of that date is
incomplete, but the current revs dump is done and has max rev 61983867
[3]). Thus, our scripts ignore several days of dailies.
Before I go and work on this, my question is whether this is an error in
our script (i.e., the number we take from sitestats is not supposed to
be the max revision) or an error in the dumps (i.e., sitestats was
exported wrongly).
Cheers,
Markus
[1] http://dumps.wikimedia.org/other/incr/wikidatawiki/20130801/maxrevid.txt
[2]
http://dumps.wikimedia.org/wikidatawiki/20130727/wikidatawiki-20130727-site…
[3] This can be seen in the comments for the dump at
http://dumps.wikimedia.org/wikidatawiki/20130727/
--
Markus Kroetzsch, Departmental Lecturer
Department of Computer Science, University of Oxford
Room 306, Parks Road, OX1 3QD Oxford, United Kingdom
+44 (0)1865 283529 http://korrekt.org/