Hi,
TL;DR: Did anybody consider using Wikidata items of Wikipedia templates to
store multilingual template parameters mapping?
Full explanation:
As in many other projects in the Wikimedia world, templates are one of the
biggest challenges in developing the ContentTranslation extension.
Translating a template between languages is tedious - many templates are
language-specific, many others have a corresponding template, but
incompatible parameters, and even if the parameters are compatible, there
is usually no comfortable mapping. Some work in that direction was done in
DBpedia, but AFAIK it's far from complete.
In ContentTranslation we have a simplistic mechanism for mapping between
template parameters in pairs of languages, with proof of concept for three
templates. We can enhance it with more templates, but the question is how
much can it scale.
Some templates shouldn't need such mapping at all - they should pull their
data from Wikidata. This is gradually being done for infoboxes in some
languages, and it's great.
But not all templates can be easily mapped to Wikidata data. For example -
reference templates, various IPA and language templates, quotation
formatting, and so on. For these, parameter mapping could be useful, but
doing this for a single language pair doesn't seem robust and reminds me of
the old ways in which interlanguage links were stored.
So, did anybody consider using Wikidata items of templates to store
multilingual template parameters mapping?
--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
“We're living in pieces,
I want to live in peace.” – T. Moore
Hi all,
Again, for already so many times, many users have complained on nl-wiki
that it is for them impossible now to add interwikilinks on Wikidata. They
are sick of all the changes all the time, especially if they experience the
new design as impossible to use, as it now is.
They are lost with the current design and can't add new interwikilinks.
That this pops-up so many times is a serious problem that needs a solution.
Wikidat is not meant for techno users only, but that is how many regular
users experience Wikidata. If regular users find themselves impossible to
add/update pages, the software needs a big change to re-enable them to work
with Wikidata again.
Ow, I am just the messenger... [1]
Romaine
[1] https://www.wikidata.org/wiki/Q2515525
OPENSYM 2015, THE 11TH INTERNATIONAL SYMPOSIUM ON OPEN COLLABORATION
August 19-21, 2015 | San Francisco, California, U.S.A.
http://opensym.org/os2015 | ACM SIWEB and ACM SIGSOFT supported
ABOUT THE CONFERENCE
The 11th International Symposium on Open Collaboration (OpenSym 2015) is the
premier conference on open collaboration research and practice, including
free/libre/open source software, open data, IT-driven open innovation
research, wikis and related open collaborative media, and Wikipedia and
related Wikimedia projects.
OpenSym brings together the different strands of open collaboration research
and practice, seeking to create synergies and inspire new collaborations
between computer science and information systems researchers, social
scientists, legal scholars, and everyone interested in understanding open
collaboration and how it is changing the world.
OpenSym 2015 will be held in San Francisco, California, on *August 19-21, 2015*.
This is the general call for papers and includes the
- research track call for submissions,
- industry and community track call for submissions, and
- doctoral symposium call for submissions.
OpenSym is held in-cooperation with ACM SIGWEB and ACM SIGSOFT. As in previous
years, the conference proceedings will be archived in the ACM digital library.
RESEARCH TRACK CALL FOR SUBMISSIONS
The conference provides the following peer-reviewed research tracks.
- Free/libre/open source software research, chaired by Carlos Jensen of Oregon
State University and Gregorio Robles of Universidad Rey Juan Carlos. This
track seeks papers on all aspects of FLOSS. For detailed topics and the
research track committee please see http://wp.me/Pezfy-IU.
- IT-driven open innovation research, chaired by Ann Majchrzak of University
of Southern California and Arvind Malhotra of University of North Carolina at
Chapel Hill. This track is devoted to research on the process of expanding
research and development activities beyond the boundaries of single company
structures. For detailed topics and the research track committee please see
http://wp.me/Pezfy-J3.
Open data research, chaired by Carl Lagoze of University of Michigan. This
track contributes to the increasing awareness on Open Data in research. For
detailed topics and the research track committee please see http://wp.me/Pezfy-J5.
- Wikis and open collaboration research, chaired by Kevin Crowston of Syracuse
University. This track is dedicated to the science and application of wikis
and open collaboration technology outside of the context of Wikipedia. For
detailed topics and the research track committee please see http://wp.me/Pezfy-J7.
- Wikipedia and related projects research, chaired by Claudia Müller-Birn of
Freie Universität Berlin and Aaron Shaw of Northwestern University. This track
addresses research specifically on Wikipedia and associated projects. For
detailed topics and the research track committee please see http://wp.me/Pezfy-J9.
Research papers present integrative reviews or original reports of substantive
new work: theoretical, empirical, and/or in the design, development and/or
deployment of novel concepts, systems, and mechanisms. Research papers will be
reviewed by a research track program committee to meet rigorous academic
standards of publication. Papers will be reviewed for relevance, conceptual
quality, innovation and clarity of presentation.
Authors can submit full papers (5-10 pages), short papers (2-4 pages), and
research posters (1-2 pages). For more details on paper types please see
http://wp.me/Pezfy-Je.
Submission deadline for all research contributions is *March 29th, 2015*.
Authors submit through EasyChair at
https://easychair.org/conferences/?conf=opensym2015. Submissions and final
contributions must follow the ACM SIG Proceedings template found at
http://www.acm.org/sigs/publications/proceedings-templates.
OpenSym seeks to accommodate the needs of the different research disciplines
it draws on. Authors whose submissions have been accepted for presentation at
the conference have a choice of having
- their paper become part of the official proceedings, archived in the ACM
Digital Library, or having
- only a short abstract included in the proceedings (rather than the full
submitted paper) in order to preserve future publication possibilities.
DOCTORAL SYMPOSIUM CALL FOR SUBMISSIONS
OpenSym seeks to explore the synergies between all strands of open
collaboration research. Thus, we will have a doctoral symposium, in which
Ph.D. students from different disciplines can present their work and receive
feedback from senior faculty and their peers.
Submission deadline for doctoral symposium position papers is *May 3rd, 2015*.
Authors submit through EasyChair at
https://easychair.org/conferences/?conf=opensym2015. Submissions and final
contributions must follow the ACM SIG Proceedings template found at
http://www.acm.org/sigs/publications/proceedings-templates.
More information is available at http://wp.me/Pezfy-Jh.
INDUSTRY AND COMMUNITY TRACK CALL FOR SUBMISSIONS
OpenSym is also seeking submissions for experience reports (full and short),
tutorials, workshops, panels, non-research posters, and demos. Such work
accepted for presentation or performance at the conference is considered part
of the industry and community track. It will be put into the proceedings in an
industry and community track section; authors can opt-out of the publication,
as with research papers, but will still have to provide an abstract (less than
one page) for the proceedings.
Submission deadline for industry and community track papers is *April 19, 2015*.
Authors submit through EasyChair at
https://easychair.org/conferences/?conf=opensym2015. Submissions and final
contributions must follow the ACM SIG Proceedings template found at
http://www.acm.org/sigs/publications/proceedings-templates.
More information is available at http://wp.me/Pezfy-Jh.
THE OPENSYM CONFERENCE EXPERIENCE
OpenSym 2015 will be held in San Francisco, California, on August 19-21, 2015.
Research, industry, and community presentations and performances will be
accompanied by keynotes, invited speakers, and a social program in one of the
most vibrant cities on this planet.
The open space track is a key ingredient of the event that distinguishes
OpenSym from other conferences. It is an integral part of the program that
makes it easy to talk to other researchers and practitioners and to stretch
your imagination and conversations beyond the limits of your own
sub-discipline, exposing you to the full breadth of open collaboration
research. The open space track is entirely participant-organized, is open for
everyone, and requires no submission or review.
The general chair of the conference is Dirk Riehle of Friedrich-Alexander
University Erlangen-Nürnberg. Feel free to contact us with any questions you
might have at info(a)opensym.org.
--
Website: http://dirkriehle.com - Twitter: @dirkriehle
Ph (DE): +49-157-8153-4150 - Ph (US): +1-650-450-8550
On Fri, Jan 23, 2015 at 6:24 AM, Bahodir Mansurov
<bmansurov(a)wikimedia.org> wrote:
> I’ve created a GNOME Shell extension that allows the user to search for Wikidata items directly from the shell. Currently you can search for simple things such as “Obama”, “Book”, etc. I plan on adding support for complex queries such as “the population of the earth” which would show the current population of the earth. In the future I also see this extension handle the submission of new entries or editing existing ones. Check it out here [1] if you’re interested. Pull requests are also welcome ;)
>
> [1] https://github.com/6ahodir/wikidata-search-provider
Thanks, Bahodir!
Forwarding to the Wikidata mailing list so they get to see this too.
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
Magnus, et al.,
I've been using parts of the Wikidata APIs to do the following:
- Get detail about an item. For example:
https://www.wikidata.org/w/api.php?action=wbgetentities&ids=Q42&props=label…
- Get items that are subclass of, instance of, or part of. For example:
http://wdq.wmflabs.org/api?q=claim[279:153]
Now I'd like to find out what items are related to an item through
various properties, as is demonstrated in Resonator. What techniques
and APIs are used?
Thanks,
James Weaver
Hi (esp. WMF people),
The JSON dumps used to be at
http://dumps.wikimedia.org/other/wikidata/
Now this directory is empty. Any hints at what is going on?
Cheers,
Markus
Dear Wikidata JSON export team,
There seems to be a sytnax error in the 20120112 JSON file that (I
think) has already been there in the previous dump. So I guess it makes
sense to report it.
In line 9374899, around column 2648 of the 20120112 JSON dump, we find
"snaks":[]
Of course, {} would be correct here instead of []. The entity that this
JSON occurs in is Q4383128.
Cheers,
Markus
2015-01-17 4:27 GMT-05:00 Lydia Pintscher <lydia.pintscher(a)wikimedia.de>:
>
> The log is at
> https://meta.wikimedia.org/wiki/IRC_office_hours/Office_hours_2015-01-16
> for anyone who couldn't make it.
Denny discusses importing all missing VIAF keys from Freebase using
"multichill" (unclear what that is from the context) on the assumption that
the error rate is low. It would be worth checking assumptions like that
with folks who are familiar with the Freebase data before acting on them.
Here are some things that I think are true about the VIAF keys in Freebase:
- they were assigned by a user, not by Google/Metawab (not necessarily a
bad thing since some of the biggest problems in Freebase were created by
G/M and some users have contributed very high quality data)
- they keys were, I believe, assigned based heavily on existing Library of
Congress identifiers that had previously been assigned by Metaweb. Those
key assignments are not as high quality as other parts of Freebase. One
easy thing to check for is people with two LC (and thus two VIAF) keys
assigned. In cases where there are more than key and the extra keys don't
represent pseudonyms, this is a clear error.
- Freebase doesn't create separate entities for pseudonyms, unlike the
library cataloging world. Depending on what decision Wikidata makes in
this regard, it's something to watch out for when reusing Freebase author
data (including VIAF keys)
- much Freebase author data was imported from OpenLibrary which has its own
set of quality issues. A bunch of this data was later deleted, leaving
that portion of the graph somewhat thready and moth-eaten. It's unclear
whether that was a net gain or loss in overall bibliographic data quality
for Freebase.
- I suspect that most VIAF keys which are in Freebase and not Wikidata
represent entities which are not in Wikidata which means they aren't useful
anyway since he wants to focus on creating new links, not new entities (a
direction that I'm not sure I agree with, but that's a whole separate
discussion).
One of the key inputs to judging the quality of assertions is their
provenance. Fortunately, this is recorded for all assertions in Freebase
and it's possible to trace a given fact back to the user, toolchain, or
process that added it to the database. Unfortunately, this information is
only available through the Freebase API, not the bulk data dump.
Hopefully, this will change before Google completely abandons Freebase.
If any Wikidata folk want to discuss VIAF keys in Freebase (or its author
data in general), feel free to get in touch.
Tom
Hi all,
I have been trying today to get the Java library Wikidata-Toolkit
going, but about to give up... I keep running with both 0.3.0 and
current master into a NullPointerException... I thought it was how I
called the code, and did add several System.out calls, and in the end
just tried to get it running from the command line... I tried the
example from the website (though replaced the Dump examples, which I
don't see in master; btw, "mvn test" runs fine) using a pristine
master:
$ cd wdtk-examples/
$ mvn compile
$ mvn exec:java
-Dexec.mainClass="org.wikidata.wdtk.examples.EntityStatisticsProcessor"
In doing so, I get the same NPE:
********************************************************************
*** Wikidata Toolkit: EntityStatisticsProcessor
***
*** This program will download and process dumps from Wikidata.
*** It will print progress information and some simple statistics.
*** Results about property usage will be stored in a CSV file.
*** See source code for further details.
********************************************************************
2015-01-17 16:53:00 INFO - Using download directory
/home/egonw/var/Projects/GitHub/Wikidata-Toolkit/wdtk-examples/dumpfiles/wikidatawiki
[WARNING]
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.codehaus.mojo.exec.ExecJavaMojo$1.run(ExecJavaMojo.java:293)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.NullPointerException
at org.wikidata.wdtk.dumpfiles.DumpProcessingController.processDumpFile(DumpProcessingController.java:470)
at org.wikidata.wdtk.dumpfiles.DumpProcessingController.processMostRecentDump(DumpProcessingController.java:456)
at org.wikidata.wdtk.dumpfiles.DumpProcessingController.processMostRecentJsonDump(DumpProcessingController.java:426)
at org.wikidata.wdtk.examples.ExampleHelpers.processEntitiesFromWikidataDump(ExampleHelpers.java:158)
at org.wikidata.wdtk.examples.EntityStatisticsProcessor.main(EntityStatisticsProcessor.java:88)
... 6 more
I tried finding what goes wrong, but cannot grasp all the magic that
is going on... the directory it reports was created, but is empty...
$ mvn --version
Apache Maven 3.0.5
Maven home: /usr/share/maven
Java version: 1.7.0_65, vendor: Oracle Corporation
Java home: /usr/lib/jvm/java-7-openjdk-i386/jre
Default locale: en_US, platform encoding: UTF-8
OS name: "linux", version: "3.16.0-4-686-pae", arch: "i386", family: "unix"
Can someone give me some pointers where and how it is testing of dump
files exist? Is this problem something platform dependent?
Thanks,
Egon
--
E.L. Willighagen
Department of Bioinformatics - BiGCaT
Maastricht University (http://www.bigcat.unimaas.nl/)
Homepage: http://egonw.github.com/
LinkedIn: http://se.linkedin.com/in/egonw
Blog: http://chem-bla-ics.blogspot.com/
PubList: http://www.citeulike.org/user/egonw/tag/papers
ORCID: 0000-0001-7542-0286
ImpactStory: https://impactstory.org/EgonWillighagen