Hi,
TL;DR: Did anybody consider using Wikidata items of Wikipedia templates to
store multilingual template parameters mapping?
Full explanation:
As in many other projects in the Wikimedia world, templates are one of the
biggest challenges in developing the ContentTranslation extension.
Translating a template between languages is tedious - many templates are
language-specific, many others have a corresponding template, but
incompatible parameters, and even if the parameters are compatible, there
is usually no comfortable mapping. Some work in that direction was done in
DBpedia, but AFAIK it's far from complete.
In ContentTranslation we have a simplistic mechanism for mapping between
template parameters in pairs of languages, with proof of concept for three
templates. We can enhance it with more templates, but the question is how
much can it scale.
Some templates shouldn't need such mapping at all - they should pull their
data from Wikidata. This is gradually being done for infoboxes in some
languages, and it's great.
But not all templates can be easily mapped to Wikidata data. For example -
reference templates, various IPA and language templates, quotation
formatting, and so on. For these, parameter mapping could be useful, but
doing this for a single language pair doesn't seem robust and reminds me of
the old ways in which interlanguage links were stored.
So, did anybody consider using Wikidata items of templates to store
multilingual template parameters mapping?
--
Amir Elisha Aharoni · אָמִיר אֱלִישָׁע אַהֲרוֹנִי
http://aharoni.wordpress.com
“We're living in pieces,
I want to live in peace.” – T. Moore
Hi everyone!
I would like to quickly introduce myself, my name is Sebastian
(https://www.wikidata.org/wiki/User:Sebotic) and I will join the lab of
Andrew Su (sulab.org) towards the end of this year. As you probably
know, our main aim regarding WikiData is to integrate human genomics and
medical data, this will also be my primary project.
We therefore thought that it might be interesting to the WikiData
community and also helpful for us, if we could integrate our sources for
the ProteinBoxBot into the Mediawiki Gerrit code review system and
receive contributions from the large community.
The projects on Mediawiki Gerrit seems to be primarily concerning
Mediawiki itself or addons to it. Would it be possible, to get our
ProteinBoxBot project integrated on Gerrit? It is currently hosted on
https://bitbucket.org/sulab/wikidatagenebot and is under strong
development by Andra Wagmeester and I am about to join him in this effort.
Thank you!
Best regards,
Sebastian
For more than a year I am asking users to add their articles to Wikidata
when they have written it. That seems succesful, they added their articles
more and more and did understand how to do that. Until recently. Now I get
more and more complaints from users that they do not understand any more
how to add a newly written article to an item. They seem to have tried, but
fail in actual getting it managed. That is a worse development!
Romaine
Hey folks :)
I'll be doing an office hour on IRC for all things Wikidata. It'll be
on December 3rd at 18:00 UTC. We'll be meeting in #wikimedia-office on
freenode.
Hope to see many of you there!
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
Hi,
I have just noticed that at this address:
https://aws.amazon.com/datasets
several public datasets are available (including Wikipedia's traffic stats).
Maybe you can find something stimulating in there.
Cristian
Hey folks :)
Good news before the end of the year!
* You can go to test.wikidata.org right now and test out statements on
properties. If everything goes well we will make it available on
Wikidata in 1 week. One thing that is missing is a new datatype to
link to other properties to indicate things like "property x is the
inverse of property y". That will come a bit later. Discussion about
proposed properties to be used on properties is happening at
https://www.wikidata.org/wiki/Wikidata:Property_proposal/Property_metadata
* Language fallbacks are coming. This is a first version. It'll only
show fallbacks of linked entities on an item. So if you are viewing an
item with a statement linking to another item you will see a label in
another language for this linked item if it doesn't have a label in
your language. Fallbacks are currently implemented to follow the
fallback for system messages. Unfortunately you can't try that out on
test.wikidata.org yet. That'll hopefully happen next week. It should
go live on Wikidata in two weeks.
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
** apologies for cross-posting **
==== First Call for Papers ====
2015 Workshop on Semantics, Analytics, Visualisation: Enhancing Scholarly Data (SAVE-SD 2015)
Date: May 19, 2015 (Half day)
Venue: Florence, Italy (co-located with WWW 2015)
Hashtag: #savesd2015
Twitter: @savesdworkshop
Site: http://cs.unibo.it/save-sd/2015/index.html
Workshop chairs:
- Francesco Osborne (Open University, UK)
- Silvio Peroni (University of Bologna, Italy, and National Research Council, Italy)
- Jun Zhao (Lancaster University, UK)
# DESCRIPTION
The main goal of the SAVE-SD workshop is to bring together publishers, companies and researchers from different fields (among which Document and Knowledge Engineering, Semantic Web, Natural Language Processing, Scholarly Communication, Bibliometrics, and Human-Computer Interaction) in order to bridge the gap between the theoretical/academic and practical/industrial aspects in regards to scholarly data.
The following fields will be addressed:
- semantics of scholarly data, i.e. how to semantically represent, categorise, connect and integrate scholarly data, in order to foster reusability and knowledge sharing;
- analytics on scholarly data, i.e. designing and implementing novel and scalable algorithms for knowledge extraction with the aim of understanding research dynamics, forecasting research trends, fostering connections between groups of researchers, informing research policies, analysing and interlinking experiments and deriving new knowledge;
- visualisation of and interaction with scholarly data, i.e. providing novel user interfaces and applications for navigating and making sense of scholarly data and highlighting their patterns and peculiarities.
# TOPICS OF INTEREST
We would encourage submission of papers covering one or more of the following topics:
Semantics:
- Data models (e.g., ontologies, vocabularies, schemas) for the description of scholarly data and the linking between scholarly data and academic papers that report or cite them
- Description of citations and citation networks
- Theoretical models describing the rhetorical and argumentative structure of scholarly papers and
their application in practice
- Description and use of provenance information of scholarly data
- From digital libraries of scholarly papers to Linked Open Datasets: models, applicability and
challenges
- Definition and description of scholarly publishing processes
- Modelling licences for scholarly documents and data
Analytics:
- Assessing the quality and/or trust of scholarly data
- Pattern discovery of scholarly data
- Citation analysis and prediction
- Scientific claims identification from textual contents
- New indicators for measuring the quality and relevance of research
- Comparison between standard metrics (e.g., h-index, impact factor, citation counting) and
alternative metrics in real-case scenarios
- Automatic or semi-automatic approaches to making sense of research dynamics
- Content- and data-based semantic similarity of scholarly papers
- Citation generation
- Automatic semantic enhancement of existing scholarly libraries and papers
- Reconstruction, forecasting and monitoring of scholarly data
Visualisation and Interaction:
- Novel user interfaces for interaction with paper, metadata, content, and data
- Visualisation of citation networks according to multiple dimensions (e.g., citation counting,
citation functions, kinds of citing/cited entities)
- Visualisation of related papers or data according to multiple dimensions (semantic similarity of
abstracts, keywords, etc.)
- Applications for making sense of scholarly data
- Usability studies on existing interfaces (e.g., Web sites, Web applications, smartphone apps) for
browsing scholarly data
- Scholarly data and ubiquity: accessing scholarly information from multiple devices (PC, tablet,
smartphones)
- Applications for the (semi-)automatic annotation of scholarly papers
# IMPORTANT DATES
- Submission deadline: January 24, 2015 (23:59 Hawaii Standard Time)
- Acceptance notification: February 22, 2015
- Camera ready deadline: March 8, 2015
# SUBMISSIONS
SAVE-SD welcomes the submission of original research and application papers dealing with the tree aforementioned field. We encourage theoretical, methodological, empirical and applications papers. We appreciate the submission of papers incorporating links to data sets and other material used for evaluation as well as to live demos and software source code.
All submissions must be written in English. Two formats are possible for the submission:
- PDF, a file formatted according to the ACM double-column instructions (http://www.acm.org/sigs/publications/proceedings-templates);
- HTML, a zip archive containing an HTML file compliant with the Research Articles in Simplified HTML (RASH) format (http://cs.unibo.it/save-sd/rash/rash.rng).
We invite four kinds of submissions:
- full research papers (max. 6 pp. in PDF or 5400 words in HTML)
- position papers (max. 4 pp. in PDF or 3600 words in HTML)
- demo papers (max. 2 pp. in PDF or 1800 words in HTML)
- poster papers (max. 2 pp. in PDF or 1800 words in HTML)
All the aforementioned limits include metadata (title, authors, keywords, ACM categories, abstract), acknowledgements, references and the whole content of the paper. In the HTML format, figures and tables count 300 words each.
Submissions and reviewing will be supported by the EasyChair system:
https://easychair.org/conferences/?conf=savesd2015
# EVALUATION OF SUBMISSIONS
In order to evaluate the submitted papers, we have three different program committees:
- The Senior PC, whose members will act as meta-reviewers and have the crucial role of balancing the scores provided by the reviews from the other two PCs (see below);
- The Industrial PC, who will evaluate the submissions from an industrial perspective mainly – by assessing how much the theories/applications described in the papers do/may influence (positively or negatively) the publishing domain and whether they could be concretely adopted by publishers and scholarly data providers;
- The Academic PC, who will evaluate the papers from an academic perspective mainly – by assessing the quality of the research described in such papers.
All submissions will be reviewed by (at least) one Senior PC member, one Industrial PC member and two Academic PC members. The final decision of acceptance/rejection will be made in consensus by the chairs.
# PUBLICATION VENUES
The proceedings of SAVE-SD will be collected in the Companion volume of the WWW 2015 conference, which will be published by ACM in its digital library. The WWW 2015 organisers will require at least one registration per paper published in the Companion volume. At the time of submission of the final camera-ready copy, authors will have to indicate the already registered person for that publication.
In addition, the authors of the best papers of the workshop will be invited to submit an extended version of their work to a special issue that will be published as part of the Journal of Documentation (http://www.emeraldinsight.com/journal/jd), one of the longest established academic journals in library and information science (2013 Impact Factor: 1.035; indexed in several citation services, among which Elsevier's Scopus and Thomson Reuters' Journal Citation Reports).
Hey folks :)
We'll be doing another office hour to answer all of the questions
about structured data for multimedia files. We'll try...
It'll happen on November 20th at 19:30 UTC in #wikimedia-office on
freenode. See http://www.timeanddate.com/worldclock/fixedtime.html?hour=19&min=30&sec=0&d…
for your timezone.
Hope to see many of you there!
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Product Manager for Wikidata
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.