I am putting ideas together for a project that I am referring to as "Structured History." The objective is to create a systematic way of accessing information of an historical nature within an ontological context that would permit historical reasoning. The elements of the project are (1) access to structured historical information, which WikiData is beginning to provide, and (2) relating the body of structured information to one or more overall ontological frameworks, and then providing the tools for accessing, displaying and analyzing the data. I have not yet identified an ontology that would be most appropriate for historical purposes, other than an EU funded project called Papyrus http://www.ict-papyrus.eu - but this seems to be used by no one. Therefore, I would suggest that the development of an historical ontology (possibly building on the work of Papyrus, which in turn built on the museum ontology CIDAR) be part of a "structured history" project.
The model consists of "historical entities" that can be anything that exists or existed in time and place (animate, human or inanimate - book, river), Places, and Events (which may be nested so that a War event can consist of a sequence of Battle events).
If a project of this scope were undertaken, it could provide the framework within which other projects, such as the history of science, could be developed. Issues with regard to WikiData are: (1) to what degree will information (that may be available in Wikipedia articles) be structured sufficiently for incorporation into the structured history project? (2) How can historical narratives be broken down into elements that can be collected in a meaningful way that retains semantic validity (veracity)? The system should be comprehensive enough that all information and artifacts in museums and libraries could be incorporated, either directly into the ontology or by cross-reference (to CIDAR for instance).
I would be interested to see if there would be support for such an endeavor, and for incorporating it into the proposal for WikiData for Research.
Thanks - Sam Smith (SammyWiki)
Michigan
-----Original Message-----
From: wikidata-l-bounces(a)lists.wikimedia.org [mailto:wikidata-l-bounces@lists.wikimedia.org] On Behalf Of wikidata-l-request(a)lists.wikimedia.org
Sent: Saturday, December 27, 2014 7:01 AM
To: wikidata-l(a)lists.wikimedia.org
Subject: Wikidata-l Digest, Vol 37, Issue 20
Send Wikidata-l mailing list submissions to
wikidata-l(a)lists.wikimedia.org
To subscribe or unsubscribe via the World Wide Web, visit
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
or, via email, send a message with subject or body 'help' to
wikidata-l-request(a)lists.wikimedia.org
You can reach the person managing the list at
wikidata-l-owner(a)lists.wikimedia.org
When replying, please edit your Subject line so it is more specific than "Re: Contents of Wikidata-l digest..."
Today's Topics:
1. Use cases for Wikidata in research contexts (Daniel Mietchen)
----------------------------------------------------------------------
Message: 1
Date: Fri, 26 Dec 2014 18:43:06 +0100
From: Daniel Mietchen <daniel.mietchen(a)googlemail.com>
To: "Discussion list for the Wikidata project."
<wikidata-l(a)lists.wikimedia.org>
Subject: [Wikidata-l] Use cases for Wikidata in research contexts
Message-ID:
<CAN6n2b0J6LYOT4dgRjeXrbq_EoDkvW1nSHkpxQkSPapi94ZSYQ(a)mail.gmail.com>
Content-Type: text/plain; charset=UTF-8
Dear all,
we are building the Wikidata for Research proposal [1] around use cases for Wikidata (or Wikibase) in research contexts (cf. Task 4.1).
So if you are using Wikidata or Wikibase in research contexts already, or are contemplating to do so, we'd appreciate your comments.
The same goes for use cases for DBpedia that would be enhanced by having a Wikidata/ Wikibase implementation.
Thanks,
Daniel
[1] https://www.wikidata.org/wiki/Wikidata:WikiProject_Wikidata_for_research
--
http://www.naturkundemuseum-berlin.de/en/institution/mitarbeiter/mietchen-d…https://en.wikipedia.org/wiki/User:Daniel_Mietchen/Publicationshttp://okfn.orghttp://wikimedia.org
------------------------------
_______________________________________________
Wikidata-l mailing list
Wikidata-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
End of Wikidata-l Digest, Vol 37, Issue 20
******************************************
Hey,
Since the Tree of Life by Denny is outdated, we thought it was a nice idea,
to have a new one, to have an overview over the biological taxonomy on
wikidata. Not only to have a nice looking tree but also to see, where
errors are and to correct and update it. Right now, there are 660775 Items
in the tree.
Even though the change of names can be seen instantly, because this is
based on the API, changes in the order need an update of the whole tree,
because it's based on wikidata dumps. (This one on the most recent one from
15.12.2014)
Here you go, this is the new tree of life, made with a lot of love:
https://tools.wmflabs.org/tree-of-life/
If you have any corrections, additions or features you want to add, feel
free to ping me, send me a mail, submit a patch or file an issue on github.
The repo for the tree is on https://github.com/frimelle/tree-of-life
Cheers,
Lucie (frimelle)
Hi everyone!
I would like to quickly introduce myself, my name is Sebastian
(https://www.wikidata.org/wiki/User:Sebotic) and I will join the lab of
Andrew Su (sulab.org) towards the end of this year. As you probably
know, our main aim regarding WikiData is to integrate human genomics and
medical data, this will also be my primary project.
We therefore thought that it might be interesting to the WikiData
community and also helpful for us, if we could integrate our sources for
the ProteinBoxBot into the Mediawiki Gerrit code review system and
receive contributions from the large community.
The projects on Mediawiki Gerrit seems to be primarily concerning
Mediawiki itself or addons to it. Would it be possible, to get our
ProteinBoxBot project integrated on Gerrit? It is currently hosted on
https://bitbucket.org/sulab/wikidatagenebot and is under strong
development by Andra Wagmeester and I am about to join him in this effort.
Thank you!
Best regards,
Sebastian
Hi,
Is there a way to list organizations from Wikidata? I can see Google
<http://www.wikidata.org/wiki/Q95>, but I can't see that it's a company or
an organization. With Freebase this was quite easy as they have lists of
categories on their homepage and I would simply need to visit
freebase.com/organization
Also, with Freebase we could add any company. Can we add any company to
Wikidata or must it have a Wikipedia article about it? I hope not as that
would neuter its usefulness for many use cases. E.g. Crunchbase is a great
resource for many people and yet most of the items in Crunchbase will never
be in Wikipedia despite that usefulness.
Thanks,
Ben
Hi all,
Could those knowledgeable about OWL or intending to use Wikidata's RDF /
OWL exports please weigh in at
https://www.wikidata.org/wiki/Wikidata:Property_proposal/Property_metadata#…
? [1]
Being able to declare certain "properties of properties" is an essential
building block for querying and inference. However, the way to declare
that a property is, say, transitive in OWL does not have a clear analog in
Wikidata syntax. We could certainly shoehorn such a statement into our
existing model (and it looks like we'll need to), but it is important to do
so in a way that complicate things as little as possible for downstream
users, e.g. outside researchers or developers using the RDF exports and
assuming standard OWL semantics.
Please make any comments on this on-wiki at the location linked above.
That way we can keep the discussion centralized.
Other discussions on that page could also benefit from input by people
knowledgeable about Semantic Web vocabulary.
Thanks,
Eric
https://www.wikidata.org/wiki/User:Emw
1. Discussion permalink:
https://www.wikidata.org/w/index.php?title=Wikidata:Property_proposal/Prope…
Hello,
I am trying to get the list of pages (on Wikidata) that have a "featured article" badge for trwiki. We are using a local template called "seçkin madde" that denotes an article is featured, and I need to check if all pages using this template has an associated badge entry on Wikidata.
Thanks.
Hi,
I'm doing a project with the content translation team as part of my FOSS
OPW internship.Our goal is to understand the proportion of translated pages
in some wiki from other wikipedias.
As a first step, I need to create a list of of articles in one language
(i.e. HE) that have corresponding articles in another language, starting
with English.
I wonder what is the best why to create this list. Possible approaches that
I thought of:
1. Using the API Sandbox and some iterative script that calls it.
2. Using the Wikimedia Dumps (specifically wiki interlanguage link records).
3. Using the Wikidata dumps (specifically wikidatawiki-latest-langlinks).
Am I missing something? Which way is the best to build the list? especially
when taking into account the possibility of inline interlanguage links?
Thanks for your help,
Neta
** apologies for cross-posting **
==== Second Call for Papers ====
2015 Workshop on Semantics, Analytics, Visualisation: Enhancing Scholarly Data (SAVE-SD 2015)
**** LATE BREAKING NEWS **********************
** Submission deadline for research, position, demo, poster papers: January 24, 2015
** Invited speaker: Paul Groth (Elsevier Labs)
** Possibility of HTML submission in RASH format (i.e., simplified HTML+RDFa)
** Best Paper Award (250 euros) sponsored by Pensoft
** Best RASH Paper Award (voucher of 150 euros in books) sponsored by Springer
** Selected papers for Journal of Documentation special issue
** All papers will be published the in ACM Digital Library
Date: May 19, 2015 (Half day)
Venue: Florence, Italy (co-located with WWW 2015)
Hashtag: #savesd2015
Twitter: @savesdworkshop
Site: http://cs.unibo.it/save-sd/2015/index.html
Workshop chairs:
- Francesco Osborne (Open University, UK)
- Silvio Peroni (University of Bologna, Italy, and National Research Council, Italy)
- Jun Zhao (Lancaster University, UK)
# DESCRIPTION
The main goal of the SAVE-SD workshop is to bring together publishers, companies and researchers from different fields (among which Document and Knowledge Engineering, Semantic Web, Natural Language Processing, Scholarly Communication, Bibliometrics, and Human-Computer Interaction) in order to bridge the gap between the theoretical/academic and practical/industrial aspects in regards to scholarly data.
The following fields will be addressed:
- semantics of scholarly data, i.e. how to semantically represent, categorise, connect and integrate scholarly data, in order to foster reusability and knowledge sharing;
- analytics on scholarly data, i.e. designing and implementing novel and scalable algorithms for knowledge extraction with the aim of understanding research dynamics, forecasting research trends, fostering connections between groups of researchers, informing research policies, analysing and interlinking experiments and deriving new knowledge;
- visualisation of and interaction with scholarly data, i.e. providing novel user interfaces and applications for navigating and making sense of scholarly data and highlighting their patterns and peculiarities.
# TOPICS OF INTEREST
We would encourage submission of papers covering one or more of the following topics:
Semantics:
- Data models (e.g., ontologies, vocabularies, schemas) for the description of scholarly data and the linking between scholarly data and academic papers that report or cite them
- Description of citations and citation networks
- Theoretical models describing the rhetorical and argumentative structure of scholarly papers and
their application in practice
- Description and use of provenance information of scholarly data
- From digital libraries of scholarly papers to Linked Open Datasets: models, applicability and
challenges
- Definition and description of scholarly publishing processes
- Modelling licences for scholarly documents and data
Analytics:
- Assessing the quality and/or trust of scholarly data
- Pattern discovery of scholarly data
- Citation analysis and prediction
- Scientific claims identification from textual contents
- New indicators for measuring the quality and relevance of research
- Comparison between standard metrics (e.g., h-index, impact factor, citation counting) and
alternative metrics in real-case scenarios
- Automatic or semi-automatic approaches to making sense of research dynamics
- Content- and data-based semantic similarity of scholarly papers
- Citation generation
- Automatic semantic enhancement of existing scholarly libraries and papers
- Reconstruction, forecasting and monitoring of scholarly data
Visualisation and Interaction:
- Novel user interfaces for interaction with paper, metadata, content, and data
- Visualisation of citation networks according to multiple dimensions (e.g., citation counting,
citation functions, kinds of citing/cited entities)
- Visualisation of related papers or data according to multiple dimensions (semantic similarity of
abstracts, keywords, etc.)
- Applications for making sense of scholarly data
- Usability studies on existing interfaces (e.g., Web sites, Web applications, smartphone apps) for
browsing scholarly data
- Scholarly data and ubiquity: accessing scholarly information from multiple devices (PC, tablet,
smartphones)
- Applications for the (semi-)automatic annotation of scholarly papers
# IMPORTANT DATES
- Submission deadline: January 24, 2015 (23:59 Hawaii Standard Time)
- Acceptance notification: February 22, 2015
- Camera ready deadline: March 8, 2015
# SUBMISSIONS
SAVE-SD welcomes the submission of original research and application papers dealing with the tree aforementioned field. We encourage theoretical, methodological, empirical and applications papers. We appreciate the submission of papers incorporating links to data sets and other material used for evaluation as well as to live demos and software source code.
All submissions must be written in English. Two formats are possible for the submission:
- PDF, a file formatted according to the ACM double-column instructions (http://www.acm.org/sigs/publications/proceedings-templates);
- HTML, a zip archive containing an HTML file compliant with the Research Articles in Simplified HTML (RASH) format (http://cs.unibo.it/save-sd/rash/index.html).
We invite four kinds of submissions:
- full research papers (max. 6 pp. in PDF or 5400 words in HTML)
- position papers (max. 4 pp. in PDF or 3600 words in HTML)
- demo papers (max. 2 pp. in PDF or 1800 words in HTML)
- poster papers (max. 2 pp. in PDF or 1800 words in HTML)
All the aforementioned limits include metadata (title, authors, keywords, ACM categories, abstract), acknowledgements, references and the whole content of the paper. In the HTML format, figures and tables count 300 words each.
Submissions and reviewing will be supported by the EasyChair system:
https://easychair.org/conferences/?conf=savesd2015
# EVALUATION OF SUBMISSIONS
In order to evaluate the submitted papers, we have three different program committees:
- The Senior PC, whose members will act as meta-reviewers and have the crucial role of balancing the scores provided by the reviews from the other two PCs (see below);
- The Industrial PC, who will evaluate the submissions from an industrial perspective mainly – by assessing how much the theories/applications described in the papers do/may influence (positively or negatively) the publishing domain and whether they could be concretely adopted by publishers and scholarly data providers;
- The Academic PC, who will evaluate the papers from an academic perspective mainly – by assessing the quality of the research described in such papers.
All submissions will be reviewed by (at least) one Senior PC member, one Industrial PC member and two Academic PC members. The final decision of acceptance/rejection will be made in consensus by the chairs.
# PUBLICATION VENUES
The proceedings of SAVE-SD will be collected in the Companion volume of the WWW 2015 conference, which will be published by ACM in its digital library. The WWW 2015 organisers will require at least one registration per paper published in the Companion volume. At the time of submission of the final camera-ready copy, authors will have to indicate the already registered person for that publication.
In addition, the authors of the best papers of the workshop will be invited to submit an extended version of their work to a special issue that will be published as part of the Journal of Documentation (http://www.emeraldinsight.com/journal/jd), one of the longest established academic journals in library and information science (2013 Impact Factor: 1.035; indexed in several citation services, among which Elsevier's Scopus and Thomson Reuters' Journal Citation Reports).
# INVITED SPEAKER
The invited speaker for the opening keynote of the workshop will be Paul Groth from Elsevier Labs.
Paul Groth holds a Ph.D. in Computer Science from the University of Southampton (2007), has done research at the University of Southern California, and has been an assistant professor in the Web and Media Group at the VU University of Amsterdam and a member of its Network Institute.
His research focuses on dealing with large amounts of diverse contextualised knowledge with a particular focus on the web and science applications. This includes research in data provenance, web and data science, knowledge and data integration and knowledge sharing. Paul was co-chair of the W3C Provenance Working Group that created a standard for provenance interchange. He is co-author of Provenance: an Introduction to PROV. Currently, he is a key contributor to Open PHACTS (http://www.openphacts.org) - a project to develop a provenance-enabled platform for large scale pharmacological information. He blogs at http://thinklinks.wordpress.com. You can find him on twitter: @pgroth
# AWARDS
An award of 250 euros, kindly sponsored by Pensoft (http://www.pensoft.net), will be assigned to the best paper of the workshop. The decision will be taken by considering both the reviews received and how the authors will address the reviewers' concerns in the camera ready.
Another award of 150 euros as a voucher for buying Springer's products, kindly sponsored by Springer (http://www.springer.com), will be assigned to the best workshop submission in RASH format (i.e., simplified HTML+RDFa). The decision will be taken by considering the quality of the markup (i.e., the less syntactical mistakes there are in the markup, the better), the number of RDF statements defined in RDFa, and the number of RDF links to LOD datasets.