Hi everybody,
We’re preparing for the June 2016 research newsletter and looking for contributors. Please take a look at: https://etherpad.wikimedia.org/p/WRN201606 and add your name next to any paper you are interested in covering. Our target publication date is Saturday July 2 UTC although actual publication might happen several days later. As usual, short notes and one-paragraph reviews are most welcome.
Highlights from this month:
• This issue won’t be published before Saturday (May 28), possibly a bit later
• Case study in political user behavior on Wikipedia
• Combining syntactic patterns and Wikipedia's hierarchy of hyperlinks to extract meronym relations
• Crowdsourcing not all sourced by the crowd: An observation on the behavior of Wikipedia participants
• Customer relationship management practices in the online community – Wikipedia
• Determining the influence of reddit posts on wikipedia pageviews
• Digital History Meets Wikipedia: Analyzing Historical Persons in Wikipedia
• Enriching Wikidata with Frame Semantics
• Manipulating Google’s Knowledge Graph Box to Counter Biased Information Processing During an Online Search on Vaccination: Application of a Technological Debiasing Strategy
• Quality Assessment of Wikipedia Articles Without Feature Engineering
• The double power law in human collaboration behavior: The case of Wikipedia
• Visualizations of relationships among knowledge? Try WikiSeeker!
• Wikipedia traffic data and electoral prediction: towards theoretically informed models
If you have any question about the format or process feel free to get in touch off-list.
Masssly, Tilman Bayer and Dario Taraborelli
[1] http://meta.wikimedia.org/wiki/Research:Newsletter
Hello everyone!
The Wikimedia Brazilian Group of Education and Research[1] will host on October 13th and 14th in Rio de Janeiro the "1st Brazilian Wikipedia Science Conference"
This event aims to bring together students, teachers and researchers with projects and publications about Wikipedia, other Wikimedia projects and their dynamics.
We are almost done and soon we will announce our official website and call for papers. Articles and poster will be accepted.
The Wikimedia Foundation has validated our logo [2] and the use of the Wikipedia brand in the event name.
We are using as inspiration the events organized by Wikimedia UK[3] and OpenSym[4].
We are trying to bring to Brazil some important keynotes from our movement. If someone from this list is interested in coming to visit us, let's keep in touch!
That event is under organization in a very organic and experimental way, with no funding from WMF to see the impact and interest from the local academic community. Based on the results and impacts we will try to do something bigger in 2017.
All the content produced during the event will be published in CC licenses (photos, videos, interviews, publications).
Soon we will have more news! You can check our planning page here[5]
[1] https://meta.wikimedia.org/wiki/BR-ER
[2] https://commons.wikimedia.org/wiki/Category:Wikipedia_Science_Conference
[3] https://wikimedia.org.uk/wiki/Wikipedia_Science_Conference
[4] http://www.opensym.org/os2016/
[5] https://meta.wikimedia.org/wiki/BR-ER/Events/Congresso_Científico_Brasileir…
Best regards
Rodrigo Padula
Coordenador de Projetos
Grupo Wikimedia Brasileiro de Educação e Pesquisa
http://www.wikimedia.org.br
21 99326-0558
You may want to follow https://phabricator.wikimedia.org/T68699 , since
it will affect how often user_touched is updated for users that choose
"Keep me logged in".
Matt
Metadata
======
Call for Papers
Submission deadline: July 7th 2016
EasyChair submission page:
https://easychair.org/conferences/?conf=blink2016
Accepted papers: Short (8 pages) and long (16 pages)
Workshop page: http://project-hobbit.eu/events/blink-2016/
Conference: ISWC - Kobe, Japan - October, 17th or 18th, 2016
Description
=======
BLINK will provide a forum where topics related to the evaluation
(included, but not limited to the performance, accuracy, expressive
power and usability) of Linked Data Technologies for different steps of
the Linked Data lifecycle can be discussed and elaborated upon.
Linked Data now part of the new data economy and Big Linked Data is
gaining in use and traction. Systems are constantly being developed in
order to support the booming exchange of data (existing in numerous
formats) in the Web and the Enterprise. Linked Data benchmarks can
function as valuable tools to objectively depict and illustrate the
level of adequacy and thus performance provided by the existing Linked
Data systems.
This workshop aims to bring together a broad range of attendants
interested in benchmarking Linked Data and aims at identifying the
specific needs and challenges of the domain in order to foster
interdisciplinary collaborations towards attaining these challenges.
More specifically the objectives of this workshop are to:
create a discussion forum where researchers and industrials can meet and
discuss topics related to the performance of Linked Data systems and
expose and initiate discussions on best practices, different application
needs and scenarios related to Linked Data management.
Topics of Interest
===============
We welcome contributions presenting experiences with benchmarking Linked
Data technologies as well as technical contributions regarding the
development of benchmarks for all aspects of the Linked Data/Big Data
lifecycle. All domains (e.g., life science, social networks, smart
cities, news, digital forensics, e-science and geo-spatial data
management) are welcome.
Topics of interest include but are not limited to:
* Linked Data benchmarks
* Novel benchmarking results
* Analysis of existing benchmarks
* Novel measures for benchmarking Linked Data
* Linked Data benchmark evaluation
* Complex benchmarking pipelines
* Application of benchmarks in academic/industrial settings
* Tools and methodologies for the linked data generation and
acquisition, analytics and processing, storage and curation,
visualization and data access.
This series of workshops are supported by H2020 European Project HOBBIT
(Holistic Benchmarking of Big Linked Data), see http://project-hobbit.eu/.
Paper Submission
===========
The workshop will accept two types of submissions: short papers (8
pages) will be either position papers or describe early works in the
area of benchmarking. Long papers (up to 16 pages) will describe
benchmarks, benchmarking techniques or benchmarking results along the
linked data lifecycle. Details on the submission process can be found at
http://project-hobbit.eu/events/blink-2016/
Important Dates:
July 7th, 2016: paper submission deadline
July 31st 2016: Notifications send to authors
August 25th 2016: Camera-ready papers for workshops
October 17th or 18th: Workshop
Submission Details
============
The workshop is now accepting paper submissions. Long papers (up to 16
pages) and short papers (up to 8 pages) describing approaches or ideas /
challenges on the topics of the workshop are invited. Submissions must
be in PDF, formatted in the style of the Springer Publications format
for Lecture Notes in Computer Science (LNCS). For details on the LNCS
style, see Springer’s Author Instructions. Papers should be submitted
through the EasyChair system
https://easychair.org/conferences/?conf=blink2016 no later than midnight
Hawaii time July 7th, 2016. Submissions will be reviewed by members of
the workshop program committee. Accepted papers will be included in the
ISWC 2016 Workshop on Benchmarking Linked Data (BLINK) proceedings.
Hello,
FYI, a few deadlines are approaching:
* 2016-06-29: Early bird registration deadline for KDD2016 (SIGKDD
Conference on Knowledge Discovery and Data Mining)
* 2016-06-30: Early bird registration deadline for OpenSym 2016
(International Symposium on Open Collaboration)
* 2016-07-11: Submission deadline for COLLA 2016 (Conference on
Advanced Collaborative Networks, Systems and Applications)
More information and links are available on the list of
research-related events on Meta:
https://meta.wikimedia.org/wiki/Research:Events#Upcoming_events
--
Guillaume Paumier
We just released a new version of Research:FAQ on Meta [1], significantly
expanded and updated, to make our processes at WMF more transparent and to
meet an explicit FDC request to clarify the role and responsibilities of
individual teams involved in research across the organization.
The previous version – written from the perspective of the (now inactive)
Research:Committee, and mostly obsolete since the release of WMF's open
access policy [2] – can still be found here [3].
Comments and bold edits to the new version of the document are welcome. For
any question or concern, you can drop me a line or ping my username on-wiki.
Thanks,
Dario
[1] https://meta.wikimedia.org/wiki/Research:FAQ
[2] https://wikimediafoundation.org/wiki/Open_access_policy
[3] https://meta.wikimedia.org/w/index.php?title=Research:FAQ&oldid=15176953
*Dario Taraborelli *Head of Research, Wikimedia Foundation
wikimediafoundation.org • nitens.org • @readermeter
<http://twitter.com/readermeter>
***DEADLINE EXTENSION***
2nd Call for Posters & Demos
SEMANTiCS 2016 - The Linked Data Conference
Transfer // Engineering // Community
12th International Conference on Semantic Systems
Leipzig, Germany
September 12 -15, 2016
http://2016.semantics.cc
Important Dates (Posters & Demos)
* Submission Deadline: extended: July 1, 2016
(11:59 pm, Hawaii time)
* Notification of Acceptance: extended: August 2, 2016
(11:59 pm, Hawaii time)
* Camera-Ready Paper: extended: August 10, 2016
(11:59 pm, Hawaii time)
Submissions via Easychair:
https://easychair.org/conferences/?conf=semantics2016research
The annual SEMANTiCS conference is the meeting place for professionals
who make semantic computing work, who understand its benefits and
encounter its limitations. Every year, SEMANTiCS attracts information
managers, IT-architects, software engineers and researchers from
organisations ranging from NPOs, through public administrations to the
largest companies in the world. Attendees learn from industry experts
and top researchers about emerging trends and topics in the fields of
semantic software, enterprise data, linked data & open data strategies,
methodologies in knowledge modelling and text & data analytics. The
SEMANTiCS community is highly diverse; attendees have responsibilities
in interlinking areas like knowledge management, technical
documentation, e-commerce, big data analytics, enterprise search,
document management, business intelligence and enterprise vocabulary
management.
The success of last year’s conference in Vienna with more than 280
attendees from 22 countries proves that SEMANTiCS 2016 will continue a
long tradition of bringing together colleagues from around the world.
There will be presentations on industry implementations, use case
prototypes, best practices, panels, papers and posters to discuss
semantic systems in birds-of-a-feather sessions as well as informal
settings. SEMANTICS addresses problems common among information
managers, software engineers, IT-architects and various specialist
departments working to develop, implement and/or evaluate semantic
software systems.
The SEMANTiCS program is a rich mix of technical talks, panel
discussions of important topics and presentations by people who make
things work - just like you. In addition, attendees can network with
experts in a variety of fields. These relationships provide great value
to organisations as they encounter subtle technical issues in any stage
of implementation. The expertise gained by SEMANTiCS attendees has a
long-term impact on their careers and organisations. These factors make
SEMANTiCS for our community the major industry related event across Europe.
SEMANTiCS 2016 will especially welcome submissions for the following hot
topics:
* Data Quality Management
* Data Science (Data Mining, Machine Learning, Network Analytics)
* Semantics on the Web, Linked (Open) Data & schema.org
* Corporate Knowledge Graphs
* Knowledge Integration and Language Technologies
* Economics of Data, Data Services and Data Ecosystems
Following the success of previous years, the ‘horizontals’ (research)
and ‘verticals’ (industries) below are of interest for the conference:
Horizontals:
* Enterprise Linked Data & Data Integration
* Knowledge Discovery & Intelligent Search
* Business Models, Governance & Data Strategies
* Big Data & Text Analytics
* Data Portals & Knowledge Visualization
* Semantic Information Management
* Document Management & Content Management
* Terminology, Thesaurus & Ontology Management
* Smart Connectivity, Networking & Interlinking
* Smart Data & Semantics in IoT
* Semantics for IT Safety & Security
* Semantic Rules, Policies & Licensing
* Community, Social & Societal Aspects
Verticals:
* Industry & Engineering
* Life Sciences & Health Care
* Public Administration
* Galleries, Libraries, Archives & Museums (GLAM)
* Education & eLearning
* Media & Data Journalism
* Publishing, Marketing & Advertising
* Tourism & Recreation
* Financial & Insurance Industry
* Telecommunication & Mobile Services
* Sustainable Development: Climate, Water, Air, Ecology
* Energy, Smart Homes & Smart Grids
* Food, Agriculture & Farming
* Safety & Security
* Transport, Environment & Geospatial
Posters & Demos Track
The Posters & Demonstrations Track invites innovative work in progress,
late-breaking research and innovation results, and smaller contributions
in all fields related to the broadly understood Semantic Web. These
include submissions on innovative applications with impact on end users
such as demos of solutions that users may test or that are yet in the
conceptual phase, but are worth discussing, and also applications or
pieces of code that may attract developers and potential research or
business partners. This also concerns new data sets made publicly available.
The informal setting of the Posters & Demonstrations Track encourages
participants to present innovations to the research community, business
users and find new partners or clients and engage in discussions about
the presented work. Such discussions can be invaluable inputs for the
future work of the presenters, while offering conference participants an
effective way to broaden their knowledge of the emerging research trends
and to network with other researchers.
Poster and demo submissions should consist of a paper of 1-4 pages that
describe the work, its contribution to the field or novelty aspects.
Submissions must be original and must not have been submitted for
publication elsewhere. All submissions should follow the ACM ICPS
guidelines for formatting. The layout templates can be found here:
http://www.acm.org/sigs/publications/proceedings-templates. The best
posters (5-6 papers) will be published in the digital library of the ACM
ICP Series. The other papers will be published in the
http://ceur-ws.org/. Papers should be submitted through EasyChair
(https://easychair.org/conferences/?conf=semantics2016research). Papers
must be submitted in PDF (Adobe's Portable Document Format) format.
Other formats will not be accepted. For the camera-ready version, the
source files (Latex, Word) will also be needed.
Submissions will be reviewed by experienced and knowledgeable
researchers and practitioners; each submission will receive a detailed
feedback. For demos, it would be beneficial to include also links
enabling the reviewers testing the application or reviewing the component.
Important Dates (Posters & Demos)
* Submission Deadline: extended: July 1, 2016 (11:59 pm,
Hawaii time)
* Notification of Acceptance: extended: August 2, 2016 (11:59
pm, Hawaii time)
* Camera-Ready Paper: extended: August 10, 2016 (11:59
pm, Hawaii time)
Poster and Demo Chairs:
* Michael Martin, University of Leipzig
* Martí Cuquet, Semantic Technology Institute, University of Innsbruck
* Erwin Folmer, University of Twente, Kadaster and Geonovum
Contact email address: semantics2016postersdemos(a)gmail.com
Conference Chairs:
* Sebastian Hellmann, AKSW/KILT, InfAI, Leipzig University
* Tassilo Pellegrini, UAS St. Pölten
Hi,
I'm doing research on Wikipedia article quality, and I take advantage of
WikiProject Assessments. But I can only get the latest quality level of an
article. I wonder how to get the quality of each revision, or how to get
the exact date when an article get a quality promotion, for example, from
A-class to FA-class.
I really need your help! Thanks!
Zhang Shiyue
--
Zhang Shiyue
*Tel*: +86 18801167900
*E-mail*: byryuer(a)gmail.com, yuer3677(a)163.com
State Key Laboratory of Networking and Switching Technology
No.10 Xitucheng Road, Haidian District
Beijing University of Posts and Telecommunications
Beijing, China.
Hi Ben,
On 6/15/16 18:24, Benjamin Good wrote:
> Hi Marco,
>
> Where might we find some statistics on the current accuracy of the
> automated claim and reference extractors? I assume that information
> must be in there somewhere, but I had trouble finding it.
The StrepHit pipeline (codebase) is ready, while the project is ongoing.
We are not there yet, and will publish performance values in the final
report.
>
> This is a very ambitious project covering a very large technical
> territory (which I applaud). It would be great if your results could be
> synthesized a bit more clearly so we can understand where the
> weak/strong points are and where we might be able to help improve or
> make use of what you have done in other domains.
Sure, this will be done in the final report.
Up to now, you can have a look at the midpoint report summary:
https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
Best,
Marco
>
> -Ben
>
>
> On Wed, Jun 15, 2016 at 9:06 AM, Marco Fossati <fossati(a)spaziodati.eu
> <mailto:fossati@spaziodati.eu>> wrote:
>
> [Feel free to blame me if you read this more than once]
>
> To whom it may interest,
>
> Full of delight, I would like to announce the first beta release of
> *StrepHit*:
>
> https://github.com/Wikidata/StrepHit
>
> TL;DR: StrepHit is an intelligent reading agent that understands
> text and translates it into *referenced* Wikidata statements.
> It is a IEG project funded by the Wikimedia Foundation.
>
> Key features:
> -Web spiders to harvest a collection of documents (corpus) from
> reliable sources
> -automatic corpus analysis to understand the most meaningful verbs
> -sentences and semi-structured data extraction
> -train a machine learning classifier via crowdsourcing
> -*supervised and rule-based fact extraction from text*
> -Natural Language Processing utilities
> -parallel processing
>
> You can find all the details here:
> https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
> https://meta.wikimedia.org/wiki/Grants:IEG/StrepHit:_Wikidata_Statements_Va…
>
> If you like it, star it on GitHub!
>
> Best,
>
> Marco
>
> _______________________________________________
> Wikidata mailing list
> Wikidata(a)lists.wikimedia.org <mailto:Wikidata@lists.wikimedia.org>
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>
>
>
>
> _______________________________________________
> Wikidata mailing list
> Wikidata(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata
>