Hi all,
Following a refactoring of some of our weekly wikidata dump jobs, we had an
issue where a stale cron entry was still hanging around, which competed
with the new dump cronjobs. After having discovered the issue we've fixed
the root cause, but this week's dumps were unable to complete as a result.
New dumps will be available this upcoming week at the usual time(s).
*(Apologies for cross-posting)*
Hi all,
This is an announcement of a significant change to 3rd party Wikibase
installs that still use lib/WikibaseLib.php as an entry point.
*On September 30*, we will remove the lib/WikibaseLib.php file from the
Wikibase code. Wikis which directly load that file in their
LocalSettings.php will then fail to load (after the 1.36 release of
MediaWiki and Wikibase, if they do not track the master branch).
The file lib/Wikibase.php used to be the entry point of the WikibaseLib
extension. Directly loading this file has not been necessary since the
earliest days of Wikibase – it was loaded automatically by the entry points
of the Client and Repo extensions. Nevertheless, lines similar to the
following found their way into some Wikibase installation instructions:
require_once "$IP/extensions/Wikibase/lib/WikibaseLib.php";
If your LocalSettings.php contains such a line, please remove it. It has
not been needed for ages, and soon it will stop working.
This refactoring is part of a project with a twofold objective: (1)
decouple the Wikibase Client and Repo as much as possible and
establish a long-term
strategy for organising the shared logic
<https://doc.wikimedia.org/Wikibase/master/php/adr_0013.html>, and (2) get
both Client and Repo to use the standard extension registration mechanism
<https://www.mediawiki.org/wiki/Special:MyLanguage/Manual:Extension_registra…>
to be loaded. The impact of the project is focused on developers who can
now have a clearer view of the code, the modules involved, and their
interactions which makes their lives easier. Also, the maintenance of the
code is now easier. This is the phabricator board
<https://phabricator.wikimedia.org/project/profile/4815/> of the project.
If you have any issues or questions, please do not hesitate to ask.
Cheers,
--
Mohammed Sadat
*Community Communications Manager for Wikidata/Wikibase*
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Hello,
I wanted to ask if there is any canonical way to identify deletion, reverts
etc in the edit history xml files. I can understand that the action of
every revision is described in the "comment" element of the xml format, but
is there a code name or number or anything else that will help me to
identify one revision for example as deletion?
Thank you,
Elisavet
Hello all,
For the past four years, I’ve been working in the software department at
Wikimedia Germany, taking care of the communication between the Wikidata
development team <https://www.wikidata.org/wiki/Wikidata:DEV> and the
community, announcing new features, collecting bug reports and feature
requests from you. On top of that, I’ve been coordinating various projects,
bringing the WikidataCon
<https://meta.wikimedia.org/wiki/Grants:Conference/WMDE/WikidataCon/Report#I…>
to
life, coordinating the Wikidata decentralized birthday
<https://www.wikidata.org/wiki/Wikidata:Eighth_Birthday>, creating a
prototype for Wikidata Train the Trainers
<https://www.wikidata.org/wiki/Wikidata:Train_the_trainers>, and taking
care of various onsite or online workshops, meetups and other events.
Over the past years, with the Wikidata community growing, the development
team growing as well, more and more events happening, and the ecosystem of
Wikibase users forming a distinct group with different needs, it became
pretty clear that one person was not enough to keep track of everything and
provide the best support for the Wikidata editors. That’s the reason why,
earlier this year, we had the pleasure to announce the arrival of a new
colleague who you already know from being an active Wikidata editor, Mohammed
Sadat (WMDE) <https://www.wikidata.org/wiki/User:Mohammed_Sadat_(WMDE)>.
We already started a smooth transition of our roles: while Mohammed will
become the main person in charge of community communications for the
Wikidata and Wikibase communities, I will focus more on organizing
Wikidata-related events and supporting community members with their own
events and projects. As you may have noticed, Mohammed already took over
editing the weekly newsletter
<https://www.wikidata.org/wiki/Wikidata:Status_updates/Next>, monitoring
the social media, and various announcements for Wikidata and Wikibase. As
for myself, I will not disappear completely from the Wikidata channels: I
will keep supporting Mohammed on community communication, for example with
projects like the Wikidata Bridge
<https://www.mediawiki.org/wiki/Wikidata_Bridge>, in which I’ve been
involved since the start.
During this transition phase, we will review and improve our existing
communication processes, and you can for example give feedback on the
experience you had while reporting bugs or feature requests
<https://www.wikidata.org/wiki/Wikidata:Contact_the_development_team/Process…>.
Feel free to reach out to Mohammed if you have any questions regarding
Wikidata’s
development roadmap
<https://www.wikidata.org/wiki/Wikidata:Development_plan>.
I’m looking forward to continuing working with you on various projects:
feel free to contact me if you want to discuss Wikidata-related events,
training, online events, or any other ideas you have in mind to gather the
Wikidata community and onboard new editors.
Cheers,
--
Léa Lacroix
Community Engagement Coordinator
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Apologies for cross-posting
Dear DBpedians, Linked Data savvies and Ontologists,
We would like to invite you to join the DBpedia Autumn Hackathon 2020 as
a new format to contribute to DBpedia, gain fame, win small prizes and
experience the latest technology provided by DBpedia Association
members. The hackathon is part of the Knowledge Graphs in Action
conference on October 6, 2020. Please check here:
https://wiki.dbpedia.org/meetings/KnowledgeGraphsInAction
# Timeline
*
Registration of participants - main communication channel will be
the #hackathon channel in DBpedia Slack (sign up
https://dbpedia-slack.herokuapp.com/, then add yourself to the
channel). If you wish to receive a reminder email on Sep 21st, you
can leave your email address in this form: https://tinyurl.com/y24ps5jt
*
Until September 14th - preparation phase, participating
organisations prepare details, track formation, additional tracks
can be proposed, please contact dbpedia-events(a)infai.org
<mailto:dbpedia-events@infai.org>
*
September 21st - Announcement of details for each track, including
prizes, participating data, demos, tools and tasks. Check updates on
hackathon website
https://wiki.dbpedia.org/events/dbpedia-autumn-hackathon-2020
*
September 21st to October 1st - hacking period, coordinated via
DBpedia slack
*
October 1st, 23:59 Hawaii Time - Submission of hacking result (3
min video and 2-3 paragraph summary with links, if not stated
otherwise in the track)
*
October 5th, 16:00 CEST - Final Event, each track chair presents a
short recap of the track, announces prizes or summarizes the result
of hacking.
*
October 6th, 9:50 - 15:30 CEST - Knowledge Graphs in Action Event
*
Results and videos are documented on the DBpedia Website and the
DBpedia Youtube channel.
# Member Tracks
The member tracks are hosted by DBpedia Association members, who are
technology leaders in the area of Knowledge Engineering. Additional
tracks can be proposed until Sep 14th, please contact
dbpedia-events(a)infai.org <mailto:dbpedia-events@infai.org>.
*
timbr SQL Knowledge Graph: Learn how to model, map and query
ontologies in timbr and then model an ontology of GDELT, map it to
the GDELT database, and answer a number of questions that currently
are quite impossible to get from the BigQuery GDELT database. Cash
prizes planned. https://www.timbr.ai/
*
GNOSS Knowledge Graph Builder: Give meaning to your organisation’s
documents and data with a Knowledge Graph.
https://www.gnoss.com/en/products/semantic-framework
*
ImageSnippets: Labeling images with semantic descriptions. Use
DBpedia spotlight and an entity matching lookup to select DBpedia
terms to describe images. Then explore the resulting dataset through
searches over inference graphs and explore the ImageSnippets dataset
through our SPARQL endpoint. Prizes planned.
http://www.imagesnippets.com
*
Diffbot: Build Your Own Knowledge Graph! Use the Natural Language
API to extract triples from natural language text and expand these
triples with data from the Diffbot Knowledge Graph (10+ billion
entities, 1+ trillion facts). Check out the demo
http://demo.nl.diffbot.com/. All participants will receive access to
the Diffbot KG and tools for (non-commercial) research for one year
($10,000 value).
# Dutch National Knowledge Graph Track
Following the DBpedia FlexiFusion approach, we are currently
flexi-fusing a huge, dbpedia-style knowledge graph that will connect
many Linked Data sources and data silos relevant to the country of the
Netherlands. We hope that this will eventually crystallize a
well-connected sub-community linked open data (LOD) cloud in the same
manner as DBpedia crystallized the original LOD cloud with some
improvements (you could call it LOD Mark II). Data and hackathon details
will be announced on 21st of September.
# Improve DBpedia Track
A community track, where everybody can participate and contribute in
improving existing DBpedia components, in particular the extraction
framework, the mappings, the ontology, data quality test cases, new
extractors, links and other extensions. Best individual contributions
will be acknowledged on the DBpedia website by anointing the WebID/Foaf
profile.
(chaired by Milan Dojchinovski and Marvin Hofer from the DBpedia
Association & InfAI and the DBpedia Hacking Committee)
# DBpedia Open Innovation Track
(not part of the hackathon, pre-announcement)
For the DBpedia Spring Event 2021, we are planning an Open Innovation
Track, where DBpedians can showcase their applications. This endeavour
will not be part of the hackathon as we are looking for significant
showcases with development effort of months & years built on the core
infrastructure of DBpedia such as the SPARQL endpoint, the data, lookup,
spotlight, DBpedia Live, etc. Details will be announced during the
Hackathon Final Event on October 5.
(chaired by Heiko Paulheim et al.)
Stay tuned and stay safe!
With kind regards,
The DBpedia Organizing-Team
Hello all!
A quick update on what's going on around our SPARQL endpoints.
* Wikimedia Commons Query Service (WCQS) [1] is available as a beta
service. We've seen a number of people starting to run queries. And a
number of examples have been added [2]. Thanks all for your help!
* We are focusing again on WDQS and improving the update process [3]. So
far, we have an end-to-end working example for simple updates (revision
create) and are working on adding support for more complex updates
(deletes, undeletes, suppressed deletes, etc...). Once this all process
is complete and working for WDQS, we'll see how we can adapt it for WCQS
and have streaming updates to WCQS.
* We are looking into the deployment constraints for the new WDQS update
process. Managing Flink at scale is non trivial, we are just starting, but
there is a lot more work to make this robust.
* We are planning to spend more time doing some analytics on our data [4].
We want to better understand the use cases and the data we have. We are
still defining exactly what question we want to answer from the data, but
the main ones are
** What are the most expensive queries, what are they trying to achieve and
is that reasonable
** Do we have performant subgraphs that we could expose indepently.
This will also require some work to improve our query logging and aggregate
more context with the queries we log.
That's all for today!
Have fun!
Guillaume
[1] https://wcqs-beta.wmflabs.org/
[2]
https://commons.wikimedia.org/wiki/Commons:SPARQL_query_service/queries/exa…
[3] https://phabricator.wikimedia.org/T244590
[4] https://phabricator.wikimedia.org/T257045
--
Guillaume Lederrey
Engineering Manager, Search Platform
Wikimedia Foundation
UTC+1 / CET