Hello everyone,
Since 2019, it is possible to add structured data to files on Wikimedia
Commons [1] (SDC = Structured Data on Commons). But there are no very
advanced and user-friendly tools yet to edit the structured data of very
large and very diverse batches of files on Commons. And there is no batch
upload tool yet that supports SDC.
The OpenRefine [2] community wants to fill this gap: in the upcoming year,
we would like to build brand new features in the open source OpenRefine
tool, allowing batch editing and batch uploading SDC :-) As these are major
new functionalities in OpenRefine, we have applied for a Project Grant [3].
Your feedback [4] and (if you support this plan) endorsements are very
welcome.
Thanks in advance, and many greetings,
Sandra (User:Spinster / User:SFauconnier) as member of the OpenRefine
steering committee
Antonin (User:Pintoch) as OpenRefine developer
[1] https://commons.wikimedia.org/wiki/Commons:Structured_data
[2] https://www.wikidata.org/wiki/Wikidata:Tools/OpenRefine
[3]
https://meta.wikimedia.org/wiki/Grants:Project/Structured_Data_on_Wikimedia…
[4]
https://meta.wikimedia.org/wiki/Grants_talk:Project/Structured_Data_on_Wiki…
Hello all,
It has been exciting to see the increased adoption and usage of the
Wikidata Query Service (WDQS) usage in the past year. To support this
growing demand, on 15 March 2021 the Search Platform team released a new
Streaming Updater to a test server https://query-preview.wikidata.org for
feedback before going to production on 15 April 2021 (pending any major
blockers discovered during testing). Once in production, WDQS will become
less of a bottleneck for Wikidata updates, and we’re looking forward to
better facilitating Wikidata’s continued growth as a more complete
knowledge graph.
Your relevant feedback on the following changes is important to us to
ensure we continue to best support your needs while scaling up the service
in production:
1. New Streaming Updater: [1]
-
This improvement to the Updater will allow WDQS to better handle the
volume of edits to Wikidata, improving data consistency and decreasing
update latency: while the existing Updater fluctuates between 5–15
updates/sec (averaging 10 updates/sec), the new Updater will be
able handle
a throughput of 40–130 updates/sec (88 updates/sec on average). Without
these performance improvements, edits to Wikidata were being throttled
<https://phabricator.wikimedia.org/T243701>, approaching the point
where they could become impossible. With the new Updater, edits
to Wikidata
will be on the whole more consistent and have less lag, reducing the WDQS
bottleneck to improving Wikidata content.
-
We don’t anticipate this to adversely affect workflows or usage, but
it is a big update, and we would like you to let us know if you find any
related bugs or problems so that we can properly address them.
2. Blank node skolemization: [2]
-
To reliably use the new Streaming Updater to minimize the throttling
of edits to Wikidata, skolemization of blank nodes was required, as
detailed in the phabricator ticket. For more detail on why this was
necessary, you can also refer to another attempt to design a “diff”
format for RDF <https://www.w3.org/2001/sw/wiki/TurtlePatch>, where
the solution suggested to handle blank nodes is also skolemization. We
understand that this solution will unfortunately potentially introduce
breaking changes to your usage of WDQS, RDF dumps, and
Special:EntityData;
however, given the severe risk of the edits to Wikidata becoming
impossible, we felt this was the best course of action to take in the
timeframe we had. We acknowledge that this approach has its shortcomings,
however, and invite you to provide us with feedback on how we can improve
future usage of Wikidata and WDQS while maintaining their scalability and
reliability.
-
From a user perspective of this change, (1) queries using isBlank()
will need to be rewritten; (2) queries using isIRI/isURI will need to be
verified; (3) WDQS results will no longer include blank nodes. If these
changes affect your workflows, and/or you need to know how to modify your
workflows to account for the blank node skolemization, please let us know
what your specific use case is.
-
For more detail on how to modify your workflows, including examples,
please refer to the following page:
https://www.mediawiki.org/wiki/Wikidata_Query_Service/Blank_Node_Skolemizat…
3. Constraint fetching [3]
-
Constraints are a Wikibase concept that allows entities to be
validated based on definable properties: i.e. all astronauts
must be human.
Ideally, constraint fetching would be used to ensure data quality for
Wikidata edits. The reality is that the current implementation of
constraints fetching is not meeting our production quality standards and
was generating detrimental noise in our logs.
-
As a result of the sub-par implementation, and the fact that the new
Flink-based Streaming Updater doesn’t support it, current constraint
fetching functionality will be disabled with the new Updater
release, until
we can expose constraint violations in a more production-ready way
[4][5][6]. We recognize that even functionality that doesn’t meet our
production quality standards is still potentially useful for some, and we
would like to hear your feedback if you are adversely affected by this
change.
We’re looking forward to these new changes improving WDQS, and your
relevant feedback on these updates will help us make sure we can continue
to support your needs. If you have any questions, issues or suggestions,
feel free to reach out to us on the WDQS contact page
<https://www.wikidata.org/wiki/Wikidata:Contact_the_development_team/Query_S…>
.
original announcement on Wikidata Project Chat:
https://www.wikidata.org/wiki/Wikidata:Project_chat#New_WDQS_Streaming_Upda…
[1] - https://phabricator.wikimedia.org/T244590
[2] - https://phabricator.wikimedia.org/T244341
[3] - https://phabricator.wikimedia.org/T274982
[4] - https://phabricator.wikimedia.org/T204024
[5] - https://phabricator.wikimedia.org/T201147
[6] - https://phabricator.wikimedia.org/T201150
—
Mike Pham (he/him)
Sr Product Manager, Search Platform
Wikimedia Foundation <https://wikimediafoundation.org/>
Hi,
Is there a way to get the child properties/sub-properties of a property?
Example: position_held(P39) property has sub-properties like start time
and end time so how to get them directly with SPARQL query and also If I
have a wiki dump which is around 65 GB so how to parse a huge wiki dump
and get the properties and their related sub-properties?
I got all the property IDs and property labels by SPARQL query but its
inner properties like start time, end time.. so on were missed.
Thanks in advance!!
Thanks and Regards,
VijayaLakshmi.
Hi,
Is there a way to get the child properties/sub-properties of a property?
Example: position_held(P39) property has sub-properties like start time
and end time so how to get them directly with SPARQL query and also If I
have a wiki dump which is around 65 GB so how to parse a huge wiki dump
and get the properties and their related sub-properties?
Thanks,
Vijaya.
Hi there.
About the new development of the lexicographical data and Abstract
Wikipedia [1]: Can you estimate how much involvement (time / effort /
energy ...) is supposed from the language community that will be selected
for focus? The description says about "particularly active feedback
channels" but different lang communities can interpret it differently :)
E.g. can you compare it to situation for first languages that have received
ContentTranslation?
In the Esperanto community we discuss our possible application, but the
topic of needed resources is important for us.
Thanks!
[1]
https://www.wikidata.org/wiki/Wikidata:Lexicographical_data/Focus_languages
KuboF Hromoslav
[My apologies for cross-posting. I realized I missed this important list.:)]
Hi all,
This email contains information about a new award that we're
introducing as well as instructions about how you can nominate
scholarly research publications for the award.
=Purpose of the award=
Recognize recent research on or about the Wikimedia projects or recent
research that is of importance to the Wikimedia projects. Recognize
the researchers behind the research.
=Call for Nominations=
We invite you to submit nominations for the Wikimedia Foundation
Research Award of the year, which will be handed out at Wiki Workshop
2021 [1].
==Eligibility criteria==
Your nomination must meet the following criteria:
* The research must be on, about, using data from, and/or of
importance to Wikipedia, Wikidata, Wikimedia Commons or other
Wikimedia projects [2].
* The publication must be available in English.
* The research must have been published between January 1, 2020 to
December 31, 2020.
==Selection criteria==
The award committee [3] will assess nominations based on
methodological excellence, quality of execution, quality of
presentation, reproducibility, ethical considerations, potential
impact on the Wikimedia projects or the Wikimedia research.
==Nomination process==
If you are interested to nominate a scholarly research publication,
please do so no later than 2021-03-22 by uploading it to Easychair via
https://easychair.org/conferences/?conf=wmfray2020 . We will ask you
to provide the following information in your nomination:
* Title of the manuscript
* A copy of the manuscript you are nominating
* A summary of the research and a clear justification for why the work
merits the award (in 350 words or fewer in English).
Note that self-nominations and nominations of others' work are both welcome.
If you have any questions, please contact wmf-ray-2020(a)easychair.org.
Best,
Leila Zia, on behalf of the Selection Committee of the Wikimedia
Foundation Research Award of the Year
[1] https://wikiworkshop.org/2021/
[2] https://wikimediafoundation.org/our-work/wikimedia-projects/
[3] comprised of myself, Benjamin Mako Hill (University of Washington)
and Aaron Shaw (Northwestern University) as of now. More people may be
added depending on the diversity of the nominations we receive and as
expertise are required.
--
Leila Zia
Head of Research
Wikimedia Foundation
Hello all,
I'm glad to announce that *30 Lexic-o-days
<https://www.wikidata.org/wiki/Wikidata:Events/30_lexic-o-days_2021>*, a
series of events, projects and challenges around *lexicographical data*,
will start on *March 15th*. There will be discussions, presentations, but
also activities like improving the documentation of Lexemes or editing
challenges. The goals of this event is to gather people editing Lexemes to
have discussions around the content and work together. You can find the
schedule and all relevant links on this page
<https://www.wikidata.org/wiki/Wikidata:Events/30_lexic-o-days_2021>.
This format is a first experiment and its content is powered by the
community: if you have ideas or wishes for the discussions, you're very
welcome to set up an appointment or to create a task on the related Phabricator
board <https://phabricator.wikimedia.org/tag/wikidata-lexicodays-2021/>!
We're also keeping an open list of ideas here
<https://etherpad.wikimedia.org/p/30Lexicodays2021>. Discussions about
Lexemes, or summaries of future discussions that will take place during the
event, should be documented on the project page
<https://www.wikidata.org/wiki/Wikidata:Lexicographical_data> or its talk
page.
If you have questions or need help to participate, feel free to contact me.
I'm looking forward to your participation!
Cheers,
--
Léa Lacroix
Community Engagement Coordinator
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/029/42207.
Hello,
As you may recall, the Wikidata development team at Wikimedia Germany requested
for a CentralNotice banner
<https://meta.wikimedia.org/wiki/CentralNotice/Request#Wikidata_Community_Su…>
earlier in February to deploy the Wikidata user diversity survey. The
request has been approved and the banner will be activated and displayed
for 10 days from Wednesday 10th March, 2021 and closed on Saturday 20th
March, 2021.
Your responses to this survey will help us to understand the core
demographics that make up the Wikidata community in order to provide us
with a baseline for future diversity efforts.
The survey is very brief and will only take 1-2 minutes to complete.
Link to Survey (open until Monday 22nd March, 2021):
https://wikimedia.sslsurvey.de/Wikidata_Community_in_2021
More information about the survey, the use of the data, and the questions
asked: Wikidata:Usability and usefulness/2021-2-Survey
<https://www.wikidata.org/wiki/Wikidata:Usability_and_usefulness/2021-2-Surv…>
Many thanks for your participation. If you have any questions please do not
hesitate to ask.
Cheers,
--
Mohammed Sadat
*Community Communications Manager for Wikidata/Wikibase*
Wikimedia Deutschland e.V.
Tempelhofer Ufer 23-24
10963 Berlin
www.wikimedia.de
Dear friends and Wikimedians,
(apologies for cross-posting!)
The Community Resources team announced in January the intention and process
to review its current grants programs (Annual Plan Grants, Simple Annual
Plan Grants, Project Grants and Rapid Grants) to ensure we are aligning
with the strategic direction of the Wikimedia movement, with a focus on
knowledge equity and building a thriving movement[1][2].
Now that we have finished our needs review, key stakeholders discussions,
community brainstorms, and grantee surveys, we are excited to share with
you the draft proposal[3] that has resulted from this work:
<
https://meta.wikimedia.org/wiki/Community_Resources/Grants_Strategy_Relaunc…
>
We are requesting your feedback on how the proposed grant programs would
impact you and how well they align with our movement strategy, which can be
provided in the following ways [4]:
* Complete a feedback form
* Participate in office hours
* Discuss on-wiki
* E-mail the Community Resources team (<communityresources(a)wikimedia.org>)
More information about these options can be found on our feedback page.[4]
After one month when the planned feedback period is over, the Community
Resources team will review all community input and make final adjustments
to complete the redesign. In April 2021, the Community Resources team will
review and incorporate suggested changes and send a final proposal for
endorsement and approval by the leadership of the Wikimedia Foundation.
Working with all our communities to update our grants programs is essential
to ensuring that they are useful and address real needs for our
communities. Thank you for all of the participation that has brought us to
this point. We look forward to the continued engagement in this next step
and listening to your ideas.
On behalf of the Community Resources team,
Chris
---
[1] https://meta.wikimedia.org/wiki/Strategy/Wikimedia_movement/2018-20
[2]
https://meta.wikimedia.org/wiki/Wikimedia_Foundation_Medium-term_plan_2019/…
[3]
https://meta.wikimedia.org/wiki/Community_Resources/Grants_Strategy_Relaunc…
[4]
https://meta.wikimedia.org/wiki/Community_Resources/Grants_Strategy_Relaunc…
*Chris Schilling* (him//they)
User:I JethroBT (WMF)
Senior Program Officer, Rapid Grants
Wikimedia Foundation