Dear Wikimedia DE,
What are the plans for the upcoming Docker releases?
When can we expect the next release with a new major MediaWiki/Wikibase
version? What MediaWiki/Wikibase version will those include, and what PHP
versions will be supported?
In particular, will there be a release for MediaWiki 1.38, or will that
version be skipped, just like 1.37?
Best
--
Jeroen De Dauw | Technical Director Professional.Wiki
<https://www.Professional.Wiki>
We are MediaWiki, Wikibase and Semantic MediaWiki experts.
Professional.Wiki - Jeroen De Dauw & Karsten Hoffmeyer GbR
Tieckstraße 24-25, 10115 Berlin | +49 (30) 55 87 42 65 | USt-IdNr.
DE322440293
Hello All!
Could anyone one give us some guidance on when to use a bot as opposed to,
perhaps, a script fired off from a client box via cron or other?
Context: this is the PhiloBiblon project
<https://database.factgrid.de/wiki/Item:Q394169>, and we are using FactGrid
<https://database.factgrid.de/wiki/Main_Page>.
The use case I have in mind is integrity checking for certain constraints
that our legacy data platform enforced but that are not directly
enforceable in our current setup. For examples:
- Detecting when our external identifier
<https://database.factgrid.de/wiki/Property:P476> has an illegal value
(which has sometimes happened and has been difficult to detect by eye)
- Detecting when one of our objects (i.e. one that is marked with that
external identifier) has an item-valued property that is filled with an
object of the wrong type
I think it would give us some comfort to know that such integrity
violations are detected. We would probably like to add more constraints as
we go along. We might extend that code eventually to try to automatically
fix such things, but for now, just detecting them would be great.
Does a bot make sense for that use case?
Thanks,
Max Ziff
Good day everyone,
We've seen a problem on a self-hosted docker-compose based installation
regarding the blazegraph or wdqs index and the effect that deletion of
items, especially mass-deletion, has on it.
Now, I would have preferred to have a deeper understanding of this
before posting to the mailing list, but in this case, the wdqs service
really is like a black box to me and on top of that, I think that we're
in a situation where, regardless of the understanding of the underlying
cause, it won't change the current factual situation.
In a nutshell, what we observe is that at times, and I believe it might
be related to mass deletion ("nuking") of items, deleted items will not
be removed from the blazegraph index.
The main issue with that, besides simply returning false results, is
that some tools use both a SPARQL query, followed by an API request to
manipulate data, e.g. wikibaseintegrator.
With items remaining in the blazegraph index, but no longer existing on
mediawiki, this of course results in situations such as
{'name': 'wikibase-validator-no-such-entity', 'parameters':
['[[Item:Q342|Q342]]']…
Here, the SPARQL endpoint returned this item, but on the mediawiki
instance, it has long been deleted, resulting in an API error.
I unfortunately do not have logs of the requests made to the mediawiki
service by the wdqs-updater service back when this mass-deletion
happened, so I cannot tell if yes or no, all necessary requests were
made. I have traced the steps involved in deleting a single items and
seeing it being successfully deleted from the index though:
1. User: POST title=Item:Q3&action=delete
2. wdqs-updater -> mediawiki: api.php?action=query&list=recentchanges
3. mediawiki -> wdqs-updater: { type:log, title:Item:Q3, revid:0, …}
4. wdqs-updater -> mediawiki: Special:EntityData/Q3.ttl
I don't have full logs for the other case where the index does not get
updated correctly, but I do have for instance a log line like
18:31:27.750 [main] INFO o.w.q.r.t.change.RecentChangesPoller - Got 1
changes, from Q22@20220630163125|6522 to Q22@20220630163125|6522
And querying the SPARQL endpoint at 18:40 the Q22 item would still be
returned.
The logs from the wdqs container either to not contain any information
relevant to this problem or, alternatively, are so verbose that I'm
drowning in a sea of messages, being unable to understand what is going
on and even whether there is anything relevant to my problem or not.
My questions are these:
1. Is a problem like this known?
2. Is there any way to manually go into the blazegraph index and delete
that one record that no longer exists in mediawiki or to make
blazegraph purge it by somehow replaying the recentchanges entry, or
3. Will we need to drop the wdqs volume and recreate the entire index
from scratch with a sufficiently large $wgRCMaxAge value?
Thank you,
David Raison
--
*TenTwentyFour S.à r.l.*
www.tentwentyfour.lu <https://www.tentwentyfour.lu>
*T*: +352 20 211 1024
*F*: +352 20 211 1023
1 place de l'Hôtel de Ville
4138 Esch-sur-Alzette
FYI
---------- Forwarded message ---------
On Fri, Aug 12, 2022 at 4:56 PM Antonin Delpeuch <antonin(a)delpeuch.eu>
wrote:
> Hello Wikimedia community,
>
> I posted a message on Wikidata's project chat explaining that I would like
> to stop working on Wikidata integration in OpenRefine (which also applies
> to Wikimedia / Wikibase integration in general):
>
> Wikidata integration in OpenRefine: could this be your project?
> <https://www.wikidata.org/wiki/Wikidata:Project_chat#Wikidata_integration_in…>
>
>
> https://www.wikidata.org/wiki/Wikidata:Project_chat#Wikidata_integration_in…
> ?
>
> I do intend to keep working on OpenRefine and focus my work on the core
> tool. I am by far not the only contributor to Wikibase integration, but I
> want to make it clear that I am keen to let other contributors take the
> lead on this, because I do not have enough time myself to grow this part of
> the tool as much as it deserves.
>
> As I explain in Wikidata's project chat, I think this is a really nice
> project to work on, and you will be supported by Sandra Fauconnier, our
> project director, who is a keen Wikimedian.
>
> So: the floor is yours!
>
> If you are interested in the topic, join us today in our Wikimania
> hackathon session:
>
> https://phabricator.wikimedia.org/T314803
>
> Cheers,
> Antonin
> _______________________________________________
> Wikimedia-l mailing list -- wikimedia-l(a)lists.wikimedia.org, guidelines
> at: https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines and
> https://meta.wikimedia.org/wiki/Wikimedia-l
> Public archives at
> https://lists.wikimedia.org/hyperkitty/list/wikimedia-l@lists.wikimedia.org…
> To unsubscribe send an email to wikimedia-l-leave(a)lists.wikimedia.org
Hi all,
I am looking into better customising my WikiBase to the needs of my
community. I want to create a separate sub-heading (just like Wikidata) for
external identifiers and found some terse documentation on it here:
https://www.mediawiki.org/wiki/Wikibase/Installation/Advanced_configuration…
I would like to know which other options this feature can give me. Do any
of you have more information on the settings possibilities here? I'll see
if I can expand the manual page slightly with the outcomes of this
discussion.
The only example given (the configuration of Wikidata):
$wgWBRepoSettings['statementSections'] = [
'item' => [
'statements' => null,
'identifiers' => [
'type' => 'dataType',
'dataTypes' => [ 'external-id' ],
],
],];
Do I read this correctly when I read that everything is under the Header
'Statements' because there is no filter (null)? And that a new section
'identifiers' is created by filtering on dataType is one of 'DataTypes'
'external-id' here?
What are the other options for 'type' and what other variables need to be
created in order to make other types work?
Kind regards,
Maarten Zeinstra
--
IP Squared | +31643053919 | @mzeinstra