Hoi,
Shopping for recognition for "your" project. Making it an issue that has to affect everything else because the quality of "your" project is highly problematic given a few basic facts. Wikidata has 59,284,641 items, this effort is about 7500 people. They drown in a sea of other people, items. Statistically the numbers involved are insignificant.

HOWEVER, when your effort has a practical application, it is that application, the use of the data that ensures that the data will be maintained and hopefully ensure that the quality of this subset is maintained. When you want quality, static quality is achieved by restricting at the gate. Dynamic quality is achieved by making sure that the data is actually used. Scholia is an example of functionality that supports existing data and everyone who uses it will see the flaws in the data. It is why we need to import additional data, merge scientist when there are duplicates, add additional authorities.  

Yes, we need to achieve quality results. They will be achieved when people use the data, find its flaws and consequently append and amend. Recognition of quality is done best by supporting and highlighting the application of our data and particularly be thankful to the consequential updates we receive. The users that help us do better are our partners, all others ensure our relevance.
Thanks,
      GerardM

On Wed, 28 Aug 2019 at 00:52, Magnus Sälgö <salgo60@msn.com> wrote:
Uwe I feel this is more and more important with quality and provenance and also communicate inside Wikidata the quality of our data. 

 I have added maybe the best source for biographies in Sweden P3217 in Wikidata on 7500 person. In Wikipedia those 7500 objects are used on > 200 different languages in Wikipedia we need to have a ”layer” explaining that data confirmed  with P3217 ”SBL from Sweden” has very high trust 

See https://phabricator.wikimedia.org/T222142

I can also see this quality problem that  Nobelprize.org and Wikidata has > 30 differencies and its sometimes difficult to understand the quality of the sources in Wikidata plus that Nobelprize.com has no sources makes the equation difficult
https://phabricator.wikimedia.org/T200668

Regards
Magnus Sälgö
0046-705937579


28 aug. 2019 kl. 03:49 skrev Uwe Jung <jung.uwe@gmail.com>:

Hello,

many thanks for the answers to my contribution from 24.8.
I think that all four opinions contain important things to consider.

@David Abián
I have read the article and agree that in the end the users decide which data is good for them or not.

@GerardM
It is true that in a possible implementation of the idea, the aspect of computing load must be taken into account right from the beginning.

Please check that I have not given up on the idea yet. With regard to the acceptance of Wikidata, I consider a quality indicator of some kind to be absolutely necessary. There will be a lot of ordinary users who would like to see something like this.

At the same time I completely agree with David;(almost) every chosen indicator is subject to a certain arbitrariness in the selection. There won't be one easy to understand super-indicator.
So, let's approach things from the other side. Instead of a global indicator, a separate indicator should be developed for each quality dimension to be considered. With some dimensions this should be relatively easy. For others it could take years until we have agreed on an algorithm for their calculation.

Furthermore, the indicators should not represent discrete values but a continuum of values. No traffic light statements (i.e.: good, medium, bad) should be made. Rather, when displaying the qualifiers, the value could be related to the values of all other objects (e.g. the value x for the current data object in relation to the overall average for all objects for this indicator). The advantage here is that the total average can increase over time, meaning that the position of the value for an individual object can also decrease over time.

Another advantage: Users can define the required quality level themselves. If, for example, you have high demands on accuracy but few demands on the completeness of the statements, you can do this.

However, it remains important that these indicators (i.e. the evaluation of the individual item) must be stored together with the item and can be queried together with the data using SPARQL.

Greetings

Uwe Jung

Am Sa., 24. Aug. 2019 um 13:54 Uhr schrieb Uwe Jung <jung.uwe@gmail.com>:
Hello,

As the importance of Wikidata increases, so do the demands on the quality of the data. I would like to put the following proposal up for discussion.

Two basic ideas:
  1. Each Wikidata page (item) is scored after each editing. This score should express different dimensions of data quality in a quickly manageable way.
  2. A property is created via which the item refers to the score value. Certain qualifiers can be used for a more detailed description (e.g. time of calculation, algorithm used to calculate the score value, etc.).

The score value can be calculated either within Wikibase after each data change or "externally" by a bot. For the calculation can be used among other things: Number of constraints, completeness of references, degree of completeness in relation to the underlying ontology, etc. There are already some interesting discussions on the question of data quality which can be used here ( see  https://www.wikidata.org/wiki/Wikidata:Item_quality; https://www.wikidata.org/wiki/Wikidata:WikiProject_Data_Quality, etc).

Advantages
  • Users get a quick overview of the quality of a page (item).
  • SPARQL can be used to query only those items that meet a certain quality level.
  • The idea would probably be relatively easy to implement.

Disadvantage:
  • In a way, the data model is abused by generating statements that no longer describe the item itself, but make statements about the representation of this item in Wikidata.
  • Additional computing power must be provided for the regular calculation of all changed items.
  • Only the quality of pages is referred to. If it is insufficient, the changes still have to be made manually.

I would now be interested in the following:
  1. Is this idea suitable to effectively help solve existing quality problems?
  2. Which quality dimensions should the score value represent?
  3. Which quality dimension can be calculated with reasonable effort?
  4. How to calculate and represent them?
  5. Which is the most suitable way to further discuss and implement this idea?

Many thanks in advance.

Uwe Jung  (UJung)


_______________________________________________
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata
_______________________________________________
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata