Thank you for making the measurements. Can you estimate the time for item
Q183 specifically? Since it is 1000 entities weighing 19 MB, this means
that on average the entities were 19 KB. Germany on the other hand is much
larger, and it makes we wonder how it scales to that size.
On Oct 7, 2014 6:47 AM, "Jeroen De Dauw" <jeroendedauw(a)gmail.com> wrote:
Hey,
I got some data to share. Walking through the dump for the first 1000
entities (~19Mb), took 0.008 seconds per item, where in each step the
following things where done:
* read line from file
* json_decode the line
* use the EntityDeserializer to turn the array into DataModel objects
Given that these entities are on average a lot bigger than the typical one
found in Wikidata, it looks like the average deserialization time is a few
milliseconds. So now I really wonder why people are blaming DataModel 1.0.
Everything seems to indicate most time is spend in Wikibase.git and
MediaWiki.
Cheers
--
Jeroen De Dauw -
http://www.bn2vs.com
Software craftsmanship advocate
Evil software architect at Wikimedia Germany
~=[,,_,,]:3
_______________________________________________
Wikidata-tech mailing list
Wikidata-tech(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-tech