What makes it so slow?
Note that you can use wbeditentity to perform complex edits with a single api
call. It's not as streight forward to use as, say, wbaddclaim, but much more
powerfull and efficient.
-- daniel
Am 30.09.2014 19:00, schrieb Andra Waagmeester:
Hi All,
I have joined the development team of the ProteinBoxBot
(
https://www.wikidata.org/wiki/User:ProteinBoxBot) . Our goal is to make
Wikidata the canonical resource for referencing and translating identifiers for
genes and proteins from different species.
Currently adding all genes from the human genome and their related identifiers
to Wikidata takes more then a month to complete. With the objective to add other
species, as well as having frequent updates for each of the genomes, it would be
convenient if we could increase this throughput.
Would it be accepted if we increase the throughput by running multiple instances
of ProteinBoxBot in parallel. If so, what would be an accepted number of
parallel instances of a bot to run? We can run multiple instances from different
geographical locations if necessary.
Kind regards,
Andra
_______________________________________________
Wikidata-l mailing list
Wikidata-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l
--
Daniel Kinzler
Senior Software Developer
Wikimedia Deutschland
Gesellschaft zur Förderung Freien Wissens e.V.