Hello,
I'm trying to write a pywikibot script which read and create items / properties on my
Wikibase instance. Following pieces of tutorials and script examples, I managed to write
something working.
1/ The idea is to read a CSV file, and create an item with its properties for each line.
So I have to loop over thousands of lines and create an item and multiple claims
associated, and it takes quite some time to do so. (atleast 1 hour to create 1000 items) I
guess it's because for each line, I create a new entity and new claims, which means
multiple requests for each line.
Some pseudo code I use in my script:
To create a new item, I use : repo.editEntity({}, {}, summary='new item') assuming
repo = site.data_repository()
To create a new claim, I use : self.user_add_claim_unless_exists(item, claim), assuming my
Bot inherit WikidataBot
Is there a better way to optimize that kind of bulk import ?
--
2/ I kind of have the same problem If I want to check if an item already exists, because
first I need to get all existing items and check if they are in my CSV or not. (the CSV
does not contain QIDs, but does contain a "custom" ID I've created and added
as a property to each item )
--
I hope I was clear enough, any relevant example, idea, advice, would be much appreciated.
Bear in mind I'm a beginner with the whole ecosystem so I'm open to any
recommendation.
Thanks !