Hello,
I'm trying to write a pywikibot script which read and create items / properties on my Wikibase instance. Following pieces of tutorials and script examples, I managed to write something working.
1/ The idea is to read a CSV file, and create an item with its properties for each line. So I have to loop over thousands of lines and create an item and multiple claims associated, and it takes quite some time to do so. (atleast 1 hour to create 1000 items) I guess it's because for each line, I create a new entity and new claims, which means multiple requests for each line.
Some pseudo code I use in my script:
To create a new item, I use : repo.editEntity({}, {}, summary='new item') assuming repo = site.data_repository()
To create a new claim, I use : self.user_add_claim_unless_exists(item, claim), assuming my Bot inherit WikidataBot
Is there a better way to optimize that kind of bulk import ?
--
2/ I kind of have the same problem If I want to check if an item already exists, because first I need to get all existing items and check if they are in my CSV or not. (the CSV does not contain QIDs, but does contain a "custom" ID I've created and added as a property to each item )
--
I hope I was clear enough, any relevant example, idea, advice, would be much appreciated. Bear in mind I'm a beginner with the whole ecosystem so I'm open to any recommendation.
Thanks !