Hello Mamdouh,
As far as I know that is not possible, but you can also download the whole dataset as a dump and process it (e.g. query over the text data or set up your own SPARQL endpoint): https://www.wikidata.org/wiki/Wikidata:Database_download (JSON or RDF dumps are probably most helpful). Depending on your usecase this might be the right direction to look into :)
Feel free to reach out if there are more questions. Best, Lucie
On Wed, 10 Apr 2019 at 09:14, Ahmed Mamdouh ahmed.mamdouh24@yahoo.com wrote:
Greetings All,
Hope this e-mail finds you well. I am currently doing a master project in NLP in JKU under the supervision of Prof. Bruno Buchberger the famous Austrian Mathematician.
I am facing a problem where I can’t get enough data for my project. So is there anything that can be done to extend the limit of queries as they timeout ?
Thanks in advance, Mamdouh _______________________________________________ Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata