Hello Mamdouh,

As far as I know that is not possible, but you can also download the whole dataset as a dump and process it (e.g. query over the text data or set up your own SPARQL endpoint): https://www.wikidata.org/wiki/Wikidata:Database_download (JSON or RDF dumps are probably most helpful).
Depending on your usecase this might be the right direction to look into :)

Feel free to reach out if there are more questions.
Best,
Lucie

On Wed, 10 Apr 2019 at 09:14, Ahmed Mamdouh <ahmed.mamdouh24@yahoo.com> wrote:
Greetings All,

Hope this e-mail finds you well. I am currently doing a master project in NLP in JKU under the supervision of Prof. Bruno Buchberger the famous Austrian Mathematician.

I am facing a problem where I can’t get enough data for my project. So is there anything that can be done to extend the limit of queries as they timeout ?

Thanks in advance,
Mamdouh
_______________________________________________
Wikidata mailing list
Wikidata@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata


--
Lucie-Aimée Kaffee