Hello Mamdouh,
As far as I know that is not possible, but you can also download the whole
dataset as a dump and process it (e.g. query over the text data or set up
your own SPARQL endpoint):
https://www.wikidata.org/wiki/Wikidata:Database_download (JSON or RDF dumps
are probably most helpful).
Depending on your usecase this might be the right direction to look into :)
Feel free to reach out if there are more questions.
Best,
Lucie
On Wed, 10 Apr 2019 at 09:14, Ahmed Mamdouh <ahmed.mamdouh24(a)yahoo.com>
wrote:
Greetings All,
Hope this e-mail finds you well. I am currently doing a master project in
NLP in JKU under the supervision of Prof. Bruno Buchberger the famous
Austrian Mathematician.
I am facing a problem where I can’t get enough data for my project. So is
there anything that can be done to extend the limit of queries as they
timeout ?
Thanks in advance,
Mamdouh
_______________________________________________
Wikidata mailing list
Wikidata(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata
--
Lucie-Aimée Kaffee