Hi!
[1] Is the service protected against internet crawlers that find such links in the online logs of this email list? It would be a pity if we would have to answer this query tens of thousands of times for many years to come just to please some spiders who have no use for the result.
That's a very good point. We currently do not have robots.txt file on the service. We should have it. I'll fix it ASAP.
GUI links do not run the query until click, so they are safe from bots anyway. But direct links to sparql endpoint do run the query (it's the API after all :) So robots.txt is needed there.