[Toolserver-l] Capacity...what?

Lars Aronsson lars at aronsson.se
Tue Mar 17 04:47:49 UTC 2009


When I asked in what way, if any, the Swedish chapter could help 
the toolserver project, the short answer I got was that a new 
server would cost 8000 euro and a year's salary for an admin would 
cost even more.  If we had 8000 euro to spend (we don't) and 
nobody asked us what we use the money for, maybe we could get away 
with that.  But when we ask people for donations, they like to 
know what it's good for.  So what is the toolserver good for, 
really?  I think I know this, but how can I explain it to others?

The annual report (Tätigkeitsbericht) for 2007 for Wikimedia 
Deutschland mentions that 60,000 euro was used to double squid 
capacity and 10,000 euro was used for the toolserver cluster.  
The report for 2008 should be due soon, I guess.

I think it could be useful if the toolserver project could find a 
way to explain its current capacity, what kind of services this 
capacity is used for, and how much this capacity could be 
increased by a certain donation.  We could make a little folder or 
poster that explains this, and distribute among our fans.

The easiest would perhaps be to count the HTTP requests.  How many 
requests (or page views) does the toolserver receive per year, 
day, or second?  How are they distributed over the various 
services?  Is it possible to count how many unique visitors (IP 
addresses) each service has?  Divided by language or referer 
language of Wikipedia? Could the toolserver produce something 
similar to Domas' hourly visitor statistics for Wikipedia?

There's already http://toolserver.org/%7Einteriot/cgi-bin/tstoc 
where the right-most column indicates the number of individual IP 
adresses in the last 7 days.  Is there a log of these data?  It 
must be wrong (surely?) that only 8 tools had any accesses this 
week.  It could be correct that "21933 unique IP's visited the 
toolserver this week", but that number sounds rather low.

The toolserver hopefully spends far more processor cycles for each 
HTTP request than a squid cache. But just how many more? Every 
server has an economic life of maybe 3 years, while its purchasing 
price is written off, and it serves a number of HTTP requests in 
that time.  So what is the hardware (and hosting) cost of each 
request?

Of course, HTTP requests is not everything.  Is there a way we 
could measure other uses of the toolserver?

Getting back to my favorite: The geographic applications. Are the 
map tiles for the WikiMiniAtlas served from the toolserver?  How 
much power does this currently use for the Swedish Wikipedia? What 
if the maps were shown inline for every visitor, instead of as a 
pop-up that very few people click on?  How much extra would that 
cost?  Could such a service still run on the toolserver, or would 
it need to be moved to the servers in Florida?


-- 
  Lars Aronsson (lars at aronsson.se)
  Aronsson Datateknik - http://aronsson.se

  Wikimedia Sverige - stöd fri kunskap - http://wikimedia.se/



More information about the Toolserver-l mailing list