[Wikipedia-l] what's being done about performance?
Neil Harris
usenet at tonal.clara.co.uk
Sat Jul 13 13:45:12 UTC 2002
Fred Bauder wrote:
>At 02:29 AM 7/13/02 -0700, maveric149 wrote:
>
>
>>Much is now being done to remedy performance problems -- so I do believe
>>
>>
>what
>
>
>>you said is needlessly rude (even if there is a grain of truth to it).
>>
>>
>
>My experience with coders has been in MUDs. Coders like to make new areas,
>tweak races and guilds, etc. Performance issues are not nearly as
>attractive nor is success in dealing with them easily attainable or
>recognizable to others. Like a MUD we are relying on voluntary coders so,
>sure, courtesy is necessary.
>
>But so is focus. Right now Wikipedia is like some Ionesco play with a dead
>body in the room that everyone inexplicably ignores.
>
>Fred Bauder
>
>[Wikipedia-l]
>To manage your subscription to this list, please go here:
>http://www.nupedia.com/mailman/listinfo/wikipedia-l
>
>
>
>
>
Hi Fred,
I'm not a Wikipedia coder, I mostly contribute articles. Wearing my
contributor hat, I've been really irritated by the disastrous
performance in the last couple of months. My style of editing is
pointillistic, with lots of tiny edits, even breaking down large
articles into a series of smaller commits. I can understand that the
poor performance will have discouraged many people.
Having said that, I am helping test the new system at the moment, and I
can say that the coders are putting a lot of effort into making the
system fast and smooth. I think they've taken the right decision, to
give up on maintaining the old system, and work instead on getting the
new system up and running really well before putting it into service
ASAP. That means testing it really hard before going live.
Here's a progress report.
The new server is a dual-Athlon multiprocessor SMP machine with 2G of
RAM, and an ext3 journalling filesystem.
At the moment, the new server is running a hugely expanded database
containing over 80,000 'articles', making a total of over 100,000
entries in the database. This consists of a snapshot of the entire
Wikipedia, together with about 50,000 machine-generated articles. It's
serving around 1.2 articles per second under as realistic a traffic load
as we can generate using lots of bots running on external servers - the
traffic is nice and bursty, and the system is regualrly peaking at 12
articles/second or more.
77% of pages are served in less than a second: what I'd consider to be
'instant'
89% of pages are served in less than 2 seconds: what I'd consider to be
'quick'
98% of pages are served in less than 5 seconds, about the start of the
threshold of irritation
99.3% of pages are served in less than 10 seconds, a significant delay
Only 0.1% of articles take longer than 20 seconds to serve: what I would
regard as failures.
About 30% of these are very big special pages. I'm trying to do a finer
analysis of the rest, and find the exact conditions that seem to trigger
the hiccups.
What this means, I hope, is that when the new server comes into service
it will perform really well under normal loads, and editing will be
smooth and lovely again.
Neil
More information about the Wikipedia-l
mailing list