I have done some work on reorganizing the squid logs and including the
stats on individual pages through AJAX. I have also done some work on a
possible system to accumulate similar logs at the server level, but only
for servers without squids in front of them. The first one is in a demo
state, the last one is very sketchy.
John
Andrew Whitworth skrev:
On Wed, Aug 27, 2008 at 3:29 PM,
<WJhonson(a)aol.com> wrote:
What is a squid hit log?
Squid is a cache server program. Wikimedia uses several squid servers
to decrease the load on the application servers. Incoming web page
requests are able to be serviced by cached versions in the squid
server without having to post a request to the application servers. A
squid log is the logfile of page accesses kept by the squid server.
Because of this aggressive use of squids, Wikipedia can't maintain a
count of page hits. The reason being that most page hits don't
interact with the application servers at all and therefore cannot be
counted. I know people have experimented with workarounds, such as
reading page hit data from the squid servers and aggregating that data
into some external website. I haven't heard of any recent efforts like
that, however, and none which are able to be put into production use
by the foundation.
--Andrew Whitworth
_______________________________________________
Wikiquality-l mailing list
Wikiquality-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikiquality-l