The WMF aggregates them as (page,views) pairs on an hourly basis:
http://dumps.wikimedia.org/other/pagecounts-raw/
I've been parsing these and storing them in a query-able DB format (for
en.wp exclusively; though the files are available for all projects I
think) for about two years. If you want to maintain such a fine
granularity, it can quickly become a terrabyte scale task that eats up a
lot of processing time.
If your looking for more coarse granularity reports (like top views for
day, week, month) a lot of efficient aggregation can be done.
See also:
http://en.wikipedia.org/wiki/Wikipedia:5000
Thanks, -AW
On 12/28/2012 07:28 PM, John Vandenberg wrote:
There is a steady stream of blogs and 'news'
about these lists
https://encrypted.google.com/search?client=ubuntu&channel=fs&q=%22S…
How does a researcher go about obtaining access logs with useragents
in order to answer some of these questions?
--
Andrew G. West, Doctoral Candidate
Dept. of Computer and Information Science
University of Pennsylvania, Philadelphia PA
Website:
http://www.andrew-g-west.com