Note that to avoid too much traffic here, I've responded to MZMcBride
privately with my code. I'd be happy to share my code with others, and
include others in its discussion -- just contact me/us privately.
Thanks, -AW
On 08/12/2011 10:30 AM, MZMcBride wrote:
Andrew G. West wrote:
I've actually been parsing the raw data from
[
http://dammit.lt/wikistats/] daily into a MySQL database for over a
year now. I also store statistics at hour-granularity, whereas
[stats.grok.se] stores them at day granularity, it seems.
I only do this for en.wiki, and its certainly not efficient enough to
open up for public use. However, I'd be willing to chat and share code
with any interested developer. The strategy and schema are a bit
awkward, but it works, and requires on average ~2 hours processing to
store 24 hours worth of statistics.
I'd certainly be interested in seeing the code and database schema you've
written, if only as a point of reference and to learn from any
bugs/issues/etc. that you've encountered along the way. Is it possible for
you to post the code you're using somewhere?
MZMcBride
_______________________________________________
Wikitech-l mailing list
Wikitech-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l
--
Andrew G. West, Doctoral Student
Dept. of Computer and Information Science
University of Pennsylvania, Philadelphia PA
Email: westand(a)cis.upenn.edu
Website:
http://www.cis.upenn.edu/~westand