I thiiink (don't quote me on this) the NT schema contains wiki data. I'll
see if I can stitch together an aggregate dataset for you, if you think
it'd help. Per-day granular enough?
On 11 June 2014 06:54, Federico Leva (Nemo) <nemowiki(a)gmail.com> wrote:
Federico Leva (Nemo), 01/06/2014 14:11:
Currently I'm in need of per-wiki and/or
diachronic/over-time plots for it.wiki:
<https://meta.wikimedia.org/wiki/Research:The_sudden_
decline_of_Italian_Wikipedia>
If nobody has ideas I guess I'll just go for the 800 lb gorilla approach
and submit a patch to have performance graphs for the top 10 wikis as we
currently do for edits (cf.
https://bugzilla.wikimedia.org/56039 )? In
the meanwhile:
https://gerrit.wikimedia.org/r/#/c/136631/
Those two are closed but I can't just steal the editswiki approach because
it's based on the series' values; absent better ideas I'll need to hardcode
the wikis' names.
Not for that goal, but mostDeviant can be of use too: <
https://graphite.readthedocs.org/en/latest/functions.html#
graphite.render.functions.mostDeviant>
For instance, a graph of sites/kinds of assets whose payload varied the
most across last week.
http://ur1.ca/hi4oa Can something like this be of
use anywhere?
Nemo
_______________________________________________
Analytics mailing list
Analytics(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics
--
Oliver Keyes
Research Analyst
Wikimedia Foundation