Bah - all covered above. I should not answer emails early in the morning. But: if you still want that data, I am happy to see if I can grab it ;).


On 11 June 2014 08:04, Oliver Keyes <okeyes@wikimedia.org> wrote:
I thiiink (don't quote me on this) the NT schema contains wiki data. I'll see if I can stitch together an aggregate dataset for you, if you think it'd help. Per-day granular enough?


On 11 June 2014 06:54, Federico Leva (Nemo) <nemowiki@gmail.com> wrote:
Federico Leva (Nemo), 01/06/2014 14:11:

Currently I'm in need of per-wiki and/or
diachronic/over-time plots for it.wiki:
<https://meta.wikimedia.org/wiki/Research:The_sudden_decline_of_Italian_Wikipedia>


If nobody has ideas I guess I'll just go for the 800 lb gorilla approach
and submit a patch to have performance graphs for the top 10 wikis as we
currently do for edits (cf. https://bugzilla.wikimedia.org/56039 )? In
the meanwhile: https://gerrit.wikimedia.org/r/#/c/136631/

Those two are closed but I can't just steal the editswiki approach because it's based on the series' values; absent better ideas I'll need to hardcode the wikis' names.

Not for that goal, but mostDeviant can be of use too: <https://graphite.readthedocs.org/en/latest/functions.html#graphite.render.functions.mostDeviant>
For instance, a graph of sites/kinds of assets whose payload varied the most across last week. http://ur1.ca/hi4oa Can something like this be of use anywhere?


Nemo

_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics



--
Oliver Keyes
Research Analyst
Wikimedia Foundation



--
Oliver Keyes
Research Analyst
Wikimedia Foundation