(cross-posting to analytics list)
On 09/29/2012 12:13 PM, Lori Phillips wrote:
[snip]
> *Action step*: I think we should create a single page, be it on meta,
> outreach, or mediawiki, where we compile all of the useful links, needs,
> and suggestions for the Analytics Team. They can then take from that what
> they can. But centralizing the information is key. This could include the
> above Google Doc I mention, links to our current guides and partnerships
> lists on Commons, info on the Europeana Toolset Project, and additional
> suggestions we may have. I'd be appreciative to hear if someone, who is
> more technically inclined than myself (and that doesn't take much) is
> willing to take this on.
It looks like people started at
https://www.mediawiki.org/wiki/Analytics/Pageviews/GLAM , right? "The
purpose of this page is to gather all the different documents,
discussions and ideas related to GLAM Analytics and to define a list
with the most important metrics. Please feel free to add any missing
data, this is document is almost certain to be incomplete."
> Hope that helps, and thanks for your patience as I attempt to faciliate
> this topic which is far out of my comfort zone.
> Lori
Thanks for working on this, Lori!
--
Sumana Harihareswara
Engineering Community Manager
Wikimedia Foundation
Cross-posting.
-------- Original Message --------
Subject: [Wikitech-l] MediaWiki community metrics
Date: Fri, 28 Sep 2012 13:10:12 -0700
From: Quim Gil <quimgil(a)gmail.com>
Reply-To: Wikimedia developers <wikitech-l(a)lists.wikimedia.org>
To: Wikimedia developers <wikitech-l(a)lists.wikimedia.org>
Hi there,
Asking for a task to volunteer, Sumana encouraged me to look at the
topic of community metrics. She pointed to
https://wikitech.wikimedia.org/view/Pentaho as a starting point.
After a first look at Pentaho and what some colleagues at the MeeGo
project did with it [1], I searched (a bit) for any wiki pages or
discussions about community metrics here, but couldn't find any.
http://www.mediawiki.org/wiki/User:Qgil/MediaWiki_Community_Metrics
Edits welcome!
I'm looking for feedback, help, and a first prototype of an
automatically refreshed report hopefully sooner than later. Something
simple to build upon.
Even if it's too tempting to define the first prototype thinking first
on tools or data available, you are encouraged to start by proposing
what questions do you want actually answered. What community trends do
you want to know?
See
http://www.mediawiki.org/wiki/User:Qgil/MediaWiki_Community_Metrics#Trends_…
- in few days we should have agreed on the first and most important
trends we want to visualize.
[1] http://wiki.meego.com/Metrics/Dashboard
--
Quim
_______________________________________________
Wikitech-l mailing list
Wikitech-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l
FYI. If you'd like an invite, shoot my an e-mail!
Forwarded message:
> From: Ori Livneh <ori(a)wikimedia.org>
> Reply To: E3 team discussion list <e3-team(a)lists.wikimedia.org>
> To: Internal E3 team discussion list <e3-team(a)lists.wikimedia.org>
> Date: Tuesday, September 25, 2012 2:25:29 PM
> Subject: [E3-team] Intel: analytics platforms
>
> Hey guys,
>
> I thought we should take a look at how the major analytics platforms do what they do: how they implement experiments, how they collect data, and how they present it for analysis. We can't use them on Wikipedia, so I created a simple web "app" that guides the user along a mock conversion funnel:
>
> http://cube.256.io/bi/
>
> That page is logging events to Google Analytics, MixPanel, KISSMetrics and Optimizely.
>
> You should have received invitations for Google Analytics already. I'll follow up with the login details for the other services.
>
> Some of these services are smart enough to filter you out of the data stream if you log in to their service, so if you can, go through a part (or all) of the funnel on random computers, so we have some data.
>
> If this provokes ideas, or if you discover interesting things, please report back.
>
>
> Ori
>
> --
> Ori Livneh
> ori(a)wikimedia.org (mailto:ori@wikimedia.org)
>
>
>
> _______________________________________________
> E3-team mailing list
> E3-team(a)lists.wikimedia.org (mailto:E3-team@lists.wikimedia.org)
> https://lists.wikimedia.org/mailman/listinfo/e3-team
>
Hello analytics,
I've set up an article edit / insert feed at http://kubo.wmflabs.org/editstream.html. It's receiving updates using a node.js WebSockets server running on stat1. I'd like to productionize it and wanted to solicit your input on how to do it right. I think it'd be useful to provide this stream as a service to the community.
Each edit event is ~300 bytes of gzipped-compressed JSON data. With ~140,000 edits a day, the bandwidth per client is 0.5kbps. No filtering or buffering happens on the server, so I think it'll scale quite well. Should I simply submit a puppet patch to configure this service to run on stat1001?
It'd be good to map the service onto a URL on bits, so that it's easily accessible from JS code running on Wikipedia.
Thoughts? Let me know!
Thanks,
Ori
--
Ori Livneh
ori(a)wikimedia.org
fyi, I know very little about IRC in general. Is this already happening?
Is there a reason we don't want to? Should it be public? I was hoping to
install irssi <http://irssi.org/> (andrew?) on stat1 and have it make logs
that we could all check when we've joined the channel too late to get the
hangout link, or need to look for link that someone posted a week ago. I'm
imagining just grep it over an ssh connection for now, though if we wanted
to share it publicly we could create a labs instance and run one of those
web interfaces that is built around irssi logs
(irclog<http://irclog.readthedocs.org/en/latest/index.html>,
webssi <http://wouter.coekaerts.be/webssi/>--suggestions welcome as none of
these look super canonical)
thanks,
evan
Yo 'lytics,
Where do I apply for access to the MySQL instance on stat1?
I affirm that I am not a past or present member of the Communist Party.
--
Ori Livneh
ori(a)wikimedia.org
Hi everybody,
The Analytics Team is happy to announce the first version of gerrit-stats.
Gerrit-stats keeps track of the backlog of codereview for Git individual
repositories.
Gerrit-stats dashboard is available at http://gerrit-stats.wmflabs.org
Currently, it has a few example charts but we can add your repo to the
dashboard as well, just let us know!
To create a new chart yourself visit
http://gerrit-stats.wmflabs.org/graphs/new
This will launch the interface to create your own graph. Click on 'Data'
and Click on 'Add Metric' and a pull down menu with all the repositories
will appear. Select the repository of your interest and select the metric
that you want to visualize. Once you have selected all the metrics of your
interest go back to 'Info' and enter a slug name. Then press 'Enter' and
then click the 'Save' button.
Currently, the following metrics are tracked (on a daily basis):
1) Number of new changesets
2) Number of changesets without any codereview per day (this excludes
automated review from lint and lint-like reviewers).
3) Number of changesets waiting for merge per day (only applies to
changesets that received only positive reviews)
4) Number of changesets self reviewed.
And for metrics 2 and 3, there is a version for volunteers and for WMF
staff.
Gerrit-stats is visualized using Limn, Limn is the data GUI developed by
the Analytics Team and lead by David Schoonover. Limn is available on
https://github.com/wikimedia/limn
This is the initial release and I am sure there will be bugs and issues. If
you have any questions, or problems using gerrit-stats then either:
1) Head over to #wikimedia-analytics on IRC and ask us
2) Send an email to the analytics mailinglist
3) Contact us directly.
Not Yet Frequently Asked Questions:
1) How do I create a visualization of the code review metrics for a repo?
Visit gerrit-stats.wmflabs.org/graphs/new
This will launch the interface to create your own graph. Click on 'Data'
and Click on 'Add Metric' and a pull down menu with all the repositories
will appear. Select the repository of your interest and select the metric
that you want to visualize. Once you have selected all the metrics of your
interest press the 'Save' button.
Your are all set and you can use this permalink for future reference.
2) How do I edit an existing chart?
Simply append /edit to the URL of your chart and you can edit it.
3) My repository is not showing up in the pull down menu, what happened?
By default, all repositories are automagically kept track of as soon they
contain a single commit. There are two exceptions:
1) If your repository name contains the string 'test' or 'private', it will
be ignored.
2) The orgchart repository is not tracked by gerrit-stats, this is a known
issue but Chad and I haven't been able to figure out what causes this.
If your repository is missing then please contact me.
4) Will you add metrics for individual committers?
Right now, the unit of analysis is a repository but it is definitely
possible to keep track of codereview metrics for individuals. However, I
would like to hear some use-cases first before embarking on this.
5) The chart looks to spikey, how can I have smoother lines?
1) Go to http://gerrit-stats.wmflabs.org/graphs/name_of_chart/edit
2) Click on 'Options'
3) Click on 'Advanced' (right side of screen)
4) Click on 'rollPeriod' (bottom of screen, yellow box)
5) This allows to create a moving average, so you can replace the 1 with 7
meaning that each datapoint is the average of the past 7 days. This option
applies to both metrics but it really smooths out the outliers.
6) I want a new metric. How do I go about it?
There are two options:
a) Clone the gerrit-stats repo and hack away,it's Python btw. We are happy
to help out!
b) Send us a suggestion for a new metric, the more precise the more useful!
On behalf of the Analytics Team,
Diederik