I filed epics for each of these:
We'll be breaking these epics down into specific actionables the sprint planning meeting in 25 minutes.

Thanks,
Dan

On 8 November 2015 at 21:51, Dan Garry <dgarry@wikimedia.org> wrote:
Summarising this discussion, it seems like the path forward which would reap the most rewards is as follows:
  1. Finish the MVP of the relevance lab; right now we can only test zero results rate for any given experiment, and the lab will help us also test result relevance.
  2. Start writing tests to switch out the language detector used in the first test with alternative ones, to see if they're better
    • This should affect the zero results rate, so lack of the relevance lab does not block this
    • This should also affect relevance (at least conceptually), so can be tested using the relevance lab also
  3. Write test to use accept-language header as a heuristic to do language switching (rather than language detection)
    • This should affect the zero results rate, so lack of the relevance lab does not block this
    • This should also affect relevance (at least conceptually), so can be tested using the relevance lab also
  4. Expand original language switching test to also switch if there are "few" results (let's say "few" = 3 or fewer).
    • Does not really affect zero results rate; this is dependent on relevance lab
Any objections to this course of action? I plan to file tasks for these mid-Monday morning.

Thanks,
Dan

On 2 November 2015 at 16:58, Erik Bernhardson <ebernhardson@wikimedia.org> wrote:
Now that we have the feature deployed (behind a feature flag), and have an initial "does it do anything?" test going out today, along with an upcoming integration with our satisfaction metrics, we need to come up with how will will try to further move the needle forward.

For reference these are our Q2 goals:
  • Run A/B test for a feature that:
    • Uses a library to detect the language of a user's search query.
    • Adjusts results to match that language.
  • Determine from A/B test results whether this feature is fit to push to production, with the aim to:
    • Improve search user satisfaction by 10% (from 15% to 16.5%).
    • Reduce zero results rate for non-automata search queries by 10%.
We brainstormed a number of possibilities here:
https://etherpad.wikimedia.org/p/LanguageSupportBrainstorming

We now need to decide which of these ideas we should prioritize. We might want to take into consideration which of these can be pre-tested with our relevancy lab work, such that we can prefer to work on things we think will move the needle the most. I'm really not sure which of these to push forward on, so let us know which you think can have the most impact, or where the expected impact could be measured with relevancy lab with minimal work.



_______________________________________________
discovery mailing list
discovery@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/discovery




--
Dan Garry
Lead Product Manager, Discovery
Wikimedia Foundation



--
Dan Garry
Lead Product Manager, Discovery
Wikimedia Foundation