I filed epics for each of these:
- T118278 - EPIC: Run additional A/B tests for the language switching functionality, with different libraries to detect the query's language, and evaluate if the other libraries are better
- T118280 - EPIC: Run A/B test to determine whether using the accept-language header of the user to switch query languages is good or not
- T118281 - EPIC: Run original language switching A/B test, but switch languages if the user has fewer than n results (for some n), and determine if that's better or worse
We'll be breaking these epics down into specific actionables the sprint planning meeting in 25 minutes.Thanks,DanOn 8 November 2015 at 21:51, Dan Garry <dgarry@wikimedia.org> wrote:Summarising this discussion, it seems like the path forward which would reap the most rewards is as follows:
- Finish the MVP of the relevance lab; right now we can only test zero results rate for any given experiment, and the lab will help us also test result relevance.
- Start writing tests to switch out the language detector used in the first test with alternative ones, to see if they're better
- This should affect the zero results rate, so lack of the relevance lab does not block this
- This should also affect relevance (at least conceptually), so can be tested using the relevance lab also
- Write test to use accept-language header as a heuristic to do language switching (rather than language detection)
- This should affect the zero results rate, so lack of the relevance lab does not block this
- This should also affect relevance (at least conceptually), so can be tested using the relevance lab also
- Expand original language switching test to also switch if there are "few" results (let's say "few" = 3 or fewer).
- Does not really affect zero results rate; this is dependent on relevance lab
Any objections to this course of action? I plan to file tasks for these mid-Monday morning.Thanks,DanOn 2 November 2015 at 16:58, Erik Bernhardson <ebernhardson@wikimedia.org> wrote:_______________________________________________Now that we have the feature deployed (behind a feature flag), and have an initial "does it do anything?" test going out today, along with an upcoming integration with our satisfaction metrics, we need to come up with how will will try to further move the needle forward.For reference these are our Q2 goals:We brainstormed a number of possibilities here:
- Run A/B test for a feature that:
- Uses a library to detect the language of a user's search query.
- Adjusts results to match that language.
- Determine from A/B test results whether this feature is fit to push to production, with the aim to:
- Improve search user satisfaction by 10% (from 15% to 16.5%).
- Reduce zero results rate for non-automata search queries by 10%.
https://etherpad.wikimedia.org/p/LanguageSupportBrainstorming
We now need to decide which of these ideas we should prioritize. We might want to take into consideration which of these can be pre-tested with our relevancy lab work, such that we can prefer to work on things we think will move the needle the most. I'm really not sure which of these to push forward on, so let us know which you think can have the most impact, or where the expected impact could be measured with relevancy lab with minimal work.
discovery mailing list
discovery@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/discovery
--Dan GarryLead Product Manager, DiscoveryWikimedia Foundation--Dan GarryLead Product Manager, DiscoveryWikimedia Foundation