Last time we trained the hewiki model on an analysis chain that was different than the one used when running the test, so the results were invalid. We re-run the test with a new model trained against the new analysis chain. We use the same sampling rate as last time.
This test ran from 02 January 2018 to 16 January 2018 on hewiki. There were 3 test groups: control, ltr-1024, ltr-1024-i. This report includes fulltext searches. Refer to Phabricator ticket T182616 for more details.
Fulltext search events: Deleted 232 duplicated events. Deleted 98498 unnecessary check-in events and only keep the last one. Deleted 5 events with negative load time. Removed 465 orphan (SERP-less) events. Removed 0 sessions falling into multiple test groups. Removed 2 sessions with more than 50 searches.
Select one of these three tabs:
Days | Events | Sessions | Page IDs | SERPs | Unique search queries | Searches | Same-wiki clicks | Other clicks |
---|---|---|---|---|---|---|---|---|
15 | 115,525 | 34,575 | 86,314 | 70,353 | 68,890 | 56,432 | 16,739 | 0 |
Select one of these sub-tabs:
Event type identifies the context in which the event was created. Every time a new search is performed a searchResultPage event is created. When the user clicks a link in the results a visitPage event is created. When the user has dwelled for N seconds a checkin event occurs. If the user clicks an interwiki result provided by TextCat language detection, there is a iwclick event. If the user clicks on a sister search result from the sidebar, that’s an ssclick. If the user interacts with a result to explore similar (pages, categories, translations), there are hover-on, hover-off, and esclick events.
The goal here is to see whether the proportions of operating system (OS) and browser usage are similar between the groups. To aid decision making, Bayes factor is computed for each row. If one group has very different OS/browser share breakdown (Bayes factor > 2), there might be something wrong with the implementation that caused or is causing the sampling to bias in favor of some OSes/browsers. Note that for brevity, we show only the top 10 OSes/browsers, and that we don’t actually expect the numbers to be different so this is included purely as a diagnostic.
Operating systems:
Browsers:
Select one of these sub-tabs:
ltr-1024 vs. control
PaulScore is a measure of search results’ relevancy which takes into account the position of the clicked results, and is computed via the following steps:
We can calculate the confidence interval of PaulScore\((F)\) by approximating its distribution via boostrapping.
Users may click back to the search result page directly after they clickthrough to an article (within 10 mins). We computed two kinds of return rate:
Among users with at least a click in their search, the proportion of searches that return to the same search page
Among users with at least a click in their search session, the proportion of sessions that return to search for different things (different search result page but in the same session)
We use a technique called interleaving to evaluate the user-perceived relevance of search results from the experimental configuration. In it, each user is their own baseline – we perform two searches behind the scenes and then interleave them together into a single set of results using the team draft algorithm described by Chapelle et al. (2012). For all the graphs in this section, A refers to control group or the first group in the interleaved group names, B referes to the name in the interleaved group names or the second group in the interleaved group names. For example, if the interleaved group name is ‘ltr-i-1024’, A is the control group and B is group ‘1024’; if the interleaved group name is ‘ltr-i-20-1024’, A is group ‘20’ and B is group ‘1024’.