In addition to tracking the detailed analytics needed for evaluating the A/B test (T405177), we also want engineering focused performance metrics to be able to assert that our feature works as expected on a technical level.
This explicitly includes tracking the runtime needed to figure out for which paragraph to show the Revise Tone Check. (See the internal slack conversation for context.)
Acceptance Criteria:
Metrics for at least the following are being recorded:
- time to get data from Data Gateway
- time to get find correct paragraph in article
- number of users declining a task with which reason in the survey
- number of errors
- suggestion no longer available (no data in data gateway)
- paragraph ambiguous (distance of 2nd-highest scoring paragraph is 90% of the distance of the 2nd-highest scoring paragraph)
In addition to the above, we should make sure that the ReviseTone task is visible on the KPI dashboard next to the other Newcomer Tasks
- ReviseTone will be visible on the KPI dashboard
Notes:
- visualization of these metrics on Grafana is also part of this task: https://grafana.wikimedia.org/d/0352e22b-1ebf-4e13-a97d-4f5f05eb8463/revise-tone-structured-task
- the metrics should be stratified by wiki, i.e. have a label with the dbname
- performance metrics for the onboarding will be done as part of T406252: 🧑💻 Instrument the Revise Tone Onboarding Quiz