To evaluate the impact of Reference Check, we evaluated the proportion of new content edits that also included a new reference.
To calculate this proportion (T342930), we leveraged the editcheck-newcontent and editcheck-newreference change tags.
Similarly, to evaluate the impact of Tone Check (T387918), we need to analyze the proportion of new content edits that include problematic language and we will use this ticket to organize the work of:
- Leveraging the Tone Check model to evaluate the presence/absence of problem language within new content edits
- Appending edit tags to edits when the Tone Check model detects non-neutral language within new content edits
Requirements
- Do not block the saving of an edit on the model returning a result
- Each time an edit is saved without the model returning a result beforehand, log this so that Megan can filter these edits out from edits where the model returned an evaluation "no problematic language present"
- Please share proposed implementation with @MNeisler for review before finalizing.
- Name new tag editcheck-tone
- Implement editcheck-tone as a hidden tag
- Only edits in languages we've evaluated the model with (English, Spanish, Japanese, Portuguese, French) by editors which would meet the requirements for Tone check is evaluated and tagged (e.g. editors with <100 edits, or whatever is set in config)
- Make sure this tag only runs on wikis that have Edit Check enabled
Open questions
- 1. To what extent – if any – is it feasible for the model to evaluate presence/absence of problematic language at save-time so that a positive evaluation can be reflected as an edit tag?
- Yes; see T388716#10783780 for more.
- 2. To what extent – if any – is evaluation happening at save-time required? Asked another way: can edit tags be appended after-the-fact?
- Per discussion with Editing Engineering, tags need to be applied at save-time
- 3. Would we consider the intervention successful if the model detected problematic language within published edits that were NOT reverted?
- Relevant for T383279
- 4. To extent would we expect a person's connection speed impact how quickly the model is able to return a result?
- Context: were a person's connection speed to be highly correlated with the model returning a result, we would risk biasing the sample.
Done
- Editing QA confirms tag is being applied as expected
- @MNeisler to verify model timeouts are being logged in VEFU


