Set up a pipeline/framework for human annotators to label relevance of different search results.
The most promising approach is to use prolific's AI task builder tool:
- This allows to easily define an annotation task with a simple interface. The only requirement is a csv-file containing the data to be annotated.
- It also allows to recruit participants for annotating each sample a given number of times.
Different available tools from similar previous experiments; however, they cant just be used off-the-shelve and likely require some work to adapt to the current use-case
- Discernatron
- Media search
- Article level image suggestion
- Annotool
- https://annotool.toolforge.org/projects/13
- https://toolsadmin.wikimedia.org/tools/id/annotool
- https://gitlab.wikimedia.org/mnz/annotool
- currently only supports annotation of diffs