Page MenuHomePhabricator

Annotate relevance of search results for sample queries
Closed, ResolvedPublic

Description

Set up a pipeline/framework for human annotators to label relevance of different search results.

The most promising approach is to use prolific's AI task builder tool:

  • This allows to easily define an annotation task with a simple interface. The only requirement is a csv-file containing the data to be annotated.
  • It also allows to recruit participants for annotating each sample a given number of times.

Different available tools from similar previous experiments; however, they cant just be used off-the-shelve and likely require some work to adapt to the current use-case

Event Timeline

  • Created a code pipeline to prepare the data for annotation using Prolific's AI task builder tool.
  • Created a mockup study on Prolific to test the functionality of the annotation UI.
  • Agreed to proceed with the Prolific framework with minor adjustments:
    • Try merging page-title/section into paragraph content -> Rename paragraph content into “result”. (to avoid scrolling).
    • Improve the paragraph splitting logic to incorporate corner cases.
    • Estimate the time of study.
  • Created the task description to use in the study.
  • Estimated the study duration at approximately 30 seconds per sample, resulting in about 6 minutes per annotator, accounting for 10 samples per annotation and additional time for instruction review.
  • Updated the data preparation pipeline for the study by merging the page title, section name, and paragraph text into a single field, improving the annotator user interface.

Final update:

We will post-process the raw data to make the dataset more usable for downstream tasks. But closing this task as we completed the goal.