Background
This fiscal year, the iOS team is seeking to increase unreverted mobile contributions. One of our hypotheses to achieve this growth is by introducing Suggested Edits to the iOS app. We would like to explore a Suggested Edits task that encourages users to add helpful and accurate alt-text to the 95% of images across Wikipedias that are missing them. Before investing too much into this idea, we will start with a proof of concept and get feedback at the GLAM conference in Uruguay and with LATAM and Caribbean users. If the task proves successful, we will build out the full feature and scale it, if it is not successful we will pivot to Image Recommendations or Article Descriptions.
Decisions Matrix
- If less than 45% of edits are scored as a 3 or higher then we will pivot to a different suggested edit. If 46%-70% of edits are scored a 3 or higher we will improve guidance or use AI to better assist users. If 71% or higher of edits are scored a 3 or higher we will scale feature.
- If average edit per unique user is under 3 we will pivot to a different suggested edit. If it is 3-6, we will consider interventions to reduce friction. If it is 7 or higher we can scale
- If user edit through feature a second day we should proceed with improvements and scaling
- If less than 55% or less of users are satisfied with feature we will not scale without making changes
- If we do not have at least 50 people try the feature we will do direct outreach to gain more edits
- If more there are more than 30% of users find the task too difficult we will create an intervention to reduce difficulty before scaling. If 80% or more users find the task too difficult we will consider abandoning depending on supplementary responses
- If the skip rate is 20% higher than image captions on Android we will consider pivoting to a different suggested edit unless evidence points to an intervention that could reduce this rate
Curiosity
- How does the predicted revert rate compare to Android's Image Captions Suggested Edit?
- Is there a difference user reported feedback based on their familiarity with editing and writing alt-text
- Is there a difference in metrics by language?
Must haves for proof of concept
- Entry point in Settings
- Prominent guidance for writing good alt-text
- Users ability to get context about the image from the article
- Users ability to access relevant metadata (can take user to Web)
- Detection of which images do not have alt-text
- Ability to store the responses we get to evaluate if they are good alt-text
- Users should be able to give feedback about the feature over all
Nice to have
- Positive Reinforcement
- Exposing to users how many tasks or edits they've completed
- Ability to publish alt-text
- Limits on number of task can be completed in a given day or session if it is being published to Wikipedia
- Entry point via Explore Feed
- Do not allow users to copy and paste in image caption
- Users should be prompted provide feedback about the feature
- Suggest Alt-text (think Machine Assisted Article Descriptions)
- Can playback what was written in preview
User stories
Primarily
- As a participant at the GLAM Conference in Uruguay, I want to test out an alt-text Suggested Edits Task, to get a concrete idea of the concept and provide meaningful feedback
- As the Director of Product, I want concrete proof an alt-text Suggested Edits would increase edits on iOS, without creating a burden for patrollers
Secondarily
- As a user with limited data, I want to read alt-text, so even if images are not loaded, I am aware of what is in the image
- As a user navigating articles with a screen reader, I want alt-text to be available , so that I have the same additional context about an article as users that are not using screen readers
Primary target audiences
- Round 1: GLAM Conference Attendees
- Round 2: Spanish and Portuguese speakers in the Americas and Caribbean
- Round 3: French and English Speakers in the Americas and Caribbean