Page MenuHomePhabricator

[medium] Understand current usage of edit summary statements on Wikipedia
Open, Needs TriagePublic

Description

Overview

When seeking to understand what changes were made by a given Wikipedia edit, there are three general sources of information: the actual diff of an edit (example), the edit summary, and edit tags. Edit tags are generally pretty simple and constrained to tool information or basic regexes -- e.g., whether an edit occurred via mobile or easy-to-define content changes like blanked a section or a new, short article). The edit diff is much richer -- showing the specific changes made in context -- but takes up too much visual space for tasks such as quick browsing of an edit history or patrolling for suspicious edits. The edit summary is intended to be a good middle ground -- a succinct but flexible, editor-provided description of what the edit did (and maybe why).

Edit summaries have a number of drawbacks but also are very valuable. Vandals may provide misleading summaries and many editors leave the summary blank or use a canned edit summary that may or may not accurately reflect the impact of the edit. Despite these drawbacks, edit summaries are invaluable for tasks such as understanding work on Wikipedia, building datasets of edits for study, and contextualizing edits.

Task

This research will focus on understanding current usage of edit summaries in order to identify potential opportunities for tool-building to better support researchers and patrollers. It will primarily be qualitative research. A potential set of steps include:

  • Read documentation on edit summaries and go through the edit history for some random articles to get a sense of how they are used.
  • Gather initial sample of last three edits to 50 random articles (sticking to just namespace 0 for now -- i.e. the standard Wikipedia article)
  • Code your sample for types of edit summaries and how well the edit summary described the edit. This should be multi-faceted -- e.g., was the summary misleading? was it complete? did it say what and why? was it manual, canned, or automated? Ideas for what sorts of questions to be asking for will initially come from the documentation -- i.e. evaluating how well edit summaries meet what they are recommended to contain. You likely will want to do this iteratively and revise what codes you apply as you see more summaries and notice new patterns too.
    • NOTE: this can be for English Wikipedia and/or any other languages the researcher can read fluently.
    • NOTE: this is combining both the conventional and directed approaches described in Hsieh and Shannon or inductive and deductive by Muller.
  • Expand sample and apply codes to new edit summaries (and identify any new codes that should be added).
    • NOTE: this can be more edits, more languages, or perhaps a stratified sample that looks at specific types of articles such as more controversial or heavily-edited articles.
    • NOTE: this is similar to the summative approach described in in Hsieh and Shannon.
  • Write up taxonomy of edit summaries, common issues with edit summaries, and how common they are based on qualitative coding.
  • For each issue, propose what approaches / tools might help to address the issue.

This task is considered [medium]. In general, it's expected that the task will take a a month or two of consistent work and is a good fit for someone with some research experience or interest in being involved in research. The actual time needed, however, will depend greatly on your level of experience.

Rationale

In order to make edit summaries more useful, they likely need to be more complete and trustworthy. This research is a first step to identifying how large of a challenge it would be to augment edit summaries -- either via editor support or addition of edit tags. Depending on the outcomes, it may lead to creating datasets for further research, new tooling, or design recommendations.

Recommended Skills

  • This task primarily requires some experience with qualitative methods and, in particular, qualitative coding.
  • It might be possible to automate aspects of the sampling -- e.g., using the Random API and Python -- but that is not necessary.

Acceptance Criteria

  • The output of this task will be a Meta report describing the research and findings (example). Depending on researcher and mentor interest, this could be expanded into a more formal publication.
  • At each stage of the analysis, the choices should be carefully grounded in past research and validated by the mentor.
  • Depending on the approach, PAWS-hosted Jupyter notebooks might also be a viable option for sharing progress.

Process

  • If you are interested in this task and it is not assigned to anyone, you may begin work on it. Please leave a comment on the task and tag @Isaac so that he is aware.
  • If you have made some progress on the task (at least 10-20 edit summaries and your initial codes) and would like to continue, share a link to your current draft and let @Isaac know so that he can assign the task to you and help you to plot out the next steps.
  • Generally, @Isaac will be able to answer any questions about the task and try to respond quickly when clarification is necessary but response times may be slow if help is needed for more general debugging etc.

Resources

Event Timeline

Hi @Isaac, confirming that I'm taking this up :)

Hi @Isaac, confirming that I'm taking this up :)

Great! An older piece of research was just brought to my attention that used edit summaries (they call them user comments) that might be worth reading in that they were trying to bring some structure and understanding to edit summaries through visualization techniques. It's also just a classic Wikipedia research paper and is a lot of fun to read regardless of its connection to this task: https://link.springer.com/content/pdf/10.1007/978-3-540-74800-7_23.pdf

Hi @Isaac,
We’ve done some initial planning for the project and set a timeline for the project that works for us. We’d like to try and wrap up everything in the scope of the task in 2 months, so roughly by the first week of October.

Our motivations:

  • Primarily we are interested in using this research to actually improve edit summaries and develop solutions to the problems mentioned above such as vandalism detection. This aligns with the possible outcomes mentioned in the task.
  • Another goal we’re interested in is trying to turn this project into a conference paper. For now we’re thinking of submitting it to CSCW 2022 but would be open to suggestions of different venues.

The immediate next steps we’ve planned are meant to be for the next two weeks:

  • Literature review: We would be covering readings about the method(including your recommendations above) and the domain. Within the domain, we plan to read about topics similar to edit summaries like open source code contributions and commit messages and also read about related wikipedia research starting with the papers you’ve linked above. Here, it would be great to have your assistance through suggested readings and guidance on what all to cover in a proper literature review for this project.
  • Data Collection: As described by the first step in the task we would be collecting edit summaries, diff pages and talk pages that we want to go through for our analysis. We plan to open code them subsequently.
  • Another question we had was whether talking to experienced Wikipedia editors would help us uncover different motivations, mental models and thoughts behind writing edit summaries. We could also identify the shortcomings of using edit summaries as an experience. Would this help us understand the context better? What are your thoughts?

We are still planning the next steps after these and would update you on those as soon as possible. Please offer any suggestions or feedback on our approach or direction.

@Tarun_Mugunthan thanks for providing these details! In particular, very useful to hear what your motivations are for taking on this project. While we would want to approach this robustly regardless of intent to publish, knowing that a venue like CSCW is something of interest to you can help with making choices about approaches, sample size, etc. In particular, after you get through the initial exploration stages, you'll want to think about the paper you want to write and what bigger themes you want to discuss -- i.e. what takes this project from just an understanding of edit summaries that is useful for Wikipedians and tool builders to an understanding of edit summaries that also informs larger questions about signaling/norms on online platforms (in which case maybe we focus on the vandalism side of things) or machine learning (in which case maybe we focus on how edit summaries are used to generate datasets of edits) etc.

Literature review: We would be covering readings about the method(including your recommendations above) and the domain. Within the domain, we plan to read about topics similar to edit summaries like open source code contributions and commit messages and also read about related wikipedia research starting with the papers you’ve linked above. Here, it would be great to have your assistance through suggested readings and guidance on what all to cover in a proper literature review for this project.

This sounds like a great idea. I'll try to do some brainstorming over the next week and add any more papers that I think of. Two papers come to mind right now:

As an aside, the Wikimedia Foundation maintains an etherpad instance, which is essentially very lightweight Google docs. Caveats being the doc is public and there is no commenting ability. If that would be useful, feel free to create a doc for e.g., tracking the lit review: https://etherpad.wikimedia.org

Data Collection: As described by the first step in the task we would be collecting edit summaries, diff pages and talk pages that we want to go through for our analysis. We plan to open code them subsequently.

Sounds good -- I'd say focus on the edit summaries right now to keep this manageable but diffs and talk pages can certainly help elucidate what editors were seeking to do.

Another question we had was whether talking to experienced Wikipedia editors would help us uncover different motivations, mental models and thoughts behind writing edit summaries. We could also identify the shortcomings of using edit summaries as an experience. Would this help us understand the context better? What are your thoughts?

It's a good thought. My feeling is that before we think about taking this step, you should see what editors have already been saying about edit summaries. This could be searching Village Pump for discussions, looking at talk pages and their archives for edit summary documentation pages, or presumably other public spaces where they might have been discussed.