Skip to content
Longterm Wiki
Back

ArXiv preprint on Community Notes helpfulness rates and timing.

paper

Authors

Zahra Arjmandi-Lari·Alexios Mantzarlis·Tom Stafford

Credibility Rating

3/5
Good(3)

Good quality. Reputable source with community review or editorial standards, but less rigorous than peer-reviewed venues.

Rating inherited from publication venue: arXiv

Relevant to AI safety discussions around scalable content moderation and human oversight mechanisms; Community Notes represents a real-world crowdsourced alignment/moderation system whose sustainability challenges inform debates about human-in-the-loop governance at scale.

Paper Details

Citations
4
0 influential
Year
2025

Metadata

Importance: 42/100arxiv preprintprimary source

Abstract

Community Notes are emerging as an important option for content moderation. The Community Notes system pioneered by Twitter, now known as X, uses a bridging algorithm to identify user-generated context with upvotes across political divides, supposedly spinning consensual gold from partisan straw. It is important to understand the nature of the community behind Community Notes, especially as the feature has now been imitated by several billion-user platforms. We look for signs of stability and disruption in the X Community Notes community and interrogate the motivations other than partisan animus (Allen, Martel, and Rand 2022) which may be driving users to contribute. We conduct a novel analysis of the impact of having a note published, which requires being considered "helpful" by the bridging algorithm, utilising a regression discontinuity design. This allows stronger causal inference than conventional methods used with observational data. Our analysis shows the positive effect on future note authoring of having a note published. This highlights the risk of the current system, where the proportion of notes considered "helpful" (and therefore shown to users on X) is low, 10%, and declining. This analysis has implications for the future of Community Notes on X and the extension of this approach to other platforms.

Summary

This paper investigates the sustainability of X's Community Notes crowdsourced moderation system, finding that note publication significantly boosts future contributor activity but that only ~10% of notes are rated 'helpful'—a rate that is declining. The authors argue this low and falling helpfulness rate threatens contributor motivation and the long-term viability of bridging-based content moderation. The findings have implications for other platforms adopting similar systems.

Key Points

  • Using regression discontinuity design, the study shows that having a note published causally increases a contributor's future note-writing activity.
  • Only ~10% of submitted Community Notes are deemed 'helpful' and displayed to users, and this helpfulness rate is declining over time.
  • The low helpfulness rate risks demotivating contributors, threatening the system's long-term sustainability.
  • Community Notes uses a bridging algorithm requiring cross-partisan agreement, making helpfulness harder to achieve as the contributor base diversifies.
  • Findings are relevant to other major platforms (e.g., Meta) that are adopting or considering similar crowdsourced moderation mechanisms.

Cited by 2 pages

PageTypeQuality
X Community NotesProject54.0
X.com Platform EpistemicsApproach20.0

Cached Content Preview

HTTP 200Fetched Mar 20, 202633 KB
[License: CC BY-SA 4.0](https://info.arxiv.org/help/license/index.html#licenses-available)

arXiv:2510.00650v1 \[cs.SI\] 01 Oct 2025

# Threats to the sustainability of Community Notes on X

Report issue for preceding element

Zahra Arjmandi-Lari1,
Alexios Mantzarlis2,
Tom Stafford3The authors assert a Creative Commons Attribution (CC BY 4.0) License for this preprint

Report issue for preceding element

###### Abstract

Report issue for preceding element

Community Notes are emerging as an important option for content moderation. The Community Notes system pioneered by Twitter, now known as X, uses a bridging algorithm to identify user-generated context with upvotes across political divides, supposedly spinning consensual gold from partisan straw. It is important to understand the nature of the community behind Community Notes, especially as the feature has now been imitated by several billion-user platforms. We look for signs of stability and disruption in the X Community Notes community and interrogate the motivations other than partisan animus (Allen, Martel, and Rand [2022](https://arxiv.org/html/2510.00650v1#bib.bib1 "")) which may be driving users to contribute. We conduct a novel analysis of the impact of having a note published, which requires being considered “helpful” by the bridging algorithm, utilising a regression discontinuity design. This allows stronger causal inference than conventional methods used with observational data. Our analysis shows the positive effect on future note authoring of having a note published. This highlights the risk of the current system, where the proportion of notes considered “helpful” (and therefore shown to users on X) is low,  10%, and declining. This analysis has implications for the future of Community Notes on X and the extension of this approach to other platforms.

Report issue for preceding element

## 1 Introduction

Report issue for preceding element

Content moderation is an essential function of a digital platform, yet it is also a highly disputed one (Gillespie [2018](https://arxiv.org/html/2510.00650v1#bib.bib12 "")). Top-down decisions to remove or label (or not) a piece of content have led to advertiser boycotts, consumer complaints, and political pressure. While large online platforms have extensive policy guidelines describing violative behavior, the diversity of the speech they govern makes it nearly impossible to apply those policies completely consistently, and their scale means that even a small percentage of errors can affect thousands or millions of users.

Report issue for preceding element

Moderation decisions can be particularly challenging when it comes to content that is harder to define, such as misinformation. There is at least a baseline of consensus around what constitutes sexual or violent content; misinformation is by definition contextual and requires high-quality contradictory evidence to be available. Still, the public (Ejaz et al. [2024](https://arxiv.org/html/25

... (truncated, 33 KB total)
Resource ID: c64c6bdd2df72736 | Stable ID: OTFkNjkzZD