Longterm Wiki
Updated 2026-02-06HistoryData
Page StatusResponse
Edited 7 days ago1.4k words
48
QualityAdequate
55
ImportanceUseful
13
Structure13/15
311505%13%
Updated every 6 weeksDue in 5 weeks
Issues1
QualityRated 48 but structure suggests 87 (underrated by 39 points)

Design Sketches for Collective Epistemics

Approach

Design Sketches for Collective Epistemics

Forethought Foundation's five proposed technologies for improving collective epistemics: community notes for everything, rhetoric highlighting, reliability tracking, epistemic virtue evals, and provenance tracing. These design sketches aim to shift society toward high-honesty equilibria.

1.4k words

Overview

In 2025, Forethought Foundation published "Design Sketches for Collective Epistemics," a research report outlining five proposed technologies that could shift society toward what they call "high-honesty equilibria"—environments where honesty is generally the best policy because lies and obfuscation are reliably caught and penalized. The report was also discussed on the EA Forum.

The core thesis is that if it becomes easier to track what's trustworthy and what isn't, the resulting equilibrium would reward honesty. This could improve collective decision-making broadly, and in particular give humanity a better shot at handling the transition to more advanced AI systems.

The Five Design Sketches

The five proposed technologies are designed to work synergistically, each reinforcing the others:

ToolCore FunctionMaturityEstimated Impact
Community Notes for EverythingCross-platform AI-generated context annotationsMedium (builds on X Community Notes, 500K+ contributors)High (billions of internet users)
Rhetoric HighlightingAutomated detection of misleading rhetoricLow-Medium (GPT-4 at 79-90% fallacy detection)Medium (high-stakes content first)
Reliability TrackingTopic-specific track record scoringLow-Medium (prediction platforms exist; general system conceptual)High (addresses accountability gap)
Epistemic Virtue EvalsAI honesty and calibration benchmarksMedium-High (TruthfulQA, SimpleQA, HELM exist)High (benchmarks drive industry behavior)
Provenance TracingTransparent claim origin and evidence chainsLow (mostly conceptual; citation infra exists)Very High (foundational if built)
Loading diagram...

Why This Matters for AI Safety

Forethought argues these tools are particularly important from an existential risk perspective:

  1. Power concentration resistance: High-honesty environments make it harder for unscrupulous actors to concentrate power through manipulation
  2. Risk tracking: Societies with better epistemic tools can more effectively track and respond to catastrophe risks
  3. Robust positioning: Epistemically healthy environments improve humanity's positioning for handling major challenges, including advanced AI

The connection to the AI transition is direct: if civilizational competence depends partly on epistemic health, then tools that improve collective epistemics reduce the risk of catastrophic outcomes during the transition to powerful AI systems.

Feasibility and Cost Considerations

The report provides rough cost estimates for each technology:

ToolEstimated Cost per UnitKey Bottleneck
Community Notes for Everything$0.01–0.10 per tweetSocial adoption, platform cooperation
Rhetoric Highlighting$1–hundreds per hour of readingSpeed and cost of multiple LLM calls
Reliability TrackingCents to hundreds per person assessedGround truth determination, legal exposure
Epistemic Virtue EvalsModerate (benchmark development)Goodharting, methodology design
Provenance TracingVariable (infrastructure cost)Scale, error accumulation in recursive LLM use

The report emphasizes that these are "design sketches" rather than final specifications, and explicitly invites builders to develop implementations that may differ substantially while pursuing the same goals.

Interaction Effects: How the Tools Reinforce Each Other

Forethought presents these as five independent tools, but their real power comes from specific interactions between them:

Reliability Tracking → Community Notes: If an author has a poor reliability score on a topic, their claims on that topic get automatically prioritized for community notes annotation. The system focuses its limited resources on content from sources with the worst track records.

Provenance Tracing → Rhetoric Highlighting: When provenance tracing reveals that a claim has mutated significantly from its original source, rhetoric highlighting can flag the specific sentences where distortion occurred. "This sentence attributes X to Study Y, but the original study actually found Z."

Epistemic Virtue Evals → Community Notes: AI-generated community notes are themselves AI outputs. Epistemic virtue evals ensure the AI systems writing notes are well-calibrated and non-sycophantic—preventing the note system from introducing its own biases.

Rhetoric Highlighting → Reliability Tracking: Aggregating rhetoric flags across an author's body of work creates a "rhetoric profile" that feeds into reliability scores. An author who consistently uses emotional manipulation or misrepresents citations gets a lower reliability score even when their factual claims are technically accurate.

Provenance Tracing → Reliability Tracking: When evaluating whether someone's factual claims are accurate, provenance tracing provides the evidence chain needed to score them. Instead of manual fact-checking, the reliability system can query the provenance database to verify claims automatically.

Reliability Tracking → Provenance Tracing: Source reliability scores feed into provenance trust propagation. A citation chain passing through multiple high-reliability sources gets a higher trust score than one passing through sources with poor track records.

Lessons from Past Attempts

Several previous projects have attempted pieces of this vision and either failed or stagnated. Understanding why is essential for new attempts:

PunditTracker (2013-?): Tracked pundit predictions with letter grades. Demonstrated strong initial interest but eventually shut down. Likely failure modes: Insufficient funding model (no clear revenue), legal risk from rating named individuals, insufficient scale to create social pressure, and no integration with content consumption (users had to visit a separate site).

Fact-checking organizations: PolitiFact, FactCheck.org, and the Washington Post Fact Checker have operated for 15+ years but haven't fundamentally shifted incentives toward honesty. Key lesson: Fact-checking as a separate activity that readers must seek out has limited impact. The information needs to be embedded at the point of content consumption—which is exactly what Community Notes for Everything proposes.

X Community Notes itself: Highly effective when notes display, but 96.7% of content spread happens before notes appear. Key lesson: Speed is essential. Human-driven consensus processes are too slow for viral content. AI assistance is not optional—it's required for the timing to work.

Academic citation analysis tools: Semantic Scholar, Google Scholar, and Connected Papers exist but are used primarily by researchers, not the general public. Key lesson: Tools built for expert users don't automatically reach mainstream audiences. Provenance tracing needs a consumer-grade interface, not just research infrastructure.

Media bias ratings: Ad Fontes Media and AllSides rate entire outlets but don't assess individual claims. Key lesson: Outlet-level ratings are too coarse. The same outlet can publish both excellent and terrible analysis. Claim-level and article-level assessment is necessary.

Common patterns across failures:

  1. Separate destination: Tools that require users to visit a new site fail. Integration into existing consumption flows is essential.
  2. Insufficient funding models: Epistemic infrastructure is a public good. Market incentives alone are usually insufficient.
  3. Legal vulnerability: Rating named individuals or organizations creates legal risk that has killed multiple projects.
  4. Speed vs. quality trade-off: Slow, high-quality assessment is useless for viral content. The Forethought sketches' emphasis on AI-driven speed directly addresses this.

Integration with Existing Approaches

Several of the proposed technologies build on or extend existing work:

  • Community Notes for Everything extends X's Community Notes system, which already demonstrates the bridging algorithm concept at scale
  • Provenance Tracing shares goals with content authentication technologies like C2PA, but focuses on knowledge claims rather than media authenticity
  • Epistemic Virtue Evals connects to existing AI safety evaluation frameworks and benchmarks like TruthfulQA and sycophancy resistance testing
  • Reliability Tracking builds on work by forecasting platforms like Metaculus and research on prediction markets

Who Builds This? Stakeholder Analysis

A key question the Forethought report raises but doesn't fully answer: who funds, builds, and maintains these tools?

StakeholderInterestLikely Role
AI labsEpistemic virtue evals improve trust in their products; but may resist unfavorable resultsCould fund evals; unlikely to fund tools that flag their own claims
Tech platformsCommunity notes reduce liability; but annotation may drive users awayKey partners for deployment; may resist external annotation
GovernmentsBetter-informed public discourse serves democratic interestsCould fund as public infrastructure; regulatory mandates possible
Philanthropic orgsEpistemic infrastructure is a classic public goodMost likely early funders (Coefficient Giving, Gates Foundation, etc.)
Academic institutionsResearch value; provenance tracing benefits researchers directlyCould build and host infrastructure; credibility advantage
News organizationsReliability tracking either validates or threatens their modelCould be early adopters or fierce opponents depending on implementation
General publicBenefits from all tools; but may not pay for them directlyAdoption depends on ease of use; browser extensions are the key interface

Key Uncertainties

Key Questions

  • ?Can these tools achieve adoption without platform cooperation, or is buy-in from major tech companies essential?
  • ?How resistant are these systems to adversarial gaming by motivated actors?
  • ?Will the economics improve fast enough (via cheaper LLM inference) to make rhetoric highlighting and provenance tracing viable at scale?
  • ?Could these tools be co-opted for censorship or political control rather than genuine epistemic improvement?
  • ?Is the 'high-honesty equilibrium' thesis correct—does better tracking actually shift incentives toward honesty?

Further Reading

  • Original Report: Design Sketches for Collective Epistemics — Forethought Foundation (2025)
  • EA Forum Discussion: Some Tools for Collective Epistemics
  • Related Wiki Pages: Epistemic Infrastructure, X Community Notes, Content Authentication

Related Pages

Top Related Pages

Approaches

Epistemic Virtue EvalsCommunity Notes for EverythingAI Content Provenance TracingAI System Reliability TrackingAI-Assisted Rhetoric Highlighting

Concepts

Epistemic HealthX Community NotesAI Content Authentication