Edited 2 months ago2.0k words2 backlinksUpdated every 3 weeksOverdue by 46 days
72QualityGood •Quality: 72/100LLM-assigned rating of overall page quality, considering depth, accuracy, and completeness.Structure suggests 10078.5ImportanceHighImportance: 78.5/100How central this topic is to AI safety. Higher scores mean greater relevance to understanding or mitigating AI risk.67ResearchModerateResearch Value: 67/100How much value deeper investigation of this topic could yield. Higher scores indicate under-explored topics with high insight potential.
Content8/13
SummarySummaryBasic text summary used in search results, entity link tooltips, info boxes, and related page cards.ScheduleScheduleHow often the page should be refreshed. Drives the overdue tracking system.EntityEntityYAML entity definition with type, description, and related entries.Edit historyEdit historyTracked changes from improve pipeline runs and manual edits.crux edit-log view <id>OverviewOverviewA ## Overview heading section that orients readers. Helps with search and AI summaries.
Tables19/ ~8TablesData tables for structured comparisons and reference material.Diagrams2/ ~1DiagramsVisual content — Mermaid diagrams, charts, or Squiggle estimate models.–Int. links9/ ~16Int. linksLinks to other wiki pages. More internal links = better graph connectivity.Add links to other wiki pagesExt. links26/ ~10Ext. linksLinks to external websites, papers, and resources outside the wiki.Footnotes0/ ~6FootnotesFootnote citations [^N] with source references at the bottom of the page.Add [^N] footnote citationsReferences8/ ~6ReferencesCurated external resources linked via <R> components or cited_by in YAML.Quotes0QuotesSupporting quotes extracted from cited sources to back up page claims.crux citations extract-quotes <id>Accuracy0AccuracyCitations verified against their sources for factual accuracy.crux citations verify <id>RatingsN:4.2 R:6.8 A:5.5 C:7.5RatingsSub-quality ratings: Novelty, Rigor, Actionability, Completeness (0-10 scale).Backlinks2BacklinksNumber of other wiki pages that link to this page. Higher backlink count means better integration into the knowledge graph.
Issues3
QualityRated 72 but structure suggests 100 (underrated by 28 points)
Links16 links could use <R> components
StaleLast edited 67 days ago - may need review
Pause / Moratorium
Concept
Pause / Moratorium
Comprehensive analysis of pause/moratorium proposals finding they would provide very high safety benefits if implemented (buying time for safety research to close the growing capability-safety gap) but face critical enforcement and coordination challenges with zero current adoption by major labs. The FLI 2023 open letter garnered 30,000+ signatures but resulted in no actual slowdown, highlighting severe tractability issues despite theoretical effectiveness.
Future of Life InstituteOrganizationFuture of Life InstituteComprehensive profile of FLI documenting $25M+ in grants distributed (2015: $7M to 37 projects, 2021: $25M program), major public campaigns (Asilomar Principles with 5,700+ signatories, 2023 Pause ...Quality: 46/100
People
Stuart RussellPersonStuart RussellStuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 univer...Quality: 30/100
Risks
AI Development Racing DynamicsRiskAI Development Racing DynamicsRacing dynamics analysis shows competitive pressure has shortened safety evaluation timelines by 40-60% since ChatGPT's launch, with commercial labs reducing safety work from 12 weeks to 4-6 weeks....Quality: 72/100
Approaches
Pause AdvocacyApproachPause AdvocacyComprehensive analysis of pause advocacy as an AI safety intervention, estimating 15-40% probability of meaningful policy implementation by 2030 with potential to provide 2-5 years of additional sa...Quality: 91/100
2k words · 2 backlinks
Overview
Pause and moratorium proposals represent the most direct governance intervention for AI safety: deliberately slowing or halting frontier AI development to allow safety research, governance frameworks, and societal preparation to catch up with rapidly advancing capabilities. These proposals range from targeted pauses on specific capability thresholds to comprehensive moratoria on all advanced AI development, with proponents arguing that the current pace of development may be outstripping humanity's ability to ensure safe deployment.
The most prominent call for a pause came in March 2023, when the Future of Life InstituteOrganizationFuture of Life InstituteComprehensive profile of FLI documenting $25M+ in grants distributed (2015: $7M to 37 projects, 2021: $25M program), major public campaigns (Asilomar Principles with 5,700+ signatories, 2023 Pause ...Quality: 46/100 (FLI) published an open letter calling for a six-month pause on training AI systems more powerful than GPT-4. Released just one week after GPT-4's launch, the letter garnered over 30,000 signatures, including prominent AI researchers such as Yoshua Bengio and Stuart RussellPersonStuart RussellStuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 univer...Quality: 30/100, as well as technology leaders like Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch and funding history, Neuralink BCI developm...Quality: 38/100 and Steve Wozniak. The letter cited risks including AI-generated propaganda, extreme automation of jobs, and a society-wide loss of control. However, no major AI laboratory implemented a voluntary pause, and the letter's six-month timeline passed without meaningful slowdown in frontier development. As MIT Technology Review noted six months later, AI companies instead directed "vast investments in infrastructure to train ever-more giant AI systems."
The fundamental logic behind pause proposals is straightforward: if AI development is proceeding faster than our ability to make it safe, slowing development provides time for safety work. As Bengio et al. wrote in Science in May 2024, "downside artificial intelligence risks must be managed effectively and urgently if posited AI benefits are to be realized safely." However, implementation faces severe challenges including competitive dynamics between nations and companies, enforcement difficulties, and concerns that pauses might push development underground or to jurisdictions with fewer safety constraints. These proposals remain controversial even within the AI safety community, with some arguing they are essential for survival and others viewing them as impractical or counterproductive.
Risk Assessment & Impact
Dimension
Assessment
Rationale
Confidence
Safety Uplift
High (if implemented)
Would buy time for safety research
High
Capability Uplift
Negative
Explicitly slows capability development
High
Net World Safety
Unclear
Could help if coordinated; could backfire if unilateral
Medium
Lab Incentive
Negative
Labs strongly opposed; competitive dynamics
High
Research Investment
$1-5M/yr
Advocacy organizations; FLI, PauseAI
Medium
Current Adoption
None
Advocacy only; no major labs paused
High
Core Arguments
Case for Pausing
Diagram (loading…)
flowchart TD
SPEED[AI Development Speed] --> GAP{Safety Gap?}
SAFETY[Safety Research Speed] --> GAP
GAP -->|Growing| RISK[Increasing Risk]
GAP -->|Stable| MANAGE[Manageable]
GAP -->|Shrinking| GOOD[Safety Catching Up]
RISK --> OPTION1[Continue Development]
RISK --> OPTION2[Pause Development]
OPTION1 --> RACE[Racing to Danger]
OPTION2 --> TIME[Buy Time for Safety]
TIME --> RESEARCH[More Safety Research]
TIME --> GOV[Better Governance]
TIME --> PREP[Societal Preparation]
RESEARCH --> SAFER[Safer Development]
GOV --> SAFER
PREP --> SAFER
style RISK fill:#ffcccc
style SAFER fill:#d4edda
style RACE fill:#ff9999
Argument
Description
Strength
Safety-Capability Gap
Safety research not keeping pace with capabilities
Strong if gap is real
Irreversibility
Some AI risks may be impossible to reverse once realized
Strong for existential risks
Precautionary Principle
Burden of proof should be on developers to show safety
Philosophically contested
Coordination Signal
Demonstrates seriousness; creates space for governance
Moderate
Research Time
Enables catch-up on interpretability, alignment
Strong
Case Against Pausing
Argument
Description
Strength
Enforcement
Unenforceable without international agreement
Strong
Displacement
Development moves to less cautious actors
Moderate-Strong
Lost Benefits
Delays positive AI applications
Moderate
Talent Dispersion
Safety researchers may leave paused organizations
Moderate
False Security
Pause without progress creates complacency
Moderate
Definition Problems
Hard to define what to pause
Strong
Pause Proposals Analyzed
FLI Open Letter (2023)
Aspect
Detail
Scope
Training systems more powerful than GPT-4
Duration
Six months (renewable)
Signatories
30,000+ including Yoshua BengioPersonYoshua BengioComprehensive biographical overview of Yoshua Bengio's transition from deep learning pioneer (Turing Award 2018) to AI safety advocate, documenting his 2020 pivot at Mila toward safety research, co...Quality: 39/100, Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch and funding history, Neuralink BCI developm...Quality: 38/100, Stuart RussellPersonStuart RussellStuart Russell (born 1962) is a British computer scientist and UC Berkeley professor who co-authored the dominant AI textbook 'Artificial Intelligence: A Modern Approach' (used in over 1,500 univer...Quality: 30/100, Steve Wozniak, Yuval Noah Harari
Notable critiques: AI researcher Andrew NgPersonAndrew NgAndrew Ng is one of the most prominent figures in AI education and research. He co-founded Google Brain, served as Chief Scientist at Baidu, and co-founded Coursera. His online courses have taught ...argued that "there is no realistic way to implement a moratorium" without government intervention, which would be "anti-competitive" and "awful innovation policy." Reid Hoffman criticized the letter as "virtue signaling" that would hurt the cause by alienating the AI developer community needed to achieve safety goals.
PauseAI Movement
Aspect
Detail
Founded
May 2023 in Utrecht, Netherlands by software entrepreneur Joep Meindertsma
Network of local organizations; US chapter led by Holly Elmore, UK by Joseph Miller (Oxford PhD)
Approach
Grassroots activism, protests at AI labs (OpenAI Feb 2024, Anthropic Nov 2024), policy advocacy
Policy Asks
Global pause enforced through international treaty; democratic control over AI development
Key Actions
International protests in May 2024 timed to Seoul AI Safety Summit; protests held in San Francisco, New York, Berlin, Rome, Ottawa, London
Academic Proposals
Proposal
Scope
Mechanism
Compute Caps
Limit training compute
Hardware governance
Capability Gates
Pause at defined capability thresholds
Eval-based triggers
Conditional Pause
Pause if safety benchmarks not met
RSP-like framework
Research Moratoria
Pause specific capability research
Targeted restrictions
Implementation Challenges
Coordination Problems
Challenge
Description
Severity
Potential Solution
International Competition
US-China dynamics; neither wants to pause first
Critical
Treaty with verification
Corporate Competition
First-mover advantages; defection incentives
High
Regulatory mandate
Verification
How to confirm compliance
High
Compute monitoring
Definition
What counts as "frontier" AI
High
Clear technical thresholds
Enforcement Mechanisms
Mechanism
Feasibility
Effectiveness
Notes
Voluntary Compliance
Low
Very Low
No incentive to comply
National Regulation
Medium
Medium
Jurisdictional limits
International Treaty
Low-Medium
High if achieved
Requires major power agreement
Compute Restrictions
Medium
Medium-High
Physical infrastructure trackable
Social Pressure
Medium
Low
Insufficient against strong incentives
Unintended Consequences
Consequence
Likelihood
Severity
Mitigation
Development Displacement
High
High
International coordination
Underground Development
Medium
Very High
Compute monitoring
Safety Researcher Exodus
Medium
Medium
Continued safety funding
Competitive Disadvantage
High
Variable
Coordinated action
Delayed Benefits
High
Medium
Risk-benefit analysis
Historical Precedents
Domain
Intervention
Outcome
Lessons
Nuclear Weapons
Various moratoria and treaties
Partial success; proliferation continued
Verification essential
Human Cloning
Research moratoria
Generally effective
Narrow scope helps
Gain-of-Function
Research pause (2014-2017)
Temporary; research resumed
Pressure to resume
Recombinant DNA
Asilomar conference (1975)
Self-regulation worked initially
Community buy-in crucial
CFCs
Montreal Protocol
Highly successful
Clear harm identification
Lessons for AI Pauses
Narrow scope is more enforceable than broad moratoria
Verification mechanisms are essential for compliance
International coordination requires identifying mutual interests
Community buy-in from researchers enables voluntary compliance
Clear triggering conditions help define when restrictions apply
Scalability Assessment
Dimension
Assessment
Rationale
International Scalability
Unknown
Depends on coordination
Enforcement Scalability
Partial
Compute monitoring possible
SI Readiness
Yes (if works)
Would prevent reaching SI until prepared
Deception Robustness
N/A
External policy; doesn't address model behavior
Strategic Considerations
When Pauses Might Work
Condition
Importance
Current Status
International Agreement
Critical
Very limited
Clear Triggers
High
Undefined
Verification Methods
High
Underdeveloped
Alternative Pathway
Medium
Safety research ongoing
Industry Buy-In
Medium-High
Very low
Alternative Approaches
Alternative
Relationship to Pause
Tradeoffs
Differential Progress
Accelerate safety, not slow capabilities
Competitive with capabilities
Responsible Scaling Policies
Conditional pauses at thresholds
Voluntary; lab-controlled
Compute Governance
Indirect slowdown through resource control
More enforceable
International Coordination
Framework for coordinated pause
Slower to achieve
Quick Assessment
Dimension
Assessment
Evidence
Tractability
Low
Severe coordination and enforcement challenges; no major lab has voluntarily paused
Effectiveness
Very High (if implemented)
Would directly address timeline concerns by buying time for safety research
Neglectedness
Medium
Active advocacy (FLI, PauseAI); major gap in implementation and enforcement mechanisms
Current Maturity
Early Advocacy
FLI letter catalyzed debate but no binding commitments achieved
Time Horizon
Immediate-Long Term
Could theoretically be implemented quickly but requires international coordination
Key Proponents
FLI, PauseAI, Yoshua BengioPersonYoshua BengioComprehensive biographical overview of Yoshua Bengio's transition from deep learning pioneer (Turing Award 2018) to AI safety advocate, documenting his 2020 pivot at Mila toward safety research, co...Quality: 39/100
Grassroots movements and prominent AI researchers
Key Opponents
Major AI Labs, Andrew Ng
Competitive dynamics and concerns about practicality
Risks Addressed
If implemented effectively, pause/moratorium would address:
Risk
Mechanism
Effectiveness
AI Development Racing DynamicsRiskAI Development Racing DynamicsRacing dynamics analysis shows competitive pressure has shortened safety evaluation timelines by 40-60% since ChatGPT's launch, with commercial labs reducing safety work from 12 weeks to 4-6 weeks....Quality: 72/100
Eliminates competitive pressure
Very High
Safety-Capability Gap
Time for safety research
Very High
Governance Lag
Time for policy development
High
Societal Preparation
Time for adaptation
High
****
Prevents deployment of unaligned systems
Very High (during pause)
Limitations
Enforcement Infeasibility: No mechanism to enforce global compliance
A widely-signed open letter published by the Future of Life Institute in March 2023, calling on all AI labs to pause for at least 6 months the training of AI systems more powerful than GPT-4. It argues that AI development has entered a dangerous uncontrolled race and calls for shared safety protocols, independent auditing, and accelerated AI governance frameworks before proceeding with more powerful systems.
Official website of Yoshua Bengio, Turing Award-winning AI researcher and deep learning pioneer, who has become one of the most prominent voices advocating for AI safety. The site highlights his current safety-focused work including chairing the International AI Safety Report, co-founding LawZero (a non-profit developing safe-by-design AI), and his 'Scientist AI' framework for building capable but non-agentic systems.
MIT Technology Review interviews Max Tegmark six months after the Future of Life Institute's open letter calling for a pause on advanced AI development. While the letter succeeded in shifting the Overton window and normalizing public discussion of existential AI risk, no meaningful U.S. regulation resulted and all major AI companies continued development at full speed. Tegmark argues that only government intervention via FDA-style oversight can create the conditions for an enforceable pause, since no single company can pause unilaterally without competitive disadvantage.
PauseAI is an advocacy movement calling for an international pause on the development of advanced AI systems until adequate safety measures and governance frameworks are in place. The organization coordinates activists, provides educational resources, and lobbies policymakers to take urgent action on AI risk. It represents a direct-action approach to AI safety that prioritizes preventing catastrophic outcomes over accelerating beneficial AI.
Carnegie Endowment analysis examines whether high-profile AI safety summits (like Bletchley Park and Seoul) translate into meaningful governance progress or remain largely ceremonial. The piece evaluates the gap between international AI governance rhetoric and substantive policy coordination, arguing that geopolitical competition risks turning AI governance into a performative arms race rather than genuine risk reduction.
This UN press release covers a Secretary-General statement regarding the establishment or activities of an international scientific panel on artificial intelligence, reflecting the UN's efforts to create a global governance and oversight body for AI. It represents part of the broader UN initiative to coordinate international AI safety and governance through multilateral institutions.
The Future of Life Institute (FLI) is a nonprofit organization focused on steering transformative technologies, particularly AI, away from catastrophic risks and toward beneficial outcomes. They operate across policy advocacy, research funding, education, and outreach to promote responsible AI development. FLI has been influential in key AI safety milestones including the open letter on AI risks and the Asilomar AI Principles.
The Centre for the Governance of AI (GovAI) is a leading research organization dedicated to helping decision-makers navigate the transition to a world with advanced AI. It produces rigorous research on AI governance, policy, and societal impacts, while fostering a global talent pipeline for responsible AI oversight. GovAI bridges technical AI safety concerns with practical policy recommendations.
Corporate AI Safety ResponsesApproachCorporate AI Safety ResponsesMajor AI labs invest $300-500M annually in safety (5-10% of R&D) through responsible scaling policies and dedicated teams, but face 30-40% safety team turnover and significant implementation gaps b...Quality: 68/100Multi-Agent SafetyApproachMulti-Agent SafetyMulti-agent safety addresses coordination failures, conflict, and collusion risks when AI systems interact. A 2025 report from 50+ researchers identifies seven key risk factors; empirical studies s...Quality: 68/100
Risks
Multipolar Trap (AI Development)RiskMultipolar Trap (AI Development)Analysis of coordination failures in AI development using game theory, documenting how competitive dynamics between nations (US $109B vs China $9.3B investment in 2024 per Stanford HAI 2025) and la...Quality: 91/100
Analysis
Anthropic Impact Assessment ModelAnalysisAnthropic Impact Assessment ModelModels Anthropic's net impact on AI safety by weighing positive contributions (safety research $100-200M/year, Constitutional AI as industry standard, largest interpretability team globally, RSP fr...Quality: 55/100
Other
Elon MuskPersonElon MuskComprehensive profile of Elon Musk's role in AI, documenting his early safety warnings (2014-2017), OpenAI founding and contentious departure, xAI launch and funding history, Neuralink BCI developm...Quality: 38/100Andrew NgPersonAndrew NgAndrew Ng is one of the most prominent figures in AI education and research. He co-founded Google Brain, served as Chief Scientist at Baidu, and co-founded Coursera. His online courses have taught ...
Concepts
Alignment Policy OverviewAlignment Policy OverviewThis is a stub overview page that lists four policy/governance topic areas (RSPs, model specs, evaluation governance, pause/moratorium) with one-line descriptions and links to deeper pages. It cont...Quality: 21/100International Compute RegimesConceptInternational Compute RegimesComprehensive analysis of international AI compute governance finds 10-25% chance of meaningful regimes by 2035, but potential for 30-60% reduction in racing dynamics if achieved. First binding tre...Quality: 67/100
Key Debates
Should We Pause AI Development?CruxShould We Pause AI Development?Comprehensive synthesis of the AI pause debate showing moderate expert support (35-40% of 2,778 researchers) and high public support (72%) but very low implementation feasibility, with all major la...Quality: 47/100AI Structural Risk CruxesCruxAI Structural Risk CruxesAnalyzes 12 key uncertainties about AI structural risks across power concentration, coordination feasibility, and institutional adaptation. Provides quantified probability ranges: US-China coordina...Quality: 66/100AI Safety Solution CruxesCruxAI Safety Solution CruxesA comprehensive structured mapping of AI safety solution uncertainties across technical, alignment, governance, and agentic domains, using probability-weighted crux frameworks with specific estimat...Quality: 65/100
Organizations
AnthropicOrganizationAnthropicComprehensive reference page on Anthropic covering financials ($380B valuation, $14B ARR at Series G growing to $19B by March 2026), safety research (Constitutional AI, mechanistic interpretability...Quality: 74/100OpenAIOrganizationOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to Public Benefit Corporation, with detailed analysis of governance crisis, 2024-2025 ownership restructuri...Quality: 62/100xAIOrganizationxAIComprehensive profile of xAI covering its founding by Elon Musk in 2023, rapid growth to $230B valuation and $3.8B revenue, development of Grok models, and controversial 'truth-seeking' safety appr...Quality: 48/100