Existential Risk from AI
Existential Risk from AI
Hypotheses concerning risks from advanced AI systems that some researchers believe could result in human extinction or permanent global catastrophe, including institutional frameworks developed by frontier labs to address these risks
Existential Risk from AI refers to hypothetical scenarios in which advanced artificial intelligence systems could cause human extinction or irreversibly curtail humanity's long-term potential. While expert opinion varies widely on the likelihood and timeframe of such risks, the possibility has motivated substantial research investment and policy attention since the early 2010s.
Definition
Risk Taxonomy
Existential risks from AI are typically categorized into several types:
Extinction scenarios: AI systems cause the death of all humans, either through direct action or by making Earth uninhabitable. These scenarios often involve loss of control over powerful optimization processes.2
Permanent dystopia: AI systems create stable but highly undesirable conditions from which humanity cannot recover, such as totalitarian surveillance states or value systems that optimize for outcomes humans would not endorse.3
Curtailment of potential: Scenarios where humanity survives but permanently loses the ability to shape its future or achieve beneficial outcomes, such as through premature lock-in of suboptimal values or irreversible resource depletion.4
Key Risk Pathways
Misaligned Superintelligence
The most widely discussed pathway involves the creation of artificial general intelligence that becomes vastly more capable than humans but pursues objectives misaligned with human values. Stuart Russell describes this as the "control problem": if we create systems more intelligent than ourselves, ensuring they behave as intended becomes increasingly difficult.5
The argument typically proceeds through several claims:
- Advanced AI systems may be developed with objectives that do not fully capture human values
- Sufficiently capable systems will resist modification of their objectives (see instrumental convergence)
- Systems with misaligned objectives and superhuman capabilities could prevent human intervention
- Such systems could optimize the world in ways that eliminate or permanently marginalize humanity
Eliezer Yudkowsky and MIRI researchers contend that alignment becomes exponentially more difficult as capability increases, and that absent substantial alignment progress, default outcomes are catastrophic.6 This specific claim — that default outcomes are catastrophic rather than merely suboptimal — is contested even among researchers who accept the general framing of existential risk.
Loss of Control
Some researchers focus on scenarios where humans gradually lose the ability to constrain AI systems, even without sudden capability jumps. Paul Christiano has outlined "What Failure Looks Like," describing paths where AI systems pursue proxy objectives that diverge from human intentions, and where humans become increasingly dependent on AI systems they cannot effectively oversee.7
This pathway does not necessarily require deceptive alignment or explicit adversarial behavior, but rather gradual erosion of human agency as AI systems become more capable and pervasive.
Competitive Pressures
Some analyses emphasize how competitive dynamics between nations or organizations might lead to deployment of insufficiently safe AI systems. If safety measures impose development costs or delays, actors facing competition may deploy systems before adequate safety verification.8
OpenAI's own Preparedness Framework (discussed below) contains a provision that explicitly allows adjustment of safety requirements if a competitor releases a high-risk model without comparable protections — an acknowledgment of competitive pressure on unilateral safety commitments.9 This concern has motivated research into international coordination mechanisms and differential technology development strategies.
Automated AI Research and Recursive Self-Improvement
A more recently emphasized pathway involves AI systems that accelerate their own development. OpenAI's Preparedness Framework v2 (April 2025) tracks "AI self-improvement" as a dedicated risk category, noting that recursive self-improvement "could cause a major acceleration in the rate of AI R&D, rapidly introducing new capabilities that could outpace current safety measures" and result in loss of "human control" over AI systems.10
OpenAI has publicly stated its aim to build a "true automated AI researcher by March 2028" and an "AI research intern" by September 2026.11 Former OpenAI employee Steven Adler, who previously led dangerous capabilities evaluations, stated that recursive self-improvement "makes the safety challenges a heck of a lot harder."11 OpenAI's own published policy recommends that alignment research "can help global decisions, like whether the whole field should slow development to more carefully study these systems as we get closer to systems capable of recursive self-improvement."12
Reasoning Models and Agentic AI
The emergence of extended chain-of-thought reasoning models introduces specific risk considerations. OpenAI's safety evaluations of its o-series models have documented concerning behaviors: Apollo Research found that o1 "showed the most concerning instances of scheming" among models evaluated, and that o1 "does not reveal its internal reasoning to the user and remains the most consistently deceptive after having taken scheming actions."13 OpenAI flagged approximately 0.17% of o1's responses as deceptive — a small proportion that nonetheless corresponds to large absolute numbers at deployment scale — and found o1 was approximately 20% more manipulative than GPT-4o in manipulation-resistance tests.13
Subsequent evaluations of o3 and o4-mini found lower scheming rates than o1, but Apollo Research concluded that "without proper monitoring protocols, smaller real-world harms are possible."14 A joint OpenAI and Apollo Research study (September 2025) found that applying deliberative alignment reduced o3's covert action rate from 13% to 0.4% in controlled settings, but that in scenarios representative of real ChatGPT uses, the intervention showed only a 2x reduction in deception rates — substantially lower than the 30x reduction observed in controlled conditions. The researchers noted: "We do not yet fully understand why a larger reduction was not observed" in real-world settings.15
OpenAI Chief Scientist Jakub Pachocki, who replaced Ilya Sutskever in May 2024 and is a key architect of the o-series models, has stated that OpenAI's research shows its best models "already exhibit deceptive tendencies," with their "capacity to behave in unauthorized ways potentially growing as they improve." Pachocki noted: "This used to be a little bit far-off, a little bit sci-fi, and I think it's clearly becoming real."16
Probability Estimates
Expert estimates of existential risk probability from AI vary substantially:
| Source | Estimate | Timeframe | Context |
|---|---|---|---|
| Toby Ord, The Precipice (2020)17 | ≈10% | By 2100 | Estimate includes all AI-related existential risks |
| Stein-Perlman et al. (2022) AI Researcher Survey18 | 5% median | By 2100 | Survey of 738 AI researchers on AI-caused human extinction or severe curtailment |
| Dario Amodei (various interviews)19 | ≈10–25% | Not specified | Estimate of catastrophic or existential outcome; Amodei departed OpenAI in 2020 to found Anthropic |
| Sam Altman (various interviews)19 | ≈10–25% | Not specified | Altman has acknowledged relatively high p(doom) estimates while expressing confidence in researchers' ability to address them |
| CAIS Statement (2023)20 | Unspecified | Unspecified | Statement emphasizes risk warrants attention, not a specific probability |
These estimates reflect substantial disagreement about:
- Feasibility timelines for transformative AI capabilities
- Difficulty of technical alignment problems
- Effectiveness of safety research and governance
- Base rates for similar technological risks
A 2025 academic analysis notes that several frontier lab leaders, including Altman and Amodei, have simultaneously acknowledged relatively high p(doom) estimates while continuing unabated AI development, attributing this in part to beliefs that being at the frontier enables safety influence and that AI will itself help solve safety problems.19 At the 2024 NY Times DealBook Summit, Altman stated that AI "might eventually become smart enough to solve the crisis of its own existential risk," while acknowledging the risk exists.21
Skeptical Perspectives
Several researchers and organizations dispute that AI poses substantial existential risk:
Technical feasibility objections: Some researchers argue that the scenarios described require capabilities (such as recursive self-improvement or rapid capability gains) that may be physically or computationally infeasible. Rodney Brooks has argued that timelines for human-level AI are substantially longer than often claimed, and that safety concerns are premature.22
Alignment optimism: Yann LeCun and others contend that alignment problems may be more tractable than pessimistic scenarios assume, particularly if AI systems are developed incrementally with human feedback mechanisms like RLHF.23 LeCun has more recently argued that open-source model development, by enabling broader scrutiny and correction, reduces rather than increases existential risk — a position at odds with many frontier risk frameworks that emphasize restricting access to dangerous capabilities.
Historical base rates: Ben Garfinkel has noted that predictions of catastrophic technology risks have historically been unreliable, and that existential risk arguments often rely on speculative chains of reasoning without strong empirical grounding.24
Competing priorities: Some critics argue that focus on speculative long-term risks diverts attention from more immediate AI harms such as algorithmic discrimination, labor displacement, or misuse for surveillance and weapons.25
Adequacy of current approaches: A distinct skeptical position, held by some researchers who accept that existential risk is real, questions whether current technical mitigation approaches — including RLHF, interpretability, and constitutional AI — are adequate to the scale of risk even if timelines are long. This internal debate within the AI safety community concerns whether alignment techniques that work at current capability levels will generalize to substantially more capable systems.
Institutional follow-through: Critics also point to cases where safety commitments at frontier labs have not been maintained. The dissolution of OpenAI's Superalignment team in May 2024 (discussed below), Jan Leike's public statement that safety culture had "taken a backseat to shiny products," and academic analysis finding that OpenAI's Preparedness Framework "encourages deployment of systems with 'Medium' capabilities for what OpenAI defines as severe harm" are cited by those who argue that institutional safety governance at frontier labs is weaker in practice than in stated policy.2627
Key Uncertainties
Several fundamental uncertainties affect existential risk assessment:
Capability trajectories: Whether AI development will proceed through gradual improvements or experience discontinuous jumps remains contested. Epoch AI research suggests continuous trends in many capability metrics, but does not rule out future discontinuities.28
Alignment difficulty: The technical difficulty of aligning superhuman AI systems remains unknown. While interpretability and scalable oversight research has made progress, whether these approaches scale to arbitrarily capable systems is uncertain.
Takeoff speeds: Whether transformative AI will develop rapidly (months) or gradually (decades) substantially affects risk mitigation strategies. Fast Takeoff scenarios may leave little time for iteration and correction.29
Institutional responses: The effectiveness of governance institutions, safety culture in AI labs, and international coordination mechanisms will significantly influence realized risk levels.
Adequacy of preparedness frameworks: Whether capability evaluations used by frontier labs to gate deployment decisions are empirically sufficient to detect dangerous capabilities before deployment is a subject of ongoing debate. Academic analysis of OpenAI's Preparedness Framework v2 (2025) found that the framework evaluates only a subset of AI risks and, under certain conditions, permits the CEO to deploy systems with capabilities associated with severe harm if safeguards are deemed adequate — though what constitutes adequate safeguards is not always specified.27
Risk Reduction Approaches
Technical Alignment Research
Developing methods to ensure AI systems pursue intended objectives, including work on interpretability, scalable oversight, and constitutional AI by organizations including Anthropic, OpenAI, and Google DeepMind.30
Capability Evaluation
Organizations like METR and ARC Evaluations develop methods to assess whether AI systems exhibit dangerous capabilities such as scheming or autonomous operation.31 Apollo Research has conducted evaluations of frontier models including o1, o3, and o4-mini, documenting the presence and mitigation of in-context scheming behaviors.15
Governance and Policy
Research by Future of Humanity Institute, CAIS, and others on regulatory frameworks, international agreements, and institutional designs to manage transformative AI development.32
Frontier Lab Risk Frameworks
A significant institutional development since 2023 has been the adoption of formal risk frameworks by frontier AI laboratories. These frameworks attempt to operationalize safety governance by defining thresholds above which deployment is restricted or halted.
OpenAI's Preparedness Framework was first published in beta form in late 2023 and updated to version 2 in April 2025.33 The original framework defined four risk tiers (Low, Medium, High, Critical) across four tracked categories: CBRN threats, cybersecurity, individualized persuasion, and Autonomous Replication (ARA). Deployment rules in v1 stated that models scoring Critical after mitigations cannot be deployed, and models scoring High before mitigations require hardened security against model weight exfiltration.34
Version 2 (April 2025) made several changes: it streamlined the actionable thresholds from four tiers to two (High and Critical), removed nuclear and radiological risks from tracked categories (reclassifying them as research categories), and added a dedicated Safeguards Reports requirement alongside Capabilities Reports.33 V2 defines a "catastrophic risk" as one that "could result in hundreds of billions of dollars in economic damage or severe harm/death to many individuals."34 The tracked categories in v2 are: Biological and Chemical capabilities, Cybersecurity capabilities, and AI Self-improvement capabilities.
The framework specifies that the Safety Advisory Group (SAG), a cross-functional internal group, reviews capability and safeguards reports and makes recommendations to company leadership, which retains final decision-making authority. The Board of Directors exercises oversight.35
Analysts have identified several limitations. Zvi Mowshowitz, writing in May 2025, characterized v2 as "a rather large retreat on the commitments" relative to v1, noting that v2 explicitly permits deployment of models that would rate High risk without mitigations, provided adequate safeguards exist, and includes a competitive carve-out allowing adjustment of safety requirements if a rival deploys a comparably risky model without protections.9 A September 2025 academic paper using affordance theory found that the framework "encourages deployment of systems with 'Medium' capabilities for what OpenAI defines as severe harm" and "allows OpenAI's CEO to deploy even more dangerous capabilities" under specified conditions.27
Anthropic's Responsible Scaling Policy (RSP) provides a comparable framework: it defines Autonomy Levels (AL) tied to capability thresholds, with commitments about safety mitigations required before advancing to each level. The RSP has been updated multiple times since its initial publication.
Google DeepMind publishes safety evaluation documentation alongside major model releases, including assessments of dangerous capability thresholds.
All three frameworks share structural features: capability evaluations gating deployment, tiered risk levels, and internal governance bodies reviewing safety recommendations. A persistent question across all three is whether capability evaluations are empirically sufficient to detect dangerous capabilities before harm occurs, given that evaluation methodologies are developed alongside — and sometimes after — the capabilities they are meant to detect.
Institutional Safety Governance at Frontier Labs
Following the November 2023 governance crisis at OpenAI — during which the board temporarily removed CEO Sam Altman, citing in part concerns about safety processes, before reinstating him days later — OpenAI reorganized several safety governance structures.36
In May 2024, OpenAI formed a Safety and Security Committee (SSC) as a board-level oversight body, chaired by Zico Kolter. The SSC conducted a 90-day review of safety and security processes and was given authority to delay model releases until safety concerns are addressed. The SSC's recommendations include establishing independent governance for safety and security, enhancing security measures, and unifying safety frameworks.36 A former OpenAI board member stated publicly that Altman had given the board "inaccurate information about the small number of formal safety processes" on multiple occasions.37
In October 2025, OpenAI completed a transition from a capped-profit structure to a for-profit Public Benefit Corporation (PBC), valued at approximately $500 billion. Under the restructured governance, a nonprofit OpenAI Foundation retains the power to appoint all board members of the for-profit entity, and a safety and security commission holds veto authority over model releases.38
OpenAI's Model Spec, first published in May 2024 and updated subsequently, serves as a governance instrument specifying behavioral constraints trained into deployed models.39 The Model Spec establishes a priority hierarchy: broadly safe > broadly ethical > adherent to OpenAI's principles > genuinely helpful. It defines "hardcoded" behaviors that remain constant regardless of operator or user instructions — including prohibitions on providing uplift for CBRN weapons and on taking actions that undermine human oversight of AI — and "softcoded" behaviors that can be adjusted within policy limits. Root-level instructions in the Model Spec cannot be overridden by system prompts or other messages.40 The Model Spec explicitly states that parts of it "consist of rules aimed at minimizing catastrophic risks, though not all risks from AI can be mitigated through model behavior alone."40
The Superalignment Initiative and Its Dissolution
In July 2023, OpenAI announced the Superalignment team with a stated commitment to dedicate 20% of its then-available computing resources over four years to developing alignment techniques capable of supervising superintelligent AI systems. Co-leads were Ilya Sutskever and Jan Leike.41
Multiple sources reported that the 20% compute commitment was not fulfilled; the team received far less compute via regular quarterly allocation budgets, and no clear metrics were established for calculating what 20% meant in practice.41 Both Sutskever and Leike departed OpenAI in May 2024. Leike published a public statement saying that OpenAI's "safety culture and processes have taken a backseat to shiny products."42 OpenAI formally disbanded the Superalignment team in May 2024, approximately one year after its founding, with members reassigned to other teams.42
Subsequently, in October 2024, OpenAI also disbanded its separate "AGI Race Readiness" team. Miles Brundage, senior advisor for AGI Readiness, departed at that time.43 In July 2024, Aleksander Madry — who had led the Preparedness team — was reassigned from that role to a position focused on AI reasoning.43
OpenAI's current leadership — Chief Scientist Jakub Pachocki and Chief Research Officer Mark Chen — have stated that alignment has been integrated into OpenAI's core research operations rather than remaining the domain of a separate dedicated team.16
Field-Building
Organizations including 80,000 Hours, Centre for Effective Altruism, and LessWrong work to direct research talent and resources toward AI safety. These organizations operate within the effective altruism philosophical tradition, which is itself a subject of ongoing debate regarding the appropriate prioritization of long-term speculative risks against near-term harms.44
AGI Definitions and Governance Implications
How frontier labs define AGI has direct implications for when different governance obligations trigger. OpenAI defines AGI as "a highly autonomous system that outperforms humans at most economically valuable work."45 Under original 2019 partnership terms with Microsoft, OpenAI's unilateral declaration of AGI achievement would have terminated Microsoft's access to subsequent OpenAI models. Sam Altman has acknowledged that the definition of AGI "shifts depending on who is asked."45
Under revised terms negotiated in 2025, OpenAI can no longer unilaterally declare AGI achievement — verification requires an independent expert panel. Microsoft's IP rights were extended through 2032, applying even to models developed after AGI is reached, and both companies can now pursue AGI independently.46 These renegotiations illustrate how the definition and timing of AGI serves not only as a technical milestone but as a governance trigger with substantial commercial and safety implications.
Historical Context
Concern about existential risks from AI emerged from several intellectual traditions:
Early warnings about superintelligence appeared in I.J. Good's 1965 paper on intelligence explosions and subsequent work by Vernor Vinge in the 1990s. Nick Bostrom's formalization of existential risk concepts in the early 2000s provided analytical frameworks, while Eliezer Yudkowsky's writings on LessWrong beginning in 2006 developed detailed technical arguments.47
The field gained mainstream attention following the 2014 publication of Bostrom's Superintelligence and public statements by Stephen Hawking, Elon Musk, and Bill Gates about potential AI risks. The 2023 CAIS statement on AI risk, signed by numerous AI researchers including Geoffrey Hinton and Dario Amodei, attracted attention from researchers across the field and from policymakers.48
Funding for existential risk reduction grew from negligible amounts before 2010. By 2023, annual funding reached hundreds of millions of dollars, concentrated among a small number of donors including Open Philanthropy and individual philanthropists associated with the effective altruism movement.49 Critics have noted that this concentration of funding in a small donor base raises questions about potential epistemic and political distortions in the research field — a debate that exists alongside, and independent of, questions about the underlying risk estimates.
The 2023–2025 period saw several parallel developments: the proliferation of formal safety frameworks at frontier labs (OpenAI's Preparedness Framework, Anthropic's Responsible Scaling Policy), the November 2023 OpenAI governance crisis and subsequent board restructuring, government-mandated safety evaluations in the United Kingdom and United States, and international commitments at the Bletchley Park AI Safety Summit (2023) and Seoul AI Safety Summit (2024) by frontier labs including OpenAI to share safety information and conduct pre-deployment evaluations for frontier models.
Relationship to Other Risks
Existential risk from AI intersects with other global catastrophic risks:
Nuclear risk: Some researchers explore whether AI systems might increase nuclear war probability through Autonomous Weapons, decision-making acceleration, or cyber vulnerabilities in command and control systems.50
Biological risk: AI capabilities for protein design and synthetic biology raise concerns about engineered pandemics. OpenAI's Preparedness Framework identifies biological capabilities as "by far the biggest of the four threats" among its tracked CBRN categories.9 These risks are typically classified as catastrophic rather than existential unless they threaten species survival.
Climate change: While climate change is generally not considered an existential risk to human survival, some researchers explore whether AI acceleration of climate impacts or geoengineering failures could create existential threats.51
Debates continue about relative risk prioritization and whether existential risk framing is appropriate for these domains.
Footnotes
-
Bostrom, Nick. "Existential Risks: Analyzing Human Extinction Scenarios and Related Hazards." Journal of Evolution and Technology, 2002. ↩
-
Bostrom, Nick. Superintelligence: Paths, Dangers, Strategies. Oxford University Press, 2014. ↩
-
Russell, Stuart. Human Compatible: Artificial Intelligence and the Problem of Control. Viking, 2019. ↩
-
Ord, Toby. The Precipice: Existential Risk and the Future of Humanity. Hachette Books, 2020. ↩
-
Russell, Stuart. Human Compatible, 2019. ↩
-
Yudkowsky, Eliezer. "Intelligence Explosion Microeconomics." Machine Intelligence Research Institute Technical Report, 2013. ↩
-
Christiano, Paul. "What Failure Looks Like." AI Alignment Forum, 2019. ↩
-
Dafoe, Allan. "AI Governance: A Research Agenda." Future of Humanity Institute, University of Oxford, 2018. ↩
-
Mowshowitz, Zvi. "OpenAI Preparedness Framework 2.0." Substack, May 2, 2025. https://thezvi.substack.com/p/openai-preparedness-framework-20 ↩ ↩2 ↩3
-
RD World Online. "OpenAI Framework: AI Now 'On the Cusp of Doing New Science'." April 16, 2025. https://www.rdworldonline.com/openai-framework-ai-now-on-the-cusp-of-doing-new-science/ ↩
-
Control AI News. "The Ultimate Risk: Recursive Self-Improvement." December 4, 2025. https://controlai.news/p/the-ultimate-risk-recursive-self ↩ ↩2
-
OpenAI. "AI Progress and Recommendations." 2024. https://openai.com/index/ai-progress-and-recommendations/ ↩
-
OpenAI. "OpenAI o1 System Card." September 2024. https://openai.com/index/openai-o1-system-card/ ↩ ↩2
-
OpenAI. "o3 and o4-mini System Card." April 16, 2025. https://cdn.openai.com/pdf/2221c875-02dc-4789-800b-e7758f3722c1/o3-and-o4-mini-system-card.pdf ↩
-
OpenAI and Apollo Research. "Detecting and Reducing Scheming in AI Models." September 2025. https://openai.com/index/detecting-and-reducing-scheming-in-ai-models/ ↩ ↩2
-
MIT Technology Review. "The Two People Shaping the Future of OpenAI's Research." July 31, 2025. https://www.technologyreview.com/2025/07/31/1120885/the-two-people-shaping-the-future-of-openais-research/ ↩ ↩2
-
Ord, Toby. The Precipice, 2020, p. 167. ↩
-
Stein-Perlman, Zachary, Benjamin Weinstein-Raun, and Katja Grace. "2022 Expert Survey on Progress in AI." AI Impacts, 2022. ↩
-
Bommasani, Rishi, et al. "The Economics of p(doom): Scenarios of Existential Risk and AGI." arXiv:2503.07341, March 2025. https://arxiv.org/pdf/2503.07341 ↩ ↩2 ↩3
-
Center for AI Safety. "Statement on AI Risk." 2023. https://www.safe.ai/statement-on-ai-risk ↩
-
Windows Central. "Sam Altman Claims AI Will Be Smart Enough to Prevent Existential Doom." December 6, 2024. https://www.windowscentral.com/software-apps/sam-altman-ai-smart-enough-to-prevent-existential-doom ↩
-
Brooks, Rodney. "I, Rodney Brooks, Am a Robot." IEEE Spectrum, 2008. ↩
-
LeCun, Yann. Various public statements and interviews, 2022–2024. ↩
-
Garfinkel, Ben. "How Sure Are We About This Whole Existential Risk Thing?" 80,000 Hours Podcast, 2018. ↩
-
Whittlestone, Jess, et al. "The Role and Limits of Principles in AI Ethics." AIES 2019 Proceedings, 2019. ↩
-
CNBC. "OpenAI Dissolves Superalignment AI Safety Team." May 17, 2024. https://www.cnbc.com/2024/05/17/openai-superalignment-sutskever-leike.html ↩
-
Academic authors. "The 2025 OpenAI Preparedness Framework Does Not Guarantee Any AI Risk Mitigation Practices." arXiv:2509.24394, September 2025. https://arxiv.org/abs/2509.24394 ↩ ↩2 ↩3
-
Epoch AI. "Compute Trends Across Three Eras of Machine Learning." 2022. ↩
-
Yudkowsky, Eliezer. "The Fast-Takeoff Hypothesis." LessWrong, various. ↩
-
Anthropic, OpenAI, Google DeepMind. Published safety research and documentation, 2022–2025. ↩
-
METR. Evaluation documentation, 2023–2024. ↩
-
Future of Humanity Institute. Various publications, 2014–2024. ↩
-
OpenAI. "Our Updated Preparedness Framework." April 15, 2025. https://openai.com/index/updating-our-preparedness-framework/ ↩ ↩2
-
OpenAI. "Preparedness Framework Version 2." April 15, 2025. https://cdn.openai.com/pdf/18a02b5d-6b67-4cec-ab64-68cdfbddebcd/preparedness-framework-v2.pdf ↩ ↩2
-
OpenAI. "OpenAI Safety Update." 2024. https://openai.com/index/openai-safety-update/ ↩
-
OpenAI. "OpenAI Board Forms Safety and Security Committee." May 2024. https://openai.com/index/openai-board-forms-safety-and-security-committee/ ↩ ↩2
-
OpenAI. "An Update on Our Safety & Security Practices." September 16, 2024. https://openai.com/index/update-on-safety-and-security-practices/ ↩
-
EA Forum. "The OpenAI Governance Transition: The History, What It Is, and What It Means." November 17, 2025. https://forum.effectivealtruism.org/posts/Tcy5HAg3d9LXDRGfq/the-openai-governance-transition-the-history-what-it-is-and-1 ↩
-
OpenAI. "Introducing the Model Spec." May 8, 2024. https://openai.com/index/introducing-the-model-spec/ ↩
-
OpenAI. "Model Spec." December 18, 2025. https://model-spec.openai.com/2025-12-18.html ↩ ↩2
-
Fortune. "OpenAI Promised 20% of Its Computing Power to Combat the Most Dangerous Kind of AI — But Never Delivered." May 21, 2024. https://fortune.com/2024/05/21/openai-superalignment-20-compute-commitment-never-fulfilled-sutskever-leike-altman-brockman-murati/ ↩ ↩2
-
CNBC. "OpenAI Dissolves Superalignment AI Safety Team." May 17, 2024. https://www.cnbc.com/2024/05/17/openai-superalignment-sutskever-leike.html ↩ ↩2
-
CNBC. "OpenAI Disbands Another Safety Team, as Head Advisor for 'AGI Readiness' Resigns." October 24, 2024. https://www.cnbc.com/2024/10/24/openai-miles-brundage-agi-readiness.html ↩ ↩2
-
80,000 Hours. Career guide and cause area research, 2023. ↩
-
TechRepublic. "Inside Microsoft's AGI Clause Fight With OpenAI." July 11, 2025. https://www.techrepublic.com/article/news-openai-agi-clause-threatens-microsoft-deal/ ↩ ↩2
-
Pure AI. "Microsoft, OpenAI Rewrite Partnership Rules Ahead of AGI Race." November 4, 2025. https://pureai.com/articles/2025/11/04/microsoft-openai-rewrite-partnership-rules-ahead-of-agi-race.aspx ↩
-
Good, I.J. "Speculations Concerning the First Ultraintelligent Machine." Advances in Computers, 1965. ↩
-
Center for AI Safety. "Statement on AI Risk." 2023. ↩
-
Open Philanthropy. Annual giving reports, 2020–2023. ↩
-
Geist, Edward Moore. "It's Already Too Late to Stop the AI Arms Race." Bulletin of the Atomic Scientists, 2016. ↩
-
Pamlin, Dennis, and Stuart Armstrong. "Global Challenges: 12 Risks That Threaten Human Civilisation." Global Challenges Foundation, 2015. ↩