AGI Development
Quick Assessment
| Dimension | Assessment | Evidence |
|---|---|---|
| Timeline Consensus | 2027-2031 median (50% probability) | Metaculus: 25% by 2027, 50% by 2031; 80,000 Hours expert synthesis |
| Industry Leader Predictions | 2026-2028 | Anthropic: "powerful AI" by late 2026/early 2027; OpenAI: "we know how to build AGI" |
| Capital Investment | $400-450B annually by 2026 | Deloitte: AI data center capex; McKinsey: $5-8T total by 2030 |
| Compute Scaling | 10^26-10^28 FLOPs projected | Epoch AI: compute trends; training runs reaching $1-10B |
| Safety-Capability Gap | 3-5 year research lag | Industry evaluations show alignment research trailing deployment capability |
| Geopolitical Dynamics | US maintains ≈5x compute advantage | CFR: China lags 3-6 months in models despite chip restrictions |
| Catastrophic Risk Concern | 25% per Amodei; 5% median (16% mean) in surveys | AI Impacts 2024: 2,778 researchers surveyed |
Key Links
| Source | Link |
|---|---|
| Official Website | blog.ktbyte.com |
| Wikipedia | en.wikipedia.org |
Overview
AGI development represents the global race to build artificial general intelligence—systems matching or exceeding human-level performance across all cognitive domains. Timeline forecasts have shortened dramatically: Metaculus forecasters now average a 25% probability of AGI by 2027 and 50% by 2031, down from a median of 50 years as recently as 2020. CEOs of major labs have made even more aggressive predictions, with Anthropic officially stating they expect "powerful AI systems" with Nobel Prize-winner level capabilities by early 2027.
Development is concentrated among 3-4 major labs investing $10-100B+ annually. This concentration creates significant coordination challenges and racing dynamics that could compromise safety research. The field has shifted from academic research to industrial competition, with OpenAI, Anthropic, DeepMind, and emerging players like xAI pursuing different technical approaches while facing similar resource constraints and timeline pressures.
AGI Development Dynamics
Diagram (loading…)
flowchart TD
subgraph Drivers["Key Drivers"]
COMPUTE[Compute Scaling<br/>10^26-28 FLOPs]
CAPITAL[Capital Investment<br/>400-450B annually]
TALENT[Talent Concentration<br/>Top researchers at labs]
end
subgraph Development["Development Race"]
LABS[Major Labs<br/>OpenAI, Anthropic, DeepMind, xAI]
COMPETITION[Racing Dynamics]
CHINA[US-China Competition]
end
subgraph Timelines["Timeline Estimates"]
SHORT[Short: 2025-2027<br/>15-25% probability]
MEDIUM[Medium: 2027-2030<br/>30-40% probability]
LONG[Long: 2030-2040<br/>25-35% probability]
end
subgraph Risks["Safety Concerns"]
GAP[Safety-Capability Gap<br/>3-5 year lag]
ALIGN[Alignment Research<br/>Underfunded relative to capabilities]
end
COMPUTE --> LABS
CAPITAL --> LABS
TALENT --> LABS
LABS --> COMPETITION
CHINA --> COMPETITION
COMPETITION --> SHORT
COMPETITION --> MEDIUM
COMPETITION --> LONG
LABS --> GAP
GAP --> ALIGN
style SHORT fill:#ffcccc
style MEDIUM fill:#ffffcc
style LONG fill:#ccffcc
style GAP fill:#ffcccc
style ALIGN fill:#ffccccAGI Timeline Forecasts
Timeline estimates have compressed dramatically over the past four years. The table below summarizes current forecasts from major sources:
Timeline Estimates Comparison
| Source | Definition Used | 10% Probability | 50% Probability | 90% Probability | Last Updated |
|---|---|---|---|---|---|
| Metaculus | Weakly general AI | 2025 | 2027 | 2032 | Dec 2024 |
| Metaculus | General AI (strict) | 2027 | 2031 | 2040 | Dec 2024 |
| AI Impacts Survey | High-level machine intelligence | 2027 | 2047 | 2100+ | Oct 2024 |
| Manifold Markets | AGI by definition | - | 47% by 2028 | - | Jan 2025 |
| Samotsvety Forecasters | AGI | - | ≈28% by 2030 | - | 2023 |
Sources: Metaculus AGI forecasts, 80,000 Hours AGI review, AI Impacts 2024 survey
Industry Leader Predictions
| Leader | Organization | Prediction | Statement Date |
|---|---|---|---|
| Sam Altman | OpenAI | AGI during 2025-2028; "we know how to build AGI" | Nov 2024 |
| Dario Amodei | Anthropic | Powerful AI (Nobel-level) by late 2026/early 2027 | Jan 2026 |
| Demis Hassabis | DeepMind | 50% chance of AGI by 2030; "maybe 5-10 years, possibly lower end" | Mar 2025 |
| Jensen Huang | NVIDIA | AI matching humans on any test by 2029 | Mar 2024 |
| Elon Musk | xAI | AGI likely by 2026 | 2024 |
Note: Anthropic is the only major lab with official AGI timelines in policy documents, stating in March 2025: "We expect powerful AI systems will emerge in late 2026 or early 2027."
Timeline Trend Analysis
The most striking feature of AGI forecasts is how rapidly they have shortened:
| Year | Metaculus Median AGI | Change |
|---|---|---|
| 2020 | ≈2070 (50 years) | - |
| 2022 | ≈2050 (28 years) | -22 years |
| 2024 | 2031 (7 years) | -19 years |
| 2025 | 2029-2031 | -2 years |
The AI Impacts survey found that the median estimate for achieving "high-level machine intelligence" shortened by 13 years between 2022 and 2023 alone.
AGI Development Assessment
| Factor | Current State | 2025-2027 Trajectory | Key Uncertainty |
|---|---|---|---|
| Timeline Consensus | 2027-2031 median | Rapidly narrowing | Compute scaling limits |
| Resource Requirements | $10-100B+ per lab | Exponential growth required | Hardware availability |
| Technical Approach | Scaling + architecture | Diversification emerging | Which paradigms succeed |
| Geopolitical Factors | US-China competition | Intensifying restrictions | Export control impacts |
| Safety Integration | Limited, post-hoc | Pressure for alignment | Research-development gap |
Source: Metaculus AGI forecasts↗🔗 web★★★☆☆MetaculusMetaculus AGI forecastsA crowd-sourced probabilistic forecast on AGI timelines with well-defined resolution criteria; useful as a real-time community sentiment indicator on near-term general AI capabilities.A Metaculus forecasting question asking when the first 'weakly general AI' system will be publicly announced, with a current community median estimate of April 2028. The questio...capabilitiesagievaluationforecasting+2Source ↗, expert surveys
Major Development Approaches
Scaling-First Strategy
Most leading labs pursue computational scaling as the primary path to AGI:
| Lab | Approach | Investment Scale | Key Innovation |
|---|---|---|---|
| OpenAI | Large-scale transformer scaling | $13B+ (Microsoft) | GPT architecture optimization |
| Anthropic | Constitutional AI + scaling | $7B+ (Amazon/Google) | Safety-focused training |
| DeepMind | Multi-modal scaling | $2B+ (Alphabet) | Gemini unified architecture |
| xAI | Rapid scaling + real-time data | $6B+ (Series B) | Twitter integration advantage |
Sources: OpenAI funding announcements↗🔗 web★★★★☆OpenAIOpenAI funding announcementsThis is OpenAI's main news index page; useful for tracking announcements but not a focused AI safety resource. Individual linked articles (e.g., on prompt injection, instruction hierarchy) may be more directly relevant to safety research.The OpenAI news blog serves as the central hub for company announcements, research publications, product launches, and safety updates. Recent posts cover model releases, securit...capabilitiesdeploymenttechnical-safetyai-safety+3Source ↗, Anthropic Series C↗🔗 web★★★★☆AnthropicAnthropic Series C Funding AnnouncementThis is a corporate funding announcement from Anthropic, relevant for understanding the organizational and financial context of a major safety-focused AI lab; limited direct technical or research content.Anthropic announced its Series C funding round, raising significant capital to advance AI safety research and develop safer AI systems. The announcement reflects investor confid...ai-safetycapabilitiesgovernancedeployment+1Source ↗, DeepMind reports↗🔗 web★★★★☆Google DeepMindGoogle DeepMind Official HomepageGoogle DeepMind is a major frontier AI lab whose research and policies are highly relevant to AI safety; this homepage provides entry point to their publications, safety frameworks, and organizational positions on AI risk.Google DeepMind is a leading AI research laboratory combining the former DeepMind and Google Brain teams, focused on developing advanced AI systems and conducting research acros...capabilitiesai-safetygovernancealignment+4Source ↗
Resource Requirements Trajectory
Current AGI development demands exponentially increasing resources:
| Resource Type | 2024 Scale | 2026 Projection | 2028+ Requirements |
|---|---|---|---|
| Training Compute | 10^25 FLOPs | 10^26-10^27 FLOPs | 10^28+ FLOPs |
| Training Cost | $100M-1B | $1-10B | $10-100B |
| Electricity | 50-100 MW | 500-1000 MW | 1-10 GW |
| Skilled Researchers | 1000-3000 | 5000-10000 | 10000+ |
| H100 Equivalent GPUs | 100K+ | 1M+ | 10M+ |
Sources: Epoch AI compute trends↗🔗 web★★★★☆Epoch AIEpoch AI - AI Research and Forecasting OrganizationEpoch AI is a key reference organization for empirical data on AI scaling trends; their compute and training run databases are widely cited in AI safety and governance discussions.Epoch AI is a research organization focused on investigating and forecasting trends in artificial intelligence, particularly around compute, training data, and algorithmic progr...capabilitiescomputegovernancepolicy+4Source ↗, RAND Corporation analysis↗🔗 web★★★★☆RAND CorporationThe AI and Biological Weapons ThreatA 2023 RAND empirical study directly relevant to catastrophic risk from AI misuse; provides early evidence on LLM dual-use risks in bioweapons contexts, informing debates about frontier model deployment safeguards and biosecurity policy.This RAND Corporation report examines the misuse risks of large language models (LLMs) in biological weapons development through a red-team methodology. Preliminary findings sho...biosecurityred-teamingcapabilitiesexistential-risk+6Source ↗
Global AI Infrastructure Investment
The capital requirements for AGI development are unprecedented. According to McKinsey, companies will need to invest $5.2-7.9 trillion into AI data centers by 2030.
| Category | 2025 | 2026 | 2028 | Source |
|---|---|---|---|---|
| AI Data Center Capex | $250-300B | $400-450B | $1T | Deloitte 2026 Predictions |
| AI Chip Spending | $150-200B | $250-300B | $400B+ | Industry analysis |
| Stargate Project | $100B (Phase 1) | Ongoing | $500B total | TechCrunch |
| OpenAI Cloud Commitments | Ongoing | $50B/year | $60B/year | Azure + Oracle deals |
Training costs have declined dramatically—ARK Investment reports costs drop roughly 10x annually, ~50x faster than Moore's Law. DeepSeek's V3 achieved 18x training cost reduction vs GPT-4o.
Key Capability Thresholds
AGI development targets specific capability milestones that indicate progress toward human-level performance:
Current Capability Gaps
- Long-horizon planning: Limited to hours/days vs. human years/decades
- Scientific research: Narrow domain assistance vs. autonomous discovery
- Real-world agentic behavior: Supervised task execution vs. autonomous goal pursuit
- Self-improvement: Assisted optimization vs. recursive enhancement
2025-2027 Expected Milestones
- PhD-level performance in most academic domains
- Autonomous software engineering at human expert level
- Multi-modal reasoning approaching human performance
- Planning horizons extending to weeks/months
Geopolitical Development Landscape
AGI development increasingly shaped by international competition and regulatory responses:
US-China Competition
| Factor | US Position | China Position | Impact |
|---|---|---|---|
| Leading Labs | OpenAI, Anthropic, DeepMind | Baidu, Alibaba, ByteDance | Technology fragmentation |
| Compute Access | H100 restrictions on China | Domestic chip development | Capability gaps emerging |
| Talent Pool | Immigration restrictions growing | Domestic talent retention | Brain drain dynamics |
| Investment | Private + government funding | State-directed investment | Different risk tolerances |
Sources: CNAS reports↗🔗 web★★★★☆CNASCenter for a New American Security (CNAS) - HomepageCNAS is a mainstream national security think tank; relevant to AI safety primarily through its Technology & National Security program covering AI governance and defense AI policy, but not an AI safety-focused organization.CNAS is a Washington D.C.-based national security think tank publishing research on defense, technology policy, economic security, and AI governance. Its Technology & National S...governancepolicyai-safetycapabilities+2Source ↗, Georgetown CSET analysis↗🔗 web★★★★☆CSET GeorgetownCSET: AI Market DynamicsCSET is a prominent DC-based think tank whose research on AI governance, compute policy, and geopolitical competition is frequently cited in AI safety and policy discussions; this is their institutional homepage.CSET (Center for Security and Emerging Technology) at Georgetown University is a policy research organization focused on the security implications of emerging technologies, part...governancepolicyai-safetycoordination+2Source ↗
Safety Research Integration
Critical gap exists between AGI development timelines and safety research readiness:
Current Safety-Capability Gap
| Domain | Development State | Safety Research State | Gap Assessment |
|---|---|---|---|
| Alignment | Production systems | Early research | 3-5 year lag |
| Interpretability | Limited deployment | Proof-of-concept | 5+ year lag |
| Robustness | Basic red-teaming | Formal verification research | 2-3 year lag |
| Evaluation | Industry benchmarks | Academic proposals | 1-2 year lag |
Industry Safety Initiatives
- OpenAI: Superalignment team (dissolved 2024), safety-by-default claims
- Anthropic: Constitutional AI, AI Safety via Debate research
- DeepMind: Scalable oversight, cooperative AI research
- Industry-wide: Responsible scaling policies, voluntary commitments
Current State & Development Trajectory
2024 Status
- GPT-4 level models becoming commoditized
- Multimodal capabilities reaching practical deployment
- Compute costs limiting smaller players
- Regulatory frameworks emerging globally
2025-2027 Projections
- 100x compute scaling attempts by major labs
- Emergence of autonomous AI researchers/engineers
- Potential capability discontinuities from architectural breakthroughs
- Increased government involvement in development oversight
Key Development Bottlenecks
- Compute hardware: H100/H200 supply constraints, next-gen chip delays
- Energy infrastructure: Data center power requirements exceeding grid capacity
- Talent acquisition: Competition for ML researchers driving salary inflation
- Data quality: Exhaustion of high-quality training data sources
Scenario Analysis
The wide range of AGI timeline estimates reflects genuine uncertainty. The following scenarios capture the range of plausible outcomes:
AGI Arrival Scenarios
| Scenario | Timeline | Probability | Key Assumptions | Implications |
|---|---|---|---|---|
| Rapid Takeoff | 2025-2027 | 15-25% | Scaling continues; breakthrough architecture; recursive self-improvement | Minimal time for governance; safety research severely underprepared |
| Accelerated Development | 2027-2030 | 30-40% | Current trends continue; major labs achieve stated goals | 2-4 years for policy response; industry-led safety measures |
| Gradual Progress | 2030-2040 | 25-35% | Scaling hits diminishing returns; algorithmic breakthroughs needed | Adequate time for safety research; international coordination possible |
| Extended Timeline | 2040+ | 10-20% | Fundamental barriers emerge; AGI harder than expected | Safety research can mature; risk of complacency |
Probabilities are rough estimates based on synthesizing Metaculus forecasts, expert surveys, and industry predictions. Significant uncertainty remains.
Scenario Implications for Safety
| Scenario | Safety Research Readiness | Governance Preparedness | Risk Level |
|---|---|---|---|
| Rapid Takeoff | Severely underprepared | No frameworks in place | Very High |
| Accelerated Development | Partially prepared; core problems unsolved | Basic frameworks emerging | High |
| Gradual Progress | Adequate research time; may achieve interpretability | Comprehensive governance possible | Medium |
| Extended Timeline | Full research maturity possible | Global coordination achieved | Lower |
The critical insight is that the probability-weighted risk is dominated by shorter timelines, even if they are less likely, because the consequences of being underprepared are severe and irreversible.
Key Uncertainties & Expert Disagreements
AI Impacts 2024 Survey Findings
The largest survey of AI researchers to date (2,778 respondents who published in top-tier AI venues) provides important calibration:
| Finding | Value | Notes |
|---|---|---|
| 50% probability of HLMI | By 2047 | 13 years earlier than 2022 survey |
| 10% probability of HLMI | By 2027 | Near-term risk not negligible |
| Median extinction risk | 5% | Mean: 16% (skewed by high estimates) |
| "Substantial concern" warranted | 68% agree | About AI-related catastrophic risks |
The survey also found researchers gave at least 50% probability that AI would achieve specific milestones by 2028, including: autonomously constructing payment processing sites, creating indistinguishable music, and fine-tuning LLMs without human assistance.
Timeline Uncertainty Factors
- Scaling law continuation: Will current trends plateau or breakthrough?
- Algorithmic breakthroughs: Novel architectures vs. incremental improvements
- Hardware advances: Impact of next-generation accelerators
- Data limitations: Quality vs. quantity tradeoffs in training
Strategic Disagreements
| Position | Advocates | Key Argument | Risk Assessment |
|---|---|---|---|
| Speed prioritization | Some industry leaders | First-mover advantages crucial | Higher accident risk |
| Safety prioritization | Safety researchers | Alignment must precede capability | Competitive disadvantage |
| International cooperation | Policy experts | Coordination prevents racing | Enforcement challenges |
| Open development | Academic researchers | Transparency improves safety | Proliferation risks |
Critical Research Questions
- Can current safety techniques scale to AGI-level capabilities?
- Will AGI development be gradual or discontinuous?
- How will geopolitical tensions affect development trajectories?
- Can effective governance emerge before critical capabilities?
Timeline & Warning Signs
Pre-AGI Indicators (2025-2028)
- Autonomous coding: AI systems independently developing software
- Scientific breakthroughs: AI-driven research discoveries
- Economic impact: Significant job displacement in cognitive work
- Situational awareness: Systems understanding their training and deployment
Critical Decision Points
- Compute threshold policies: When scaling restrictions activate
- International agreements: Multilateral development frameworks
- Safety standard adoption: Industry-wide alignment protocols
- Open vs. closed development: Transparency vs. security tradeoffs
Sources & Resources
Timeline Forecasting Resources
| Source | Type | URL | Key Contribution |
|---|---|---|---|
| Metaculus AGI Questions | Prediction market | metaculus.com | Crowd forecasts with 25% by 2027, 50% by 2031 |
| 80,000 Hours AGI Review | Expert synthesis | 80000hours.org | Comprehensive review of expert forecasts |
| AI Impacts Survey | Academic survey | arxiv.org/abs/2401.02843 | 2,778 researchers surveyed; 50% HLMI by 2047 |
| AGI Timelines Dashboard | Aggregator | agi.goodheartlabs.com | Real-time aggregation of prediction markets |
| Epoch AI Scaling Analysis | Technical research | epoch.ai | Compute scaling projections through 2030 |
Research Organizations
| Organization | Focus | Key Publications |
|---|---|---|
| Epoch AI↗🔗 web★★★★☆Epoch AIEpoch AI - AI Research and Forecasting OrganizationEpoch AI is a key reference organization for empirical data on AI scaling trends; their compute and training run databases are widely cited in AI safety and governance discussions.Epoch AI is a research organization focused on investigating and forecasting trends in artificial intelligence, particularly around compute, training data, and algorithmic progr...capabilitiescomputegovernancepolicy+4Source ↗ | Compute trends, forecasting | Parameter counts, compute analysis |
| RAND Corporation↗🔗 web★★★★☆RAND CorporationRAND Provides Objective Research Services and Public Policy AnalysisRAND Corporation's homepage serves as an entry point to a large body of policy-relevant research on AI governance, national security, and emerging technology risks, useful as a reference for policymakers and researchers in the AI safety space.RAND Corporation is a nonprofit research organization providing objective analysis and policy recommendations across a wide range of topics including national security, technolo...governancepolicyai-safetycybersecurity+4Source ↗ | Policy analysis | AGI governance frameworks |
| Georgetown CSET↗🔗 web★★★★☆CSET GeorgetownCSET: AI Market DynamicsCSET is a prominent DC-based think tank whose research on AI governance, compute policy, and geopolitical competition is frequently cited in AI safety and policy discussions; this is their institutional homepage.CSET (Center for Security and Emerging Technology) at Georgetown University is a policy research organization focused on the security implications of emerging technologies, part...governancepolicyai-safetycoordination+2Source ↗ | Technology competition | US-China AI competition analysis |
| Future of Humanity Institute↗🔗 web★★★★☆Future of Humanity Institute**Future of Humanity Institute**FHI was a pioneering institution in AI safety and existential risk; this archived homepage is useful for historical context and understanding the institutional origins of the field, though the site is no longer actively updated following its April 2024 closure.The official website of the Future of Humanity Institute (FHI), an Oxford University research center that was foundational in establishing the fields of existential risk researc...ai-safetyexistential-riskalignmentgovernance+3Source ↗ | Existential risk | AGI timeline surveys |
Industry Analysis
| Source | Coverage | Key Insights |
|---|---|---|
| Metaculus↗🔗 web★★★☆☆MetaculusMetaculus Forecasting PlatformMetaculus is widely used in the AI safety and EA communities as a reference for probabilistic forecasts on AI timelines and risk-relevant events; useful for grounding strategic discussions in calibrated uncertainty estimates.Metaculus is a collaborative online forecasting platform where users make probabilistic predictions on future events across domains including AI development, biosecurity, and gl...existential-riskai-safetygovernanceevaluation+4Source ↗ | Crowd forecasting | AGI timeline predictions |
| Our World in Data↗🔗 web★★★★☆Our World in DataArtificial Intelligence - Our World in DataA general-audience reference site useful for grounding AI safety discussions in empirical data about AI progress and adoption; best used as a starting point or citation source rather than a deep technical resource.Our World in Data provides an empirically grounded, data-driven overview of artificial intelligence development, tracking metrics around AI capabilities, adoption, and societal ...capabilitiesgovernancepolicydeployment+2Source ↗ | Capability trends | Historical scaling patterns |
| AI Index↗🔗 webStanford HAI AI Index ReportA key annual reference for AI safety researchers tracking capability trends, policy developments, and broader AI ecosystem dynamics; useful for situating safety concerns within the wider landscape of AI progress.The Stanford HAI AI Index is an annual report providing comprehensive, data-driven analysis of global AI developments spanning research output, technical capabilities, economic ...governancepolicycapabilitiesevaluation+4Source ↗ | Industry metrics | Investment, capability benchmarks |
| Anthropic Constitutional AI↗📄 paper★★★★☆AnthropicAnthropic's Work on AI SafetyThis is Anthropic's research landing page, useful as a starting point for discovering their published work on safety and alignment, but not a standalone paper or primary source in itself.Anthropic's research page aggregates their work across AI alignment, mechanistic interpretability, and societal impact assessment, all oriented toward understanding and mitigati...ai-safetyalignmentinterpretabilitytechnical-safety+4Source ↗ | Safety-focused development | Alternative development approaches |
Government Resources
| Agency | Role | Key Reports |
|---|---|---|
| NIST AI Risk Management↗🏛️ government★★★★★NISTNIST AI Risk Management FrameworkThe NIST AI RMF is a widely referenced U.S. government standard for AI risk governance, frequently cited in policy discussions and used by organizations building internal AI safety and compliance programs; relevant to AI safety researchers tracking institutional governance approaches.The NIST AI RMF is a voluntary, consensus-driven framework released in January 2023 to help organizations identify, assess, and manage risks associated with AI systems while pro...governancepolicyai-safetydeployment+4Source ↗ | Standards development | AI risk frameworks |
| UK AI Safety Institute | Safety evaluation | AGI evaluation protocols |
| US AI Safety Institute | Research coordination | Safety research priorities |
| EU AI Office↗🔗 web★★★★☆European UnionEU AI Office - European CommissionThe EU AI Office is a key regulatory institution for AI safety practitioners and developers operating in Europe; its mandates and guidelines directly shape how frontier AI models must be evaluated and deployed under the EU AI Act framework.The EU AI Office is the European Commission's central body responsible for overseeing and implementing the EU AI Act, particularly for general-purpose AI models. It coordinates ...governancepolicyai-safetydeployment+3Source ↗ | Regulatory oversight | AI Act implementation |
References
RAND Corporation is a nonprofit research organization providing objective analysis and policy recommendations across a wide range of topics including national security, technology, governance, and emerging risks. It produces influential studies on AI policy, cybersecurity, and global governance challenges. RAND's work is frequently cited by governments and policymakers worldwide.
Google DeepMind is a leading AI research laboratory combining the former DeepMind and Google Brain teams, focused on developing advanced AI systems and conducting research across capabilities, safety, and applications. The organization is one of the most influential labs in AI development, working on frontier models including Gemini and publishing widely-cited safety and capabilities research.
Epoch AI is a research organization focused on investigating and forecasting trends in artificial intelligence, particularly around compute, training data, and algorithmic progress. They produce empirical analyses and datasets to inform understanding of AI development trajectories and support better decision-making in AI governance and safety.
The official website of the Future of Humanity Institute (FHI), an Oxford University research center that was foundational in establishing the fields of existential risk research and AI safety. FHI closed on 16 April 2024 after approximately two decades of influential work. The site now serves as an archived record of the institution's history, research agenda, and legacy.
Our World in Data provides an empirically grounded, data-driven overview of artificial intelligence development, tracking metrics around AI capabilities, adoption, and societal impact. The resource aggregates research and statistics to help general audiences understand AI's trajectory and implications. It serves as a reference hub for understanding AI progress in broader economic and social context.
The OpenAI news blog serves as the central hub for company announcements, research publications, product launches, and safety updates. Recent posts cover model releases, security research, AI agent development, and safety initiatives across multiple domains.
The Stanford HAI AI Index is an annual report providing comprehensive, data-driven analysis of global AI developments spanning research output, technical capabilities, economic impact, policy, and societal effects. It serves as a widely cited reference for policymakers, researchers, and the public seeking objective benchmarks on AI progress. The report tracks trends over time, enabling longitudinal analysis of AI's trajectory.
The NIST AI RMF is a voluntary, consensus-driven framework released in January 2023 to help organizations identify, assess, and manage risks associated with AI systems while promoting trustworthiness across design, development, deployment, and evaluation. It provides structured guidance organized around core functions and is accompanied by a Playbook, Roadmap, and a Generative AI Profile (2024) addressing risks specific to generative AI systems.
CNAS is a Washington D.C.-based national security think tank publishing research on defense, technology policy, economic security, and AI governance. Its Technology & National Security program produces policy-relevant work on AI, cybersecurity, and emerging technologies with implications for AI safety and governance.
A Metaculus forecasting question asking when the first 'weakly general AI' system will be publicly announced, with a current community median estimate of April 2028. The question defines precise resolution criteria including passing a Turing test variant, 90%+ on Winograd Schema, 75th percentile SAT math, and mastering Montezuma's Revenge, all within a single unified system.
This RAND Corporation report examines the misuse risks of large language models (LLMs) in biological weapons development through a red-team methodology. Preliminary findings show that while LLMs haven't provided explicit weapon-creation instructions, they do offer guidance useful for planning biological attacks, including agent selection and acquisition strategies. The authors caution that AI's rapid advancement may outpace regulatory oversight, closing historical information gaps that previously hindered bioweapon development.
Anthropic announced its Series C funding round, raising significant capital to advance AI safety research and develop safer AI systems. The announcement reflects investor confidence in Anthropic's safety-focused approach to building large language models and reinforces the company's mission to ensure AI systems are safe, beneficial, and understandable.
Metaculus is a collaborative online forecasting platform where users make probabilistic predictions on future events across domains including AI development, biosecurity, and global catastrophic risks. It aggregates crowd wisdom and expert forecasts to produce calibrated probability estimates on complex questions relevant to long-term planning and existential risk assessment.
CSET (Center for Security and Emerging Technology) at Georgetown University is a policy research organization focused on the security implications of emerging technologies, particularly AI. It produces research on AI policy, workforce, geopolitics, and governance. The content could not be fully extracted, limiting detailed analysis.
The EU AI Office is the European Commission's central body responsible for overseeing and implementing the EU AI Act, particularly for general-purpose AI models. It coordinates AI governance across member states, enforces compliance with AI safety requirements, and supports the development of AI standards and testing methodologies.
Anthropic's research page aggregates their work across AI alignment, mechanistic interpretability, and societal impact assessment, all oriented toward understanding and mitigating risks from increasingly capable AI systems. It serves as a central hub for their published findings and ongoing safety-focused investigations.
A Metaculus forecasting question asking community members to predict when the first generally capable AI system will be created and publicly announced. The question aggregates probabilistic forecasts from many forecasters to produce a community estimate on AGI arrival timing.
A comprehensive synthesis by 80,000 Hours reviewing expert predictions on AGI timelines from multiple groups including AI lab leaders, researchers, and forecasters. The review finds a notable convergence toward shorter timelines, with many estimates suggesting AGI could arrive before 2030. Different expert communities that previously disagreed are now showing increasingly similar estimates.
Anthropic CEO Dario Amodei presents an optimistic vision of what a world with powerful AI could look like if development goes well, covering transformative potential in medicine, biology, mental health, economic development, and governance. He argues that most people underestimate both the upside potential and the downside risks of advanced AI, and explains why Anthropic has historically focused more on risks than benefits despite holding genuinely positive expectations.
Epoch AI analyzes the key constraints and bottlenecks that could limit continued AI scaling through 2030, examining factors such as compute availability, energy infrastructure, data availability, and algorithmic progress. The analysis assesses whether current scaling trends in large language models and other AI systems can realistically be sustained over the next several years.
This CFR analysis examines the technological gap between Huawei's domestic AI chips and Nvidia's leading GPUs, arguing that China's semiconductor capabilities remain significantly behind and that US export controls are effectively constraining China's AI development. The piece assesses Huawei's progress in chip design and manufacturing while highlighting persistent bottlenecks in yields, software ecosystems, and advanced packaging.
A survey of 2,778 AI researchers from top-tier venues reveals significant shifts in timelines for AI capabilities and widespread concern about advanced AI risks. Researchers predict at least 50% probability of several AI milestones by 2028 (including autonomous site construction and LLM fine-tuning), and estimate a 50% chance of AI outperforming humans in all tasks by 2047—13 years earlier than predicted in 2022. While 68% believe good outcomes from superhuman AI are more likely than bad, substantial majorities express concern about extinction-level risks (38-51% give ≥10% probability) and other scenarios like misinformation and inequality. Notably, researchers broadly agree that AI risk research should be prioritized more, despite disagreement on whether faster or slower progress is preferable.
An interactive dashboard aggregating and visualizing AGI timeline forecasts from major prediction markets and forecasting platforms including Metaculus, Manifold Markets, and Kalshi. It displays median year predictions and probability distributions for milestones such as 'weakly general AI,' 'general AI,' and passing the Turing Test, allowing users to download underlying data.