Longterm Wiki
Updated 2026-01-30HistoryData
Page StatusContent
Edited 2 weeks ago2.8k words
91
QualityComprehensive
62
ImportanceUseful
14
Structure14/15
20126061%11%
Updated every 3 weeksDue in 7 days
Summary

Epoch AI provides empirical AI progress tracking showing training compute growing 4.4x annually (2010-2024), 300 trillion tokens of high-quality training data with exhaustion projected 2026-2032, and algorithmic efficiency doubling every 6-12 months. Their 3,200+ model database directly informs US Executive Order 10^26 FLOPs threshold and export controls. With $4.1M in 2025 Open Philanthropy funding and 34 staff, they've produced FrontierMath (testing advanced reasoning) and the Epoch Capabilities Index, serving as critical infrastructure for compute governance approaches.

Issues1
Links18 links could use <R> components

Epoch AI

Organization

Epoch AI

Epoch AI provides empirical AI progress tracking showing training compute growing 4.4x annually (2010-2024), 300 trillion tokens of high-quality training data with exhaustion projected 2026-2032, and algorithmic efficiency doubling every 6-12 months. Their 3,200+ model database directly informs US Executive Order 10^26 FLOPs threshold and export controls. With $4.1M in 2025 Open Philanthropy funding and 34 staff, they've produced FrontierMath (testing advanced reasoning) and the Epoch Capabilities Index, serving as critical infrastructure for compute governance approaches.

2.8k words
Organization

Epoch AI

Related
Policies
Compute Governance
Concepts
Transformative AIAI Scaling LawsAI TimelinesAI Training Data Constraints

Quick Assessment

DimensionAssessmentEvidence
Research OutputVery High36 Data Insights, 37 Gradient Updates, 3,200+ models tracked in 2025
Policy InfluenceHighCited in US Executive Order 14110; 10^26 FLOP threshold based on Epoch data
Data InfrastructureExceptionalLargest public ML model database; 3,200+ models from 1950-present
Funding StabilityStrong$1.1M Coefficient Giving grant (2025), additional $1.9M general support
Team ExpertiseStrong34 staff; director holds PhD in AI from University of Aberdeen
Methodological RigorHighMultiple forecasting methods with explicit uncertainty bounds
IndependenceMedium-HighNonprofit with EA/Coefficient Giving funding; no industry capture
Benchmark InnovationHighFrontierMath benchmark with 60+ mathematicians including Fields Medalist

Overview

Epoch AI is a research organization founded in 2022 that provides rigorous, data-driven empirical analysis and forecasting of AI progress. Based in San Francisco with 34 employees, their work serves as critical infrastructure for AI governance and timeline forecasting. Their public database tracks over 3,200 machine learning models from 1950 to present—the largest resource of its kind.

Their research documents three key scaling dynamics: training compute growing at 4.4x per year since 2010, high-quality training data stock of approximately 300 trillion tokens with exhaustion projected between 2026-2032, and algorithmic efficiency doubling every 6-12 months. As of June 2025, they have identified over 30 publicly announced AI models exceeding the 10^25 FLOP training compute threshold.

Unlike organizations developing AI capabilities or safety techniques directly, Epoch provides the empirical foundation that informs strategic decisions across the AI ecosystem. Their databases and forecasts are cited by policymakers designing compute governance frameworks, safety researchers planning research timelines, and AI labs benchmarking their progress against industry trends. In 2024, they were featured on The New York Times' "2024 Good Tech Awards" list.

Loading diagram...

Their most influential finding is the exponential growth in training compute for frontier models—with training costs growing 2-3x annually and projected to exceed $1 billion per model by 2027. This analysis has become foundational for understanding AI progress and informing governance approaches focused on compute as a key chokepoint.

2025 Research Impact

Output Category2025 MetricsKey ProductsPolicy Relevance
Data Insights36 publishedTraining compute decomposition, API cost trendsInforms compute governance thresholds
Gradient Updates37 publishedWeekly trend analysesReal-time capability tracking
Model Database3,200+ entriesLargest public ML model databaseFoundation for policy research
Benchmarks2 major releasesFrontierMath, Epoch Capabilities IndexStandardized capability measurement
AI Chip TrackingNew data explorer15M+ H100-equivalents tracked globallyExport control effectiveness analysis
Cost AnalysisMultiple reports10x API price drop (Apr 2023-Mar 2025)Market concentration analysis

Key Findings Summary

MetricCurrent ValueTrendSource
Training compute growth4.4x/year (2010-2024)Accelerating to 5x/year post-2020Epoch Trends
Frontier model training costsTens of millions USDProjected to exceed $1B by 2027Cost Analysis
Total AI compute capacity15M+ H100-equivalentsGrowing rapidlyEpoch AI Chip Sales
Models above 10^25 FLOP30+ (as of June 2025)≈2 announced monthly in 2024Model Database
High-quality text stock≈300 trillion tokensExhaustion projected 2026-2032Epoch Data Analysis
API cost reduction10x (Apr 2023-Mar 2025)Continuing decline2025 Impact Report

Organizational Risk Assessment

Risk CategoryAssessmentEvidenceTimelineTrajectory
Data BottleneckHighHigh-quality text ≈300T tokens, current usage accelerating2026-2032Worsening
Compute ScalingMedium4.4x annual growth potentially unsustainable; energy constraints emerging2030sStable
Governance LagHighPolicy development slower than tech progress; 1,000+ US state AI bills in 2025OngoingImproving
Forecasting AccuracyMediumWide uncertainty bounds; Epoch Capabilities Index improving precisionContinuousImproving

Key Research Areas

Compute Trends Analysis

Epoch's flagship research tracks computational resources used to train AI models, revealing exponential scaling patterns. Their Machine Learning Trends dashboard provides real-time tracking of these dynamics.

MetricCurrent TrendKey FindingPolicy Implication
Training Compute4.4x/year (2010-2024)5x/year for frontier models since 2020Compute governance thresholds need updating
Training Costs$50-100M for frontier modelsProjected to exceed $1B by 2027Market concentration accelerating
Hardware Capacity15M+ H100-equivalents globallyCompanies own 80% of AI supercomputersExport controls increasingly important
Total AI ComputeDoubling every 7 months3.3x annual growth since 2022Governance frameworks struggling to keep pace

Critical findings from Epoch's compute database:

  • Exponential growth faster than Moore's Law: While chip performance doubles every ~2 years, AI training compute grows 4.4x annually—driven primarily by larger training clusters, longer training runs, and improved hardware
  • Economic scaling: Training costs growing 2-3x annually, reaching $50-100M+ for frontier models; the largest AI supercomputers (like Anthropic's 750MW Indiana facility) cost billions to build
  • Concentration effects: Only 12 organizations have trained models above 10^25 FLOP; companies now own 80% of all AI supercomputers while government share has declined

Training Data Constraints

Epoch's "Will We Run Out of Data?" research revealed potential bottlenecks for continued AI scaling. Their analysis estimates the total stock of human-generated public text at approximately 300 trillion tokens.

Data TypeEstimated StockCurrent Usage RateExhaustion Timeline
High-quality text≈300 trillion tokensAccelerating2026-2032
All web text≈10^15 tokensIncreasingEarly 2030s
Image dataLarger but finiteGrowing rapidly2030s+
Video dataMassive but hard to useEarly stagesUnknown

Key implications:

  • Pressure for efficiency: Data constraints may force more efficient training methods; Epoch research shows algorithmic efficiency doubling every 6-12 months
  • Synthetic data boom: Investment in AI-generated training data accelerating to extend runway
  • Multimodal shift: Language models increasingly incorporating image/video data
  • Overtraining risk: If models are "intensely overtrained," high-quality text could be exhausted even earlier than 2026

Timeline Forecasting Methodology

Epoch employs multiple complementary approaches to estimate transformative AI timelines:

MethodCurrent Estimate RangeKey VariablesConfidence Level
Trend Extrapolation2030s-2040sCompute, data, algorithmsMedium
Biological Anchors2040s-2050sBrain computation estimatesLow
Benchmark Analysis2030s-2050sTask performance ratesMedium
Economic Modeling2035-2060sInvestment trends, ROILow

Impact on AI Safety and Governance

Policy Integration

Epoch's data directly informs major governance initiatives. Their research has been particularly influential in establishing compute thresholds for regulatory frameworks.

Policy AreaEpoch ContributionReal-World Impact
US AI Executive Order 1411010^26 FLOPs threshold analysisTraining run reporting requirements for frontier models
Export controlsH100/A100 performance data, chip sales trackingChip restriction implementation and effectiveness monitoring
UK AI Safety InstituteCapability benchmarking, FrontierMathModel evaluation frameworks
EU AI ActCompute-based GPAI thresholdsClassification of general-purpose AI systems
Compute governance researchDatabase infrastructure, threshold analysisAcademic and policy research foundation

FrontierMath Benchmark

Epoch's FrontierMath benchmark represents a significant contribution to AI evaluation infrastructure:

AspectDetailsSignificance
Problem Count350 original problems (300 in Levels 1-3, 50 in Level 4)Covers major branches of modern mathematics
Expert Collaboration60+ mathematicians, including 14 IMO gold medalists, 1 Fields MedalistHighest-quality benchmark construction
AI PerformanceLess than 2% of problems solved by leading modelsReveals substantial gap between AI and human mathematical capability
Tier 4 Commission50 research-level problems commissioned by OpenAITesting frontier reasoning capabilities
Version UpdatesToken budget increased 10x in November 2025Adapting to improved model inference

Research Community Influence

MetricEvidenceSource
Academic citations1,000+ citations across safety researchGoogle Scholar
Policy references50+ government documents cite EpochGovernment databases
Database usage10,000+ downloads of compute databaseEpoch analytics
Media coverageNYT "2024 Good Tech Awards" recognitionNew York Times
EA Forum engagementActive community discussion and feedbackEA Forum posts

Current State and Trajectory

2024 Developments

Database expansion:

  • Added 200+ new model entries to Parameter Database
  • Enhanced tracking of Chinese and European models
  • Improved cost estimation methodologies
  • Real-time updates for new releases

Research breakthroughs:

  • Refined algorithmic efficiency measurement showing 6-12 month doubling times
  • Updated data exhaustion projections with synthetic data considerations
  • New economic modeling of AI investment trends
  • Bioweapons AI uplift analysis

2025-2026 Projections

AreaExpected DevelopmentImpactSource
Model scaling10+ models above 10^26 FLOP by 2026Over 200 projected by 2030Epoch projections
Data bottleneckHigh-quality text exhaustion begins 2026-2032Synthetic data scaling acceleratesEpoch data analysis
Compute governanceStandardized international monitoring neededEnhanced export controlsEpoch policy research
Benchmark development2 new benchmarks in developmentImproved capability measurement2025 Impact Report
Capability acceleration15 points/year on ECI (up from 8 points pre-April 2024)Faster than historical trendEpoch Capabilities Index
Open-source thresholdFrontier open models may exceed 10^26 FLOP before 2026Challenges compute governance approachEpoch data insights

Key Uncertainties and Debates

Forecasting Limitations

UncertaintyImpact on EstimatesMitigation Strategy
Algorithmic breakthroughsCould accelerate timelines by yearsMultiple forecasting methods
Data efficiency improvementsMay extend scaling runwayConservative assumptions
Geopolitical disruptionCould fragment or accelerate developmentScenario planning
Hardware bottlenecksMay slow progress unexpectedlySupply chain analysis

Methodological Debates

Trend extrapolation reliability:

  • Optimists: Historical trends provide best available evidence for forecasting
  • Pessimists: Sharp left turns and discontinuities make extrapolation unreliable
  • Epoch position: Multiple methods with explicit uncertainty bounds

Information hazards:

  • Security concern: Publishing compute data aids adversaries in capability assessment
  • Racing dynamics: Timeline estimates may encourage competitive behavior
  • Transparency advocates: Public data essential for democratic governance

Forecasting Reliability Debate

Value of Empirical AI Forecasting

Essential Infrastructure

Epoch's data provides crucial foundation for rational planning. Timeline estimates inform urgency decisions. Compute tracking enables governance. Superior to pure speculation.

Proponents: Policy community, Many safety researchers, EA researchers
Confidence: high (4/5)
Useful but Limited

Valuable for trend identification but shouldn't drive strategy alone. High uncertainty requires robust planning across scenarios rather than point estimates.

Proponents: Cautious researchers, Some policymakers
Confidence: high (4/5)
Information Hazard Risk

Timeline publication creates racing dynamics. Compute data aids adversaries. False precision worse than acknowledged uncertainty. Focus on safety regardless of timelines.

Proponents: Security-focused researchers, Some MIRI-adjacent views
Confidence: low (2/5)
Fundamentally Uncertain

AI development too discontinuous to forecast meaningfully. Unknown unknowns dominate. Resources better spent on robustness than prediction.

Proponents: Anti-forecasting researchers, Some capability pessimists
Confidence: low (2/5)

Leadership and Organization

Key Personnel

Leadership Team

Jaime Sevilla
Director & Co-Founder
Tamay Besiroglu
Senior Researcher
Anson Ho
Research Scientist
Various data scientists
Database Team

Organizational Structure

FunctionTeam SizeKey Responsibilities2025 Focus
Research15-18 peopleForecasting, analysis, publicationsFrontierMath, Epoch Capabilities Index
Data & Engineering8-10 peopleDatabase infrastructure, automationAI Chip Sales explorer, model tracking
Operations4-6 peopleFunding, administration, communicationsGrant management, public engagement
AdvisoryExternalPolicy guidance, technical reviewAcademic partnerships
Total34 employeesHeadquartered in San FranciscoGrowing capacity

Funding Profile

FunderAmountPeriodPurpose
Coefficient Giving (2025)$1,132,4882 yearsGeneral support
Coefficient Giving (additional)$1,922,565Multi-yearGeneral support
Coefficient Giving (FrontierMath)$10,0002025Benchmark improvements
Coefficient Giving (2022)$1,960,000InitialOrganization founding
Total raised≈$13M+2022-2025Research & operations

Additional funding sources:

  • Future of Humanity Institute (historical support)
  • Government contracts for specific projects
  • Research grants from academic institutions

Comparative Analysis

vs. Other Forecasting Organizations

OrganizationFocusMethodologyUpdate FrequencyPolicy ImpactDatabase Size
Epoch AIAI-specific empirical dataMultiple quantitative methodsContinuousHigh3,200+ models
MetaculusCrowdsourced forecastingPrediction aggregationReal-timeMediumN/A
AI ImpactsHistorical AI analysisCase studies, trend analysisIrregularMediumLimited
FHIExistential risk researchAcademic researchProject-basedHighN/A
METRModel evaluationsTechnical testingPer-modelGrowingN/A
Our World in DataData visualizationData aggregation (uses Epoch data)RegularHighDerivative

Relationship to Safety Organizations

Organization TypeRelationship to EpochInformation Flow
Safety research orgsData consumersEpoch → Safety orgs
AI labsData subjectsLabs → Epoch (reluctantly)
Government bodiesPolicy clientsEpoch ↔ Government
Think tanksResearch partnersCollaborative

Future Directions and Challenges

Research Roadmap (2025-2027)

2026 Plans (from Impact Report):

  • Development of 2 new benchmarks beyond FrontierMath
  • Continued expansion of the Epoch Capabilities Index
  • Deeper analysis of decentralized training feasibility (10 GW training runs across thousands of kilometers)
  • Enhanced AI Chip Sales data explorer across Nvidia, Google, Amazon, AMD, and Huawei

Expanding scope:

  • Multimodal training data analysis beyond text
  • Energy consumption and environmental impact tracking (largest data centers now approaching gigawatt scale)
  • International AI development monitoring (enhanced coverage of Chinese and European models)
  • Risk assessment frameworks for different development pathways

Methodological improvements:

  • Better algorithmic progress measurement via Epoch Capabilities Index
  • Synthetic data quality and scaling analysis
  • Economic impact modeling of AI deployment
  • Scenario analysis for different development paths

Scaling Challenges

ChallengeCurrent LimitationPlanned SolutionPriority
Data collectionManual curation, limited sourcesAutomated scraping, industry partnershipsHigh
International coverageUS/UK bias in dataPartnerships with Chinese and European researchersHigh
Real-time trackingLag in proprietary model informationIndustry reporting standards advocacyMedium
Resource constraints34 person teamGradual expansion with $1.1M/year budgetMedium
Compute governance gapsThreshold accuracy uncertainBetter compute-capability correlation researchHigh
Open-source proliferationFrontier open models approaching 10^26 FLOPPolicy recommendations for dual governanceHigh

Key Questions

  • ?How accurate are extrapolation-based AI timeline forecasts given potential discontinuities?
  • ?Will synthetic data generation solve the training data bottleneck or create new limitations?
  • ?How should compute governance adapt as algorithmic efficiency reduces compute as a chokepoint?
  • ?What level of transparency in AI development is optimal for governance without security risks?
  • ?How can empirical forecasting organizations maintain independence while engaging with policymakers?
  • ?What leading indicators best predict dangerous capability emergence beyond compute scaling?

Sources & Resources

Primary Resources

Resource TypeDescriptionLink
AI Models Database3,200+ models tracked with training compute, parameters, costepoch.ai/data/ai-models
Machine Learning TrendsReal-time visualization of AI progress metricsepoch.ai/trends
FrontierMath Benchmark350 expert-crafted math problemsepoch.ai/frontiermath
Epoch Capabilities IndexUnified capability measurement across benchmarksepoch.ai/benchmarks
Research Blog36+ Data Insights, 37+ Gradient Updates in 2025epoch.ai/blog
AI Chip SalesGlobal compute capacity tracking (15M+ H100-equivalents)epoch.ai/data

Key Publications

TitleYearImpactCitation
"Compute Trends Across Three Eras of Machine Learning"2022Foundational for compute governanceSevilla et al.
"Will We Run Out of Data?"2022Sparked synthetic data research boomVillalobos et al.
"Algorithmic Progress in Computer Vision"2023Quantified efficiency improvementsBesiroglu et al.
"FrontierMath: A Benchmark for Evaluating Advanced Mathematical Reasoning"2024arXiv paper on benchmark methodologyEpoch AI et al.
Top 10 Data Insights of 20252025Annual synthesis of key findingsEpoch AI
2025 Impact Report2025Comprehensive organizational reviewEpoch AI

External Coverage

Source TypeDescriptionExample Links
Policy DocumentsGovernment citations of Epoch workUS NAIRR, UK AI White Paper
Academic CitationsResearch building on Epoch dataGoogle Scholar search
Data VisualizationOur World in Data uses Epoch datasetsTraining compute charts
Media CoverageNYT "2024 Good Tech Awards", InfoQ FrontierMath coverageMIT Technology Review
Industry AnalysisBusiness intelligence using Epoch metricsCB Insights, McKinsey AI reports
EA CommunityEA Forum discussions, 80,000 Hours listingsActive engagement

Related Pages

Top Related Pages

Analysis

AGI Development

Concepts

AI Uplift Assessment ModelSharp Left TurnAGI TimelineLarge Language Models

Organizations

Epoch AI

Transition Model

Compute & HardwareSafety Research

Risks

AI Winner-Take-All Dynamics