Long-Timelines Technical Worldview
Long-Timelines Technical Worldview
Comprehensive overview of the long-timelines worldview (20-40+ years to AGI, 5-20% P(doom)), arguing for foundational research over rushed solutions based on historical AI overoptimism, current systems' limitations, and scaling constraints. Provides concrete career and research prioritization guidance but lacks novel synthesis—primarily organizes existing arguments from Brooks, Marcus, and Mitchell.
Quick Assessment
| Dimension | Assessment | Evidence |
|---|---|---|
| Current Consensus | Minority position among AI researchers | Metaculus median AGI prediction: 2031; industry leaders predict 2-5 years |
| Expert Support | 20-30% of surveyed researchers | 76% believe scaling alone insufficient for AGI |
| Historical Track Record | Strong precedent for skepticism | AI predictions consistently wrong for 60+ years |
| Core Crux | Paradigm sufficiency | Whether deep learning + scaling reaches general intelligence |
| Research Priority | Foundational theory | Agent foundations, interpretability theory, formal verification |
| Risk Assessment (P(doom)) | 5-20% by 2100 | Lower than doomer estimates (25-80%) due to extended timeline |
| Funding Alignment | Supports longer-term research | AI safety research receives only $110-170M/year vs $252B corporate AI investment |
Core belief: Transformative AI is further away than many think. This gives us time for careful, foundational research rather than rushed solutions.
Timeline to AGI
The long-timelines worldview predicts significantly longer development horizons than short-timelines perspectives, fundamentally altering strategic priorities for AI safety research and intervention planning. As of December 2024, Metaculus forecasters average a 25% chance of AGI by 2027 and 50% by 2031—down from a median of 50 years away as recently as 2020.
| Source | AGI Estimate | Methodology | Confidence Level |
|---|---|---|---|
| Long-timelines view | 2045-2065+ | Historical pattern analysis + paradigm skepticism | Medium-High |
| Metaculus forecasters (2024) | 2031 (50% median) | Aggregated prediction market | 1,700 forecasters |
| AI researcher survey (2023) | 2047-2116 | Academic survey | Median varies 70 years by framing |
| Epoch AI Direct Approach | 2033 (50%) | Compute trend extrapolation | Model-based estimate |
| Industry leaders (OpenAI, Anthropic) | 2027-2030 | Internal capability assessment | Shortest estimates, potential bias |
| Rodney Brooks | Far, far further than claimed | Historical track record analysis | Publicly tracked predictions |
Key divergence: The 13-year median shift in AI researcher surveys between 2022-2023 suggests high uncertainty and susceptibility to framing effects. Long-timelines proponents argue this volatility reflects hype cycles rather than genuine technical progress toward AGI.
P(AI existential catastrophe by 2100)
While taking AI risk seriously, the long-timelines worldview assigns lower probabilities to existential catastrophe due to extended opportunity for alignment research, iterative testing, and institutional adaptation.
| Expert/Source | Estimate | Reasoning |
|---|---|---|
| Long-timelines view | 5-20% | Extended timelines provide multiple advantages for safety: decades for careful foundational research on agent alignment theory, time to observe warning signs in increasingly capable systems, opportunity for international coordination and governance development, and ability to iterate on alignment techniques across multiple generations of AI systems. The lower bound reflects that alignment remains genuinely difficult even with more time; the upper bound acknowledges we might still fail to solve core technical challenges. |
Key Links
| Source | Link |
|---|---|
| Official Website | commons.wikimedia.org |
| Wikipedia | en.wikipedia.org |
Overview
The long-timelines technical worldview holds that transformative AI is decades away rather than years. This isn't mere optimism or wishful thinking - it's based on specific views about the difficulty of achieving human-level intelligence, skepticism about current paradigms, and historical patterns in AI progress.
Diagram (loading…)
flowchart TD
subgraph Evidence["Evidence Base"]
HIST[Historical Overoptimism]
LIMITS[Current System Limitations]
SCALE[Scaling Diminishing Returns]
end
subgraph CoreBeliefs["Core Beliefs"]
PARADIGM[Paradigm Shift Required]
SLOW[Slow Takeoff Expected]
TIME[20-40+ Years to AGI]
end
subgraph Priorities["Research Priorities"]
FOUND[Agent Foundations]
INTERP[Deep Interpretability]
THEORY[Theoretical Safety]
FIELD[Field Building]
end
subgraph Outcomes["Strategic Outcomes"]
LOWER[Lower P doom: 5-20%]
ROBUST[Robust Solutions]
COORD[Time for Coordination]
end
HIST --> PARADIGM
LIMITS --> PARADIGM
SCALE --> PARADIGM
PARADIGM --> TIME
TIME --> SLOW
TIME --> FOUND
TIME --> INTERP
TIME --> THEORY
TIME --> FIELD
FOUND --> ROBUST
INTERP --> ROBUST
THEORY --> ROBUST
FIELD --> COORD
ROBUST --> LOWER
COORD --> LOWER
style HIST fill:#fee
style LIMITS fill:#fee
style SCALE fill:#fee
style TIME fill:#e6f3ff
style LOWER fill:#dfd
style ROBUST fill:#dfdThis extended timeline fundamentally changes strategic priorities. Instead of rushing to patch current systems or advocating for immediate pause, long-timelines researchers can pursue deep, foundational work that might take decades to bear fruit.
Key distinction: This is not the same as the optimistic worldview. Long-timelines researchers take alignment seriously and don't trust current techniques to scale. They're not optimistic about alignment being easy - they're pessimistic about timelines being short.
Characteristic Beliefs
| Crux | Long-Timelines Position | Short-Timelines Position | Key Evidence |
|---|---|---|---|
| Timelines | AGI 20-40+ years (2045-2065) | AGI 2-10 years (2027-2035) | Survey framing effects cause 70-year median variance |
| Paradigm | New paradigms required beyond scaling | Scaling + engineering solves remaining gaps | 76% of experts say scaling alone insufficient |
| Takeoff | Slow, observable over years | Fast or discontinuous possible | Historical technology adoption rates |
| Scaling outlook | Diminishing returns imminent | Continued exponential gains | Ilya Sutskever: "Age of scaling" may be ending |
| Alignment difficulty | Hard, but sufficient time to solve | Hard, and racing against clock | Depends on timeline beliefs |
| Current LLM relevance | Uncertain if informs future AGI | Direct path to AGI | Architectural discontinuity question |
| Deceptive alignment | Relevant but not imminent threat | Critical near-term concern | Capability threshold dependent |
| Coordination feasibility | More feasible with extended time | Difficult under time pressure | AI safety funding: $150-170M vs $252B AI investment |
| P(doom) | 5-20% by 2100 | 25-80% by 2100 | Extended time for iteration and response |
Timeline Arguments
Several independent arguments support longer timelines:
1. Intelligence is harder than it looks
Current AI systems are impressive but lack capabilities that Melanie Mitchell argues are fundamental to general intelligence:
- Robust generalization: Systems fail in novel contexts despite strong benchmark performance
- Abstract reasoning: Mitchell's 2024 research shows current AI lacks humanlike abstraction and analogy capabilities
- World models: AI lacks "rich internal models of the world" that reflect causes rather than correlations
- Efficient learning: Humans learn from limited examples; LLMs require massive data
- Common sense: Fundamental gaps in causal and physical reasoning persist
Each of these might require breakthroughs that scaling alone cannot provide.
2. Historical track record
AI predictions have consistently been overoptimistic—Rodney Brooks has publicly tracked failed predictions since 2017:
| Era | Prediction | Reality | Years Off |
|---|---|---|---|
| 1960s | Human-level AI by 1985 | First AI winter | 40+ years |
| 1980s | Expert systems would transform economy | Brittleness, second AI winter | 30+ years |
| 2017 | Full self-driving by 2020 | GM shut Cruise after $10B investment (2024) | Ongoing |
| 2023 | LLMs are path to near-term AGI | Scaling showing diminishing returns | TBD |
As Brooks notes: "None" of the 2017 near-term predictions have materialized.
3. Scaling might not be enough
While scaling has driven recent progress, multiple experts warn of limits:
Ilya Sutskever (OpenAI co-founder, Safe Superintelligence Inc.): "From 2012 to 2020, it was the age of research. From 2020 to 2025, it was the age of scaling... I don't think that's true anymore."
| Constraint | Current Status | Long-Term Trajectory |
|---|---|---|
| Compute costs | GPT-4 training: $100M+; next-gen: $1B+ | Superlinear cost growth per capability unit |
| Data availability | Already training on most of internet | Synthetic data quality issues uncertain |
| Energy requirements | Data centers consuming city-scale power | Environmental and infrastructure limits |
| Algorithmic efficiency | 2024 gains primarily in post-training | Pre-training scaling laws potentially breaking down |
Gary Marcus coined "diminishing returns" in 2022; recent observations suggest "adding more data does not actually solve the core underlying problems."
4. Economic and institutional barriers
Even if technically feasible, deployment faces substantial friction:
- Compute costs: Training frontier models now costs $100M-1B+, limiting who can participate
- Energy requirements: Data centers require gigawatts; infrastructure buildout takes years
- Capital requirements: Global AI investment reached $252B in 2024, but concentrated in few actors
- Regulatory barriers: EU AI Act, emerging US state legislation creating compliance costs
- Adoption timelines: Brooks notes even IPv4→IPv6 transition, started in 2001, is still only 50% complete
Takeoff Speed
Long-timelines researchers typically expect slow takeoff:
Gradual progress: Incremental improvements across many years
- Can observe AI getting more capable
- Time to respond to warning signs
- Opportunities to iterate on alignment
Multiple bottlenecks: Progress limited by many factors
- Hardware constraints
- Data availability
- Algorithmic insights
- Integration challenges
- Social and regulatory adaptation
Continuous deployment: AI capabilities integrated gradually
- Society adapts incrementally
- Institutions evolve alongside AI
- Norms and regulations co-develop
This contrasts sharply with fast takeoff scenarios where recursive self-improvement leads to rapid capability explosion.
Key Proponents and Perspectives
Academic Researchers
Rodney Brooks (former MIT CSAIL director, iRobot/Robust.AI founder)
"Even if it is possible I personally think we are far, far further away from understanding how to build AGI than many other pundits might say."
| Prediction Area | Brooks' Assessment | Track Record |
|---|---|---|
| Self-driving cars | Too optimistic by 10+ years | GM shut Cruise after $10B investment |
| LLM-to-AGI path | "Hubris similar to 2017 self-driving" | Publicly tracking since 2017 |
| Technology adoption | Consistently overestimated speed | IPv4→IPv6: 23+ years, still 50% complete |
| Current AI research | "Stuck on same issues for 50 years" | Common sense, reasoning gaps persist |
Brooks warns of "FOBAWTPALSL"—Fear of Being a Wimpy Techno-Pessimist and Looking Stupid Later—driving uncritical AI optimism.
Gary Marcus (NYU emeritus, cognitive scientist)
Published "Deep Learning: A Critical Appraisal" (2018) identifying 10 fundamental limitations, and "Taming Silicon Valley" (2024) arguing "we are not on the best path right now, either technically or morally."
Key arguments:
- Brittleness: Systems fail unpredictably on slight distribution shifts
- Hybrid AI necessity: Combining neural networks with symbolic reasoning (e.g., AlphaFold2) works better than pure deep learning
- Generalization failures: Pattern matching is not understanding
- Financial bubble: "People are valuing AI companies as if they're going to solve AGI. I don't think we're anywhere near AGI."
Melanie Mitchell (Santa Fe Institute)
Author of Artificial Intelligence: A Guide for Thinking Humans (2019); published four major papers in 2024 on AI limitations including work in Science.
Key research findings:
- Abstraction gap: "No current AI system is anywhere close to a capability of forming humanlike abstractions or analogies"
- World model deficit: AI lacks "rich internal models of the world that reflect the causes of events rather than merely correlations"
- Benchmark failure: "AI systems ace benchmarks yet stumble in the real world"
- AGI skepticism: "Today's AI is far from general intelligence, and I don't believe that machine 'superintelligence' is anywhere on the horizon"
Survey Evidence
A 2024 survey chaired by Brooks found:
- 76% of 475 respondents said scaling current approaches will not be sufficient for AGI
- This challenges the dominant industry narrative that "more compute = AGI"
Alignment Researchers with Longer Timelines
Not all alignment researchers believe in short timelines:
- Focus on foundational theory requiring 10-20+ year research programs
- Skeptical current LLM architectures inform future AGI systems
- Prefer robust solutions over patches that may not transfer
Priority Approaches
Given long-timelines beliefs, research priorities differ from short-timelines views. The extended horizon allows investment in high-risk, high-reward research that requires decades to mature.
Research Priority Comparison
| Approach | Long-Timelines Priority | Short-Timelines Priority | Funding Status (2024-25) |
|---|---|---|---|
| Agent foundations | Very High | Low-Medium | $5-15M/year via MIRI, academic grants |
| Mechanistic interpretability | High | High | $30-50M/year via labs + Coefficient Giving |
| RLHF/current alignment | Low-Medium | Very High | $100M+/year via frontier labs |
| Formal verification | High | Low | $10-20M/year, primarily academic |
| Field building/education | Very High | Medium | $20-40M/year via foundations |
| Pause/moratorium advocacy | Low | High | Variable, advocacy-funded |
| Compute governance | Medium | Very High | Government + policy focus |
| Rapid deployment safety | Low | Very High | Lab-funded, urgent framing |
Funding context: AI safety research receives only $150-170M/year total (up 36% from 2024), while corporate AI investment reached $252B in 2024. Long-timelines proponents argue foundational work is underfunded relative to its importance if timelines extend.
1. Agent Foundations
Deep theoretical work on fundamental questions:
Decision theory:
- How should rational agents behave?
- Logical uncertainty
- Updateless decision theory
- Embedded agency
Value alignment theory:
- What does it mean for an agent to have values?
- How can values be specified?
- Corrigibility and interruptibility
- Utility function construction
Ontological crises:
- How do agents update when their world model changes fundamentally?
- Preserving values across paradigm shifts
Advantage of long timelines: This work might take 10-20 years to mature, which is fine if AGI is 30+ years away.
2. Interpretability and Understanding
Deep understanding of how AI systems work:
Mechanistic interpretability:
- Reverse-engineer neural networks
- Understand individual neurons and circuits
- Build comprehensive models of model internals
Theoretical foundations:
- Why do neural networks generalize?
- What are the fundamental limits?
- Mathematical theory of deep learning
Conceptual understanding:
- What are models actually learning?
- Representations and abstractions
- Transfer and generalization
Advantage of long timelines: Can build interpretability tools gradually, improving them over decades.
3. Foundational Research
First-principles approaches without time pressure:
Alternative paradigms:
- Explore architectures beyond current deep learning
- Investigate hybrid systems
- Study biological intelligence for insights
Robustness and verification:
- Formal methods for AI
- Provable safety properties
- Mathematical guarantees
Comprehensive testing:
- Extensive empirical research
- Long-term studies of AI behavior
- Edge case exploration
Advantage of long timelines: Can pursue high-risk, high-reward research without urgency.
4. Field Building
Growing the community for long-term impact:
Academic infrastructure:
- University departments and programs
- Curriculum development
- Textbooks and educational materials
Talent pipeline:
- Undergraduate and graduate training
- Interdisciplinary programs
- Career paths in alignment
Research ecosystem:
- Conferences and workshops
- Journals and publications
- Collaboration networks
Advantage of long timelines: Field-building pays off over decades.
5. Careful Empirical Work
Thorough investigation of current systems:
Understanding limitations:
- Where do current approaches fail?
- What are fundamental vs. contingent limits?
- Generalization studies
Alignment properties:
- How do current alignment techniques work?
- What are their scaling properties?
- When do they break down?
Transfer studies:
- Will current insights transfer to future AI?
- What's paradigm-specific vs. general?
Advantage of long timelines: Can be thorough rather than rushed.
Deprioritized Approaches
Given long-timelines beliefs, some approaches are less urgent:
| Approach | Why Less Urgent |
|---|---|
| Pause advocacy | Less immediate urgency |
| RLHF improvements | May not transfer to future paradigms |
| Current-system safety | Systems may not be path to AGI |
| Race dynamics | More time reduces racing pressure |
| Quick fixes | Can pursue robust solutions instead |
Note: "Less urgent" doesn't mean "useless" - just different prioritization given beliefs.
Strongest Arguments
1. Historical Overoptimism
AI predictions have been systematically wrong for 60+ years. 80,000 Hours analysis shows median expert estimates have shortened by 13 years between 2022-2023 surveys alone—suggesting high volatility driven by hype rather than genuine progress.
| Period | Prediction | Outcome | Investment Lost/Delayed |
|---|---|---|---|
| 1965-1975 | Machine translation "solved in 5 years" | ALPAC report ended funding | $20M+ wasted |
| 1980-1987 | Expert systems market $5B by 1990 | Second AI winter; Lisp machine collapse | $1B+ industry crash |
| 2012-2017 | Self-driving by 2020 | GM shut Cruise after $10B | $100B+ industry-wide |
| 2020-2023 | LLM scaling → AGI in 3-5 years | Scaling hitting diminishing returns | TBD |
Pattern: Each generation thinks they're on the path to AGI. Each is wrong. Current optimism about LLMs may repeat this pattern—76% of surveyed experts believe scaling alone insufficient.
2. Current Systems' Fundamental Limitations
Despite impressive performance, current AI lacks:
Robust generalization:
- Adversarial examples fool vision systems
- Out-of-distribution failures
- Brittle in novel situations
True understanding:
- Pattern matching vs. comprehension
- Lack of world models
- No common sense reasoning
Efficient learning:
- Require massive data (humans learn from few examples)
- Don't transfer knowledge well across domains
- Can't explain their reasoning reliably
Abstract reasoning:
- Struggle with novel problems requiring insight
- Limited analogical reasoning
- Poor at systematic generalization
These might require fundamental breakthroughs, not just scaling.
3. Scaling Has Limits
Current progress relies on scaling, but:
Compute constraints:
- Energy costs grow exponentially
- Chip production has physical limits
- Economic viability uncertain at extreme scales
Data constraints:
- Already training on most of internet
- Synthetic data has quality issues
- Diminishing returns from more data
Algorithmic efficiency:
- Gains are uncertain and irregular
- May hit fundamental limits
- Efficiency improvements are hard to predict
Returns diminishing:
- Each order of magnitude improvement costs more
- Performance gains may be slowing
- Knee of the curve might be near
4. Intelligence Requires More Than Current Approaches
Cognitive science and neuroscience suggest:
Embodiment: Intelligence might require physical interaction with world
Development: Human intelligence develops through years of experience
Architecture: Brain has specialized structures deep learning lacks
Mechanisms: Biological learning uses mechanisms we don't understand
Consciousness: Role of consciousness in intelligence unclear
If any of these are necessary, current approaches are missing key ingredients.
5. Slow Takeoff Is Likely
Multiple bottlenecks slow progress:
Integration challenges: Deploying AI into real systems takes time
Social adaptation: Society needs to adapt to new capabilities
Institutional barriers: Regulation, cultural resistance, coordination
Economic constraints: Funding and resources are limited
Technical obstacles: Each capability advance requires solving multiple problems
No reason to expect rapid discontinuities - smooth progress is default.
6. Time for Solutions Reduces Risk
Longer timelines mean:
Iterative improvement: Can refine alignment techniques over decades
Warning signs: Early systems give us data about problems
Coordination: More time for international cooperation
Institution building: Governance can develop alongside technology
Research maturation: Alignment solutions can be thoroughly tested
P(doom) is lower because we have time to get it right.
Main Criticisms and Counterarguments
"This Is Just Wishful Thinking"
Critique: Long-timelines view is motivated by hoping for more time, not actual evidence.
Response:
- Based on specific technical arguments, not hope
- Historical track record supports skepticism
- Many long-timelines people still take risk seriously
- If anything, short timelines might be motivated by excitement/fear
"Might Miss Critical Window"
Critique: If wrong about timelines, current window to shape AI development is missed.
Response:
- Can have uncertainty and hedge bets
- Foundational work pays off even in shorter timelines
- Better to have robust solutions late than rushed solutions now
- Can shift priorities if evidence changes
"Current Progress Is Different"
Critique: Unlike past failed approaches, deep learning and scaling are actually working. This time is different.
Response:
- Every generation thinks "this time is different"
- Deep learning has made progress but also has clear limits
- Scaling can't continue indefinitely
- Path from current systems to AGI remains unclear
"LLMs Show Emergent Capabilities"
Critique: Large language models show unexpected emergent abilities, suggesting scaling might reach AGI.
Response:
- "Emergent" capabilities often just smooth trends that appear suddenly in metrics
- Still lack robust reasoning, planning, and understanding
- Emergence in narrow tasks doesn't imply general intelligence
- May hit ceiling well below human-level
"Moravec's Paradox Resolved"
Critique: Deep learning solved perception problems thought to be hardest (vision, language). The rest will follow.
Response:
- Perception was hard for symbolic AI, not necessarily hardest overall
- Reasoning and planning might be fundamentally harder
- "Harder" tasks (like abstract reasoning) remain difficult for current AI
- Different problems might require different solutions
"Missing Urgency"
Critique: Even if timelines are long, should work urgently to be safe.
Response:
- Urgency doesn't mean rushing to bad solutions
- Careful work is more valuable than hasty work
- Can be thorough without being complacent
- False urgency leads to wasted effort
"Paradigm Shifts Can Be Rapid"
Critique: Even if deep learning isn't enough, sudden breakthroughs could change timelines overnight.
Response:
- Breakthroughs still require years to commercialize
- Integration takes time even if insight is sudden
- Most progress is gradual, not revolutionary
- Can update if breakthrough occurs
What Evidence Would Change This View?
Long-timelines researchers would update toward shorter timelines given specific, measurable developments:
Evidence That Would Strongly Update Toward Shorter Timelines
| Evidence Type | Specific Threshold | Current Status (2025) | Update Magnitude |
|---|---|---|---|
| Scaling continuation | 2+ more OOMs without diminishing returns | Returns appear diminishing | Very strong update |
| Robust reasoning | Pass novel math/science problems consistently | Fails on out-of-distribution | Strong update |
| Transfer learning | Same model excels across 10+ very different domains | Still domain-specific fine-tuning needed | Strong update |
| Common sense | Pass adversarial physical reasoning tests | Mitchell's research shows consistent failures | Strong update |
| Expert consensus shift | Greater than 70% of surveyed researchers predict AGI within 10 years | Currently approximately 30-40% | Moderate update |
| Prediction market movement | Metaculus median drops below 2028 | Currently 2031 median | Moderate update |
Theoretical Breakthroughs That Would Update
- Clear path to generalization: Formal demonstration that current architectures can achieve human-level abstraction
- World model success: AI systems building accurate causal models (not just correlations)
- Efficient learning: Systems learning as efficiently as humans (100x-1000x data reduction)
Economic/Investment Indicators
Current investment levels (2024: $252B corporate, $150-170M safety) already suggest serious commitment. Further indicators:
- Government Manhattan Project: $50B+/year coordinated government program (currently $3.3B federal)
- Energy breakthrough: Fusion or next-gen nuclear enabling 10x cheaper compute
- Chip breakthrough: 100x efficiency gains beyond current trajectory
What Has Already Updated Timelines
Several developments have shortened some long-timelines estimates:
- GPT-4/Claude-level reasoning capabilities (2023-2024)
- Chain-of-thought and reasoning improvements
- Multimodal integration success
- Test-time compute scaling (o1, etc.)
However, these haven't addressed the core limitations Mitchell identifies—abstraction, world models, efficient learning—that long-timelines proponents consider fundamental.
Implications for Action and Career
If you hold long-timelines beliefs, strategic implications include:
Research Career Paths
Academic research:
- PhD programs in AI alignment
- Theoretical research with long time horizons
- Building foundational knowledge
Deep technical work:
- Agent foundations
- Interpretability theory
- Formal verification
- Mathematical approaches
Interdisciplinary work:
- Cognitive science and AI
- Neuroscience-inspired AI
- Philosophy of mind and AI
Advantage: Can pursue questions requiring 5-10 year research programs
Field Building
Education and training:
- Develop curricula
- Write textbooks
- Train next generation
Community building:
- Organize conferences
- Build research networks
- Create institutions
Public scholarship:
- Explain AI alignment to broader audiences
- Attract talent to the field
- Build prestige and legitimacy
Advantage: Field-building investments pay off over decades
Careful Empirical Work
Current systems research:
- Thorough investigation of limitations
- Understanding what transfers to future systems
- Building tools and methodologies
Comprehensive testing:
- Long-term studies
- Edge case exploration
- Robustness analysis
Advantage: Can be thorough rather than rushed
Strategic Positioning
Flexibility:
- Build skills that remain valuable across scenarios
- Create options for different timeline outcomes
- Hedge uncertainty
Sustainable pace:
- Marathon, not sprint
- Avoid burnout from false urgency
- Build career that lasts decades
Leverage points:
- Focus on work with long-term impact
- Build infrastructure others can use
- Create knowledge that persists
Internal Diversity
The long-timelines worldview includes significant variation:
Timeline Estimates
Medium (20-30 years): More cautious, still somewhat urgent
Long (30-50 years): Standard long-timelines position
Very long (50+ years): Highly skeptical of current approaches
Risk Assessment
Moderate risk, long timelines: Still concerned but have time
Low risk, long timelines: Technical problem is tractable with time
High risk, long timelines: Hard problem, fortunately have time
Research Focus
Pure theory: Agent foundations, decision theory
Applied theory: Interpretability, verification
Empirical: Understanding current systems
Hybrid: Combination of approaches
Attitude Toward Current Work
Skeptical: Current LLM work likely irrelevant to AGI
Uncertain: Might be relevant, worth studying
Engaged: Working on current systems while believing AGI is far
Relationship to Other Worldviews
vs. Doomer
Disagreements:
- Fundamental disagreement on timelines
- Different urgency levels
- Different research priorities
Agreements:
- Alignment is hard
- Current techniques may not scale
- Take risk seriously
vs. Optimistic
Disagreements:
- Long-timelines folks more worried about alignment difficulty
- Don't trust market to provide safety
- More skeptical of current approaches
Agreements:
- Have time for solutions
- Catastrophe is not inevitable
- Research can make progress
vs. Governance-Focused
Disagreements:
- Less urgency about policy
- More focus on technical foundations
- Different time horizons
Agreements:
- Multiple approaches needed
- Coordination is valuable
- Institutions matter
Practical Considerations
Career Planning
Skill development: Can pursue deep expertise
Network building: Relationships develop over years
Institution building: Create enduring organizations
Work-life balance: Sustainable pace over decades
Research Strategy
Patient capital: Pursue high-risk, long-horizon research
Foundational work: Build knowledge infrastructure
Replication and verification: Be thorough
Documentation: Create resources for future researchers
Community Norms
Thorough review: Take time for peer review
Replication: Verify important results
Education: Train people properly
Standards: Build quality norms
Representative Quotes
"Every decade, people think AGI is 20 years away. It's been this way for 60 years. Maybe we should update on that." - Rodney Brooks
"Current AI is like a high school student who crammed for the test - impressive performance on specific tasks, but lacking deep understanding." - Gary Marcus
"The gap between narrow AI and general intelligence is not about scale - it's about fundamental architecture and learning mechanisms we don't yet understand." - Melanie Mitchell
"I'd rather solve alignment properly over 20 years than rush to a solution in 5 years that fails catastrophically." - Long-timelines researcher
"The best research takes time. If we have that time, we should use it wisely rather than pretending we don't." - Academic alignment researcher
Common Misconceptions
"Long-timelines people aren't worried about AI risk": False - they take it seriously but believe we have time
"It's just procrastination": No - it's a belief about technology development pace
"They're not working on alignment": Many do foundational alignment work
"They think alignment is easy": No - they think it's hard but we have time to solve it
"They're out of touch with recent progress": Many are deep in the technical details
Strategic Implications
If Long Timelines Are Correct
Good news:
- Time for careful research
- Can build robust solutions
- Opportunity for coordination
- Field can mature properly
Challenges:
- Maintaining focus over decades
- Avoiding complacency
- Sustaining funding and interest
- Adapting as technology evolves
If Wrong (Timelines Are Short)
Risks:
- Missing critical window
- Foundational work not finished
- Solutions not ready
- Institutions not built
Mitigations:
- Maintain some urgency even with long-timelines belief
- Monitor leading indicators
- Be prepared to shift priorities
- Hedge with faster-payoff work
Recommended Reading
Arguments for Longer Timelines
- AI Impacts: Likelihood of Discontinuous Progress↗🔗 web★★★☆☆AI ImpactsAI Impacts: Likelihood of Discontinuous ProgressPublished by AI Impacts, this piece informs the classic 'fast vs. slow takeoff' debate by grounding predictions in historical data on technological discontinuities, making it a useful reference for safety researchers assessing risk timelines.https://aiimpacts.org/author/katja/ (2018)An AI Impacts analysis examining the probability that progress toward AGI will be discontinuous—featuring sudden jumps or takeoffs—rather than gradual. It surveys historical pre...ai-safetycapabilitiesexistential-riskalignment+2Source ↗
- Gary Marcus: Deep Learning Alone Won't Get Us to AGI↗🔗 web★★★☆☆WIREDGary Marcus: Deep Learning Alone Won't Get Us to AGIGary Marcus is a prominent critic of deep learning maximalism; this Wired piece summarizes his longstanding argument that neural networks alone are insufficient for AGI, relevant to debates about AI timelines and capability trajectories in AI safety discussions.Condé NastGary Marcus argues in Wired that deep learning, despite its impressive achievements, has fundamental limitations that prevent it from achieving human-level artificial general in...agicapabilitiesai-safetyalignment+1Source ↗
- Melanie Mitchell: Why AI Is Harder Than We Think↗📄 paper★★★☆☆arXivMelanie Mitchell: Why AI Is Harder Than We ThinkA widely-cited critical analysis of overconfidence in AI progress by cognitive scientist Melanie Mitchell; useful for grounding discussions of AI timelines, capability evaluation, and the gap between narrow AI performance and general intelligence.Melanie Mitchell (2021)126 citationsMelanie Mitchell argues that AI progress has repeatedly been derailed by four fallacies about the nature of intelligence, leading researchers to underestimate the difficulty of ...capabilitiesai-safetyalignmentevaluation+1Source ↗
- Rodney Brooks' Predictions Scorecard (2025) - Tracking failed AI predictions since 2017
- Gary Marcus: Taming Silicon Valley (MIT Press, 2024) - Comprehensive critique of AI hype
Timeline Forecasts and Analysis
- 80,000 Hours: Shrinking AGI Timelines (2025) - Comprehensive expert forecast review
- Epoch AI: Literature Review of Transformative AI Timelines - Model-based timeline analysis
- Metaculus AGI Forecasts - Aggregated prediction market
- Stanford HAI: 2025 AI Index Report - Comprehensive AI progress metrics
Technical Limitations
- On the Measure of Intelligence↗📄 paper★★★☆☆arXivOn the Measure of IntelligenceFoundational work on defining and measuring intelligence, critical for AI safety as it addresses how we evaluate and compare AI systems' capabilities and safety relevant properties.François Chollet (2019)This paper argues that current AI benchmarking practices, which measure skill at specific tasks like games, fail to capture true intelligence because skill can be artificially i...capabilitiestrainingevaluationcapability+1Source ↗ - François Chollet
- Shortcut Learning in Deep Neural Networks↗📄 paper★★★☆☆arXivShortcut Learning in Deep Neural NetworksThis paper identifies shortcut learning as a fundamental problem in deep neural networks where models exploit spurious correlations instead of learning robust features, directly relevant to AI safety concerns about model reliability and robustness.Robert Geirhos, Jörn-Henrik Jacobsen, Claudio Michaelis et al. (2020)2,668 citations · Nature Machine IntelligenceThis perspective paper identifies shortcut learning as a unifying explanation for many limitations in deep neural networks. Shortcuts are decision rules that achieve high perfor...capabilitiesevaluationbiosecuritySource ↗
- Underspecification in Machine Learning↗📄 paper★★★☆☆arXivUnderspecification in Machine LearningA highly influential paper from Google Brain demonstrating that ML reliability problems often stem from underspecification rather than overfitting, directly relevant to AI safety concerns about model predictability and deployment robustness.Alexander D'Amour, Katherine Heller, Dan Moldovan et al. (2020)815 citationsThis paper introduces 'underspecification' as a fundamental problem in ML pipelines: many different models with equivalent training performance can behave very differently in de...technical-safetydeploymentevaluationcapabilities+2Source ↗
- Deep Learning: A Critical Appraisal (Marcus, 2018) - Foundational critique identifying 10 limitations
- AI's Challenge of Understanding the World (Mitchell, 2024) - Latest research on abstraction gaps
Scaling Laws and Diminishing Returns
- Can AI Scaling Continue Through 2030? (Epoch AI) - Analysis of scaling constraints
- TechCrunch: AI Scaling Laws Showing Diminishing Returns (2024) - Industry reporting on scaling limits
Cognitive Science Perspectives
- Building Machines That Learn and Think Like People↗📄 paper★★★☆☆arXivBuilding Machines That Learn and Think Like PeopleFoundational cognitive science paper comparing human and machine learning approaches, examining how deep learning systems diverge from human intelligence in crucial ways—relevant for understanding AI alignment and developing more human-like AI reasoning.Brenden M. Lake, Tomer D. Ullman, Joshua B. Tenenbaum et al. (2016)855 citations · Behavioral and Brain SciencesThis paper argues that while deep neural networks have achieved impressive performance on tasks like object recognition and game-playing, they fundamentally differ from human in...capabilitiesbiosecuritySource ↗
- Melanie Mitchell: Abstraction and Analogy in AI - ArXiv paper on fundamental gaps
Foundational Research
- Agent Foundations for Aligning Machine Intelligence↗🔗 web★★★☆☆MIRIAgent Foundations for Aligning Machine IntelligenceThis is MIRI's official research guide, useful for understanding the agent-foundations approach to alignment and identifying open technical problems; best paired with MIRI's technical papers and the Embedded Agency sequence.Kolya T (2024)MIRI's research guide outlines the theoretical foundations and open problems in agent-based AI alignment, focusing on decision theory, logical uncertainty, corrigibility, and re...ai-safetyalignmenttechnical-safetycorrigibility+6Source ↗
- Embedded Agency↗✏️ blog★★★☆☆Alignment ForumEmbedded Agency (Sequence)A key MIRI-affiliated research sequence; essential reading for understanding agent-foundations approaches to alignment, particularly the theoretical gaps in classical agent models that motivate much of MIRI's and Redwood's technical research agendas.A foundational sequence by Scott Garrabrant and Abram Demski examining the deep theoretical challenges that arise when AI agents are embedded within—rather than external to—the ...agent-foundationsdecision-theorytechnical-safetyalignment+4Source ↗
- Future of Life Institute: 2025 AI Safety Index - Funding and progress tracking
Field Building
- A Guide to Writing High-Quality LessWrong Posts↗✏️ blog★★★☆☆LessWrongA Guide to Writing High-Quality LessWrong PostsA meta-resource for contributors to LessWrong, relevant to anyone seeking to share AI safety or rationality research on the platform; useful for understanding community writing standards but not directly about AI safety content.This guide provides practical advice for writing clear, well-structured, and intellectually rigorous posts on LessWrong, covering topics like argumentation, clarity, and communi...educationalcoordinationai-safetySource ↗
- LessWrong: AI Futures Timelines Model (Dec 2025) - Comprehensive timeline modeling
References
Gary Marcus argues in Wired that deep learning, despite its impressive achievements, has fundamental limitations that prevent it from achieving human-level artificial general intelligence. He contends that deep learning systems lack robust reasoning, generalization, and common sense, and that reaching AGI will require integrating additional approaches beyond neural networks. Marcus calls for hybrid systems that combine deep learning with symbolic reasoning and other AI paradigms.
This paper introduces 'underspecification' as a fundamental problem in ML pipelines: many different models with equivalent training performance can behave very differently in deployment. The authors demonstrate that standard ML training procedures return a set of equally good models, with no mechanism to prefer more robust or reliable ones, leading to unpredictable real-world failures.
Melanie Mitchell argues that AI progress has repeatedly been derailed by four fallacies about the nature of intelligence, leading researchers to underestimate the difficulty of achieving general AI. The paper examines historical overconfidence in AI timelines and capabilities, diagnosing systematic conceptual errors including conflating narrow task performance with general intelligence and underestimating the complexity of human cognition.
4Building Machines That Learn and Think Like PeoplearXiv·Brenden M. Lake, Tomer D. Ullman, Joshua B. Tenenbaum & Samuel J. Gershman·2016·Paper▸
This paper argues that while deep neural networks have achieved impressive performance on tasks like object recognition and game-playing, they fundamentally differ from human intelligence in important ways. The authors review cognitive science research and propose that human-like AI systems must go beyond pattern recognition to: (1) build causal models supporting explanation and understanding, (2) ground learning in intuitive theories of physics and psychology, and (3) leverage compositionality and learning-to-learn for rapid knowledge acquisition and generalization. The paper advocates for combining neural network advances with more structured cognitive models to achieve truly human-like machine learning.
5AI Impacts: Likelihood of Discontinuous ProgressAI Impacts·https://aiimpacts.org/author/katja/·2018▸
An AI Impacts analysis examining the probability that progress toward AGI will be discontinuous—featuring sudden jumps or takeoffs—rather than gradual. It surveys historical precedents of discontinuous progress in technology and science to inform predictions about how AI development might unfold. The piece is relevant to debates about fast vs. slow AI takeoff scenarios and associated safety implications.
A foundational sequence by Scott Garrabrant and Abram Demski examining the deep theoretical challenges that arise when AI agents are embedded within—rather than external to—the environments they reason about. It addresses decision theory, world-modeling, and alignment under the realistic condition that an agent is itself a physical subsystem of the world it must model and act upon.
This perspective paper identifies shortcut learning as a unifying explanation for many limitations in deep neural networks. Shortcuts are decision rules that achieve high performance on standard benchmarks but fail to generalize to real-world conditions or more challenging test scenarios. The authors argue that shortcut learning is a common characteristic across biological and artificial learning systems, drawing parallels from comparative psychology, education, and linguistics. The paper proposes recommendations for model interpretation, benchmarking practices, and robustness improvements to enhance transferability from laboratory settings to practical applications.
This paper argues that current AI benchmarking practices, which measure skill at specific tasks like games, fail to capture true intelligence because skill can be artificially inflated through prior knowledge and training data. The authors propose a formal definition of intelligence based on Algorithmic Information Theory, conceptualizing it as skill-acquisition efficiency across diverse tasks. They introduce the Abstraction and Reasoning Corpus (ARC), a benchmark designed with human-like priors to enable fair comparisons of general fluid intelligence between AI systems and humans, addressing the need for appropriate feedback signals in developing more intelligent and human-like artificial systems.
MIRI's research guide outlines the theoretical foundations and open problems in agent-based AI alignment, focusing on decision theory, logical uncertainty, corrigibility, and related mathematical challenges. It provides a roadmap for researchers interested in contributing to foundational alignment work. The guide situates these problems within the broader goal of ensuring advanced AI systems remain safe and beneficial.
This guide provides practical advice for writing clear, well-structured, and intellectually rigorous posts on LessWrong, covering topics like argumentation, clarity, and community norms. It aims to help authors communicate complex ideas effectively to the rationalist and AI safety community. Following these guidelines helps elevate discourse quality on the platform.
A Metaculus community forecasting question tracking predictions for when weakly general AI will be publicly known. Aggregates probabilistic estimates from forecasters on a key AI development milestone, providing crowd-sourced timeline predictions updated through December 2024.
A comprehensive synthesis by 80,000 Hours reviewing expert predictions on AGI timelines from multiple groups including AI lab leaders, researchers, and forecasters. The review finds a notable convergence toward shorter timelines, with many estimates suggesting AGI could arrive before 2030. Different expert communities that previously disagreed are now showing increasingly similar estimates.
Epoch AI reviews and compares quantitative models and expert judgment-based forecasts predicting when transformative AI will arrive, including biological anchors, semi-informative priors, and prediction market aggregates. Inside-view models tend to predict shorter timelines (median ~2052) while outside-view models predict longer timelines (median >2100), with expert judgment forecasts often more aggressive than either. The review also provides Epoch AI team's subjective weightings on the relative trustworthiness of each approach.
AI labs are confronting diminishing returns from traditional compute and data scaling (pretraining), prompting a shift toward 'test-time compute' scaling—giving models more computational resources to reason during inference. Industry figures including Ilya Sutskever, Marc Andreessen, and Satya Nadella have acknowledged this transition, pointing to OpenAI's o1 model as exemplifying the new paradigm.
The Future of Life Institute's AI Safety Index Summer 2025 systematically evaluates leading AI companies on safety practices, finding widespread deficiencies across risk management, transparency, and existential safety planning. Anthropic receives the highest grade of C+, indicating that even the best-performing company falls significantly short of adequate safety standards. The report serves as a comparative benchmark for industry accountability.
Epoch AI analyzes the key constraints and bottlenecks that could limit continued AI scaling through 2030, examining factors such as compute availability, energy infrastructure, data availability, and algorithmic progress. The analysis assesses whether current scaling trends in large language models and other AI systems can realistically be sustained over the next several years.
The 2025 Stanford AI Index Report tracks global AI private investment trends, organizational adoption rates, and economic impacts, finding the U.S. leads in AI funding and that 78% of organizations have adopted AI in at least one business function. It provides comprehensive data on how AI is reshaping labor markets, productivity, and industry sectors.
Gary Marcus, cognitive scientist and AI researcher, publishes critical analysis of AI systems with a focus on their limitations, hype, and risks. His work challenges overblown claims about large language models and advocates for responsible, hybrid approaches to AI development. The blog serves as a prominent skeptical voice in mainstream AI discourse.
The 2025 Stanford HAI AI Index Report provides a comprehensive annual survey of AI development across technical performance, economic investment, global competition, and responsible AI adoption. It synthesizes data from academia, industry, and government to track AI progress and societal impact. The report serves as a key reference for understanding where AI stands today and emerging trends shaping the field.