Longterm Wiki
Updated 2026-02-01HistoryData
Page StatusContent
Edited 12 days ago1.9k words
60
QualityGood
18
ImportancePeripheral
12
Structure12/15
705052%23%
Summary

Comprehensive tracking of Sam Altman's predictions shows he's directionally correct on AI trajectory and cost declines (10x/year validated) but consistently wrong on specific timelines (self-driving 2015, ChatGPT Pro profitability, GPT-5 launch). Pattern analysis reveals rhetoric shift from 'end of world' (2015) to 'will matter less than people think' (2024-25), with 4-5 clearly correct predictions, 3-4 wrong, and 10+ pending testable claims for 2025-2030.

Issues2
QualityRated 60 but structure suggests 80 (underrated by 20 points)
Links2 links could use <R> components

Sam Altman: Track Record

Person

Sam Altman: Track Record

Comprehensive tracking of Sam Altman's predictions shows he's directionally correct on AI trajectory and cost declines (10x/year validated) but consistently wrong on specific timelines (self-driving 2015, ChatGPT Pro profitability, GPT-5 launch). Pattern analysis reveals rhetoric shift from 'end of world' (2015) to 'will matter less than people think' (2024-25), with 4-5 clearly correct predictions, 3-4 wrong, and 10+ pending testable claims for 2025-2030.

1.9k words

This page documents Sam Altman's public predictions and testable claims to assess his forecasting accuracy. For biographical information, communication patterns, and credibility context, see the main Sam Altman page.

Summary Assessment

CategoryCountNotes
Clearly Correct4-5AI needing massive capital, cost declines, legal/medical AI assistance, compute as precious commodity
Partially Correct3-4GPT-4 limitations, AI productivity gains, agents emerging
Pending/Testable10+AGI by 2025-2029, superintelligence by 2030, job displacement, 10x scientific progress
Clearly Wrong3-4Self-driving cars (2015), ChatGPT Pro profitability, GPT-5 launch, AI election manipulation (2024)
Self-Corrected1-2AI creativity (acknowledged wrong), o3 AGI hype walkback

Overall pattern: Directionally correct on AI trajectory; consistently overoptimistic on specific timelines; rhetoric has shifted from "existential threat" (2015) to "will matter less than people think" (2024-2025).


Predictions: Resolved

Timeline Predictions

DateClaimTypeWhat HappenedStatusSource
2015Self-driving cars "in 3-4 years"InterviewFull self-driving still not achieved as of 2026❌ WrongTechCrunch
Pre-2020AI would never be "a really great creative thinker"InterviewDALL-E, Sora, and LLM creative writing proved this wrong❌ Wrong (self-acknowledged)Fortune
July 2020"The GPT-3 hype is way too much"Social mediaGPT-3 was limited but led to transformative ChatGPT⚠️ Interesting self-restraintHacker News
2021AI could read legal documents and give medical advice within 5 yearsEssayAI can now assist with legal and medical analysis✅ Largely correctMoore's Law for Everything
2021AI development would need massive capitalEssayOpenAI raised $20+ billion; compute costs enormous✅ CorrectSame
2021Cost of AI would fall dramaticallyEssayToken costs dropped dramatically (Altman later cited ≈150x from GPT-4 to GPT-4o)✅ CorrectMoore's Law for Everything; cost claim from Three Observations
2023GPT-4 "kind of sucks... relative to where we need to get to"PodcastGPT-4 was transformative but has clear limitations✅ Directionally correctLex Fridman Podcast #367
Dec 2024ChatGPT Pro at $200/month would be profitableBusiness claimOpenAI losing money on Pro due to heavy usage❌ WrongTechCrunch
Aug 2025GPT-5 launchProduct launchAdmitted they "totally screwed up" the rollout❌ Acknowledged failureFortune
May 2023Warned AI could manipulate voters in 2024 electionSenate testimonyAI had "negligible impact" on 2024 elections per Munich Security Conference analysis⚠️ Concern reasonable but didn't materializeSenate testimony

Cost Decline Prediction (Validated)

Quote (2025): "The cost to use a given level of AI falls about 10x every 12 months... Moore's law changed the world at 2x every 18 months; this is unbelievably stronger."

Source: "Three Observations" blog

Status: ✅ Largely validated by token pricing data.

Compute as Future Currency (2023)

Quote: "I think compute is going to be the currency of the future. I think it will be maybe the most precious commodity in the world."

Source: Lex Fridman Podcast

Status: ✅ Ongoing trend - increasingly validated by AI infrastructure investment.


Predictions: Pending

AGI Timeline Evolution

DateClaimTypeTestable ByCurrent StatusSource
2015Set "totally random" AGI date of 2025Interview2025Approaching test; now claims AGI achievable in 2025Bloomberg
Sept 2024"Superintelligence in a few thousand days"Essay≈2030-2038"Few thousand days" = 5.5-14 yearsThe Intelligence Age
Nov 2024OpenAI has "clear roadmap for achieving AGI by 2025"Interview2025Very aggressive; pendingY Combinator interview
Dec 2024"AGI will probably get developed during [Trump's] term" (2025-2029)Interview2029PendingBloomberg
Jan 2025"We are now confident we know how to build AGI"Blog post-Unfalsifiable without clear AGI definitionReflections blog
2025Superintelligence by 2030Interview2030"I would be very surprised if we haven't developed a superintelligent model capable of performing tasks beyond human reach by the end of 2030"TIME

"Short Timelines, Slow Takeoff" Position (Feb 2023)

Quote: "Many of us think the safest quadrant in this two-by-two matrix is short timelines and slow takeoff speeds; shorter timelines seem more amenable to coordination and more likely to lead to a slower takeoff due to less of a compute overhang."

Source: "Planning for AGI and beyond"

"The Gentle Singularity" (June 2025)

Quote: "We are past the event horizon; the takeoff has started. Humanity is close to building digital superintelligence, and at least so far it's much less weird than it seems like it should be."

Source: Sam Altman Blog

Product and Economic Predictions

DateClaimTypeTestable ByCurrent StatusSource
Sept 2024AI agents "doing real cognitive work" in 2025Essay2025Agents emerging but not yet transformativeThe Intelligence Age
Sept 2024Systems that can "figure out novel insights" by 2026Essay2026PendingSame
July 2025Customer support jobs "totally, totally gone"Federal Reserve conferenceOngoingKlarna reversed course on AI customer service; Cavell research predicts agent demand will grow through 2029CX Today
2024AI could replace 30-40% of jobs by 2030Interview2030PendingMIT Technology Review
Jan 2025"In 2025, we may see the first AI agents 'join the workforce' and materially change the output of companies"Blog post2025Mixed - Cal Newport's New Yorker piece argues agents "failed to live up to their hype"Reflections blog; Cal Newport analysis
2025GPT 5.2x level intelligence by end of 2027 for "at least 100x less" than current pricingRoadmap2027PendingFortune
2025GPT-6 in Q1 2026 - "timeline between GPT-5 and 6 would be much shorter than GPT-4 and 5"Press dinnerQ1 2026PendingYahoo Finance
2025AI will compress "10 years of scientific progress into a single year" within a few yearsInterview≈2028PendingTIME
2025GPT-5 "smarter than me" / "In many ways, GPT-5 is already smarter than me"ConferenceSubjectiveDifficult to verifyFortune

"Moore's Law for Everything" Predictions (2021)

PredictionTypeTestable ByStatusSource
AI could generate enough wealth to pay every US adult $13,500/year within 10 yearsEssay2031PendingMoore's Law for Everything
Everything (housing, education, food) becomes half as expensive every two yearsEssayOngoing❌ Not materializing for housing, healthcare, educationSame

UBI Study Results (2024): Altman-funded 3-year study by OpenResearch giving $1,000/month to 1,000 participants (with 2,000 person control group receiving $50/month) found payments had "virtually no impact" on "quality of employment" and didn't lead to greater upward mobility. (Gizmodo)


Self-Corrections and Walkbacks

DateOriginal ClaimCorrectionTypeSource
Pre-2020AI wouldn't be "a really great creative thinker"Acknowledged he was wrong after DALL-E, SoraSelf-correctionFortune
Dec 2024Weeks of AGI teasers leading up to o3 launch"Twitter hype is out of control again... We are not gonna deploy AGI next month, nor have we built it"WalkbackDecrypt
May 2023Threatened to leave Europe over AI Act: "We will try to comply, but if we can't comply we will cease operating"Later said "no plans to leave" and intends to cooperateWalkbackCNBC

Safety Risk Statements

DateQuoteTypeSource
2015"I think that AI will probably, most likely, sort of lead to the end of the world. But in the meantime, there will be great companies created with serious machine learning."ConferenceTom's Guide
2023The worst-case scenario is "lights out for all of us"PodcastLex Fridman Podcast
2023"I think that there's some chance of that [AI killing all humans]. And it's really important to acknowledge it"PodcastSame
2024"The road to AGI should be a giant power struggle"PodcastLex Fridman Podcast #419
2024-2025"AGI will probably hit sooner than most people think and it will matter much less"InterviewBloomberg

Pattern: Rhetoric shifted from "probably lead to end of world" (2015) → "lights out for all of us" (2023) → "will matter much less than people think" (2024-2025). For detailed analysis of this shift, see the main Sam Altman page.


Accuracy Analysis

Where Altman tends to be right:

  • General trajectory of AI importance and capabilities
  • AI capital requirements and infrastructure needs
  • Cost decline trajectory ("10x every 12 months")
  • Compute becoming precious commodity

Where Altman tends to be wrong:

  • Specific product timelines (self-driving 2015, GPT-5 launch)
  • Profitability assumptions (ChatGPT Pro)
  • Near-term transformation claims (agents in 2025)

Confidence calibration:

  • Vague language as hedge: Uses "few thousand days" (5.5-14 year range), "AGI as we have traditionally understood it" (undefined)
  • Moving goalposts: AGI framing shifted from "transformative event" to "will matter much less than people think"
  • Overoptimism on timelines: Self-driving (2015), specific product launches

Pattern: Directionally correct on AI's importance; consistently overoptimistic on specific timelines; rhetoric shifts from existential concern to dismissal as deployment continues.


Key Testable Claims to Watch

By 2025-2026:

  • Does OpenAI achieve anything resembling "AGI"?
  • Do AI agents transform the workforce as predicted?
  • Is GPT-6 released in Q1 2026?

By 2029-2030:

  • Does superintelligence arrive within "a few thousand days"?
  • Is 30-40% of work displaced?
  • Does scientific progress accelerate 10x?

By 2031:

  • Could AI-generated wealth fund $13,500/year per US adult?

Sources

Primary Sources

Secondary Sources

Related Pages

Top Related Pages

Other

Yann LeCun: Track Record