Skip to content
Longterm Wiki
Navigation
Updated 2026-01-29HistoryData
Page StatusContent
Edited 2 months ago1.8k words25 backlinksUpdated every 6 weeksOverdue by 21 days
40QualityAdequate •29.5ImportancePeripheral40ResearchLow
Content8/13
SummaryScheduleEntityEdit history1Overview
Tables14/ ~7Diagrams0/ ~1Int. links55/ ~14Ext. links0/ ~9Footnotes0/ ~5References17/ ~5Quotes0Accuracy0RatingsN:2 R:4.5 A:2 C:6Backlinks25
Change History1
Fix factual errors found in wiki audit7 weeks ago

Systematically audited ~35+ high-risk wiki pages for factual errors and hallucinations using parallel background agents plus direct reading. Fixed 13 confirmed errors across 11 files.

Issues2
QualityRated 40 but structure suggests 73 (underrated by 33 points)
StaleLast edited 66 days ago - may need review

Holden Karnofsky

Person

Holden Karnofsky

Holden Karnofsky directed $300M+ in AI safety funding through Coefficient Giving (formerly Open Philanthropy), growing the field from ~20 to 400+ FTE researchers and developing influential frameworks like the 'Most Important Century' thesis (15% transformative AI by 2036, 50% by 2060). His funding decisions include a $580M Anthropic investment and establishment of 15+ university AI safety programs.

AffiliationAnthropic
RoleMember of Technical Staff, Anthropic
Known ForDirecting billions toward AI safety, effective altruism leadership, AI timelines work
Related
People
Toby Ord
1.8k words · 25 backlinks

Overview

Holden Karnofsky was co-CEO of Coefficient Giving (formerly Open Philanthropy), the most influential grantmaker in AI safety and existential risk. Through Coefficient, he directed over $300 million toward AI safety research and governance, fundamentally transforming it from a fringe academic interest into a well-funded field with hundreds of researchers. In 2025, he joined Anthropic.

His strategic thinking has shaped how the effective altruism community prioritizes AI risk through frameworks like the "Most Important Century" thesis. This argues we may live in the century that determines humanity's entire future trajectory due to transformative AI development.

Funding AchievementAmountImpact
Total AI safety grants$300M+Enabled field growth from ~dozens to hundreds of researchers
Anthropic investment$580M+Created major safety-focused AI lab
Field building grants$50M+Established academic programs and research infrastructure

Risk Assessment

Risk CategoryKarnofsky's AssessmentEvidenceTimeline
Transformative AI~15% by 2036, ≈50% by 2060Bio anchors frameworkThis century
Existential importance"Most important century"AI could permanently shape humanity's trajectory2021-2100
TractabilityHigh enough for top priorityOpen Phil's largest focus area allocationCurrent
Funding adequacySeverely underfundedStill seeking to grow field substantiallyOngoing

Career Evolution and Major Achievements

Early Career (2007-2014): Building Effective Altruism

PeriodRoleKey Achievements
2007-2011Co-founder, GiveWellPioneered rigorous charity evaluation methodology
2011-2014Launch Coefficient GivingExpanded beyond global health to cause prioritization
2012-2014EA movement buildingHelped establish effective altruism as global movement

Transition to AI Focus (2014-2018)

Initial AI engagement:

  • 2014: First significant AI safety grants through Open Philanthropy (now Coefficient Giving)
  • 2016: Major funding to Center for Human-Compatible AI (CHAI)
  • 2017: Early OpenAI funding (before pivot to for-profit)
  • 2018: Increased conviction leading to AI as top priority

AI Safety Leadership (2018-2025)

Major funding decisions:

  • 2021: $580M investment in Anthropic to create safety-focused lab
  • 2022: Establishment of AI safety university programs
  • 2023: Expanded governance funding addressing AI regulation

Departure from Coefficient Giving and Impact (2023-2025)

Karnofsky's departure from Coefficient Giving in 2023 had significant ripple effects on the organization's AI safety work. Ajeya Cotra, who worked closely with Karnofsky for nine years, described losing "an incredibly engaged partner... someone who would read 30 pages of analysis and give you deep feedback." His departure removed a key source of intellectual partnership that had driven Coefficient's AI strategy, including the Bio Anchors framework and the organization's approach to technical AI safety grantmaking. Karnofsky subsequently joined Anthropic, where he continues working on AI safety from within a frontier lab.

Strategic Frameworks and Intellectual Contributions

The "Most Important Century" Thesis

Core argument structure:

ComponentClaimImplication
Technology potentialTransformative AI possible this centuryCould exceed agricultural/industrial revolution impacts
Speed differentialAI transition faster than historical precedentsLess time to adapt and coordinate
Leverage momentOur actions now shape outcomesUnlike past revolutions where individuals had little influence
ConclusionThis century uniquely importantJustifies enormous current investment

Supporting evidence:

  • Biological anchors methodology for AI timelines
  • Historical analysis of technological transitions
  • Economic modeling of AI impact potential

Bio Anchors Framework

Developed with Ajeya Cotra, this framework estimates AI development timelines by comparing required computation to biological systems:

Anchor TypeComputation EstimateTimeline Implication
Human brain≈10^15 FLOP/sMedium-term (2030s-2040s)
Human lifetime≈10^24 FLOPLonger-term (2040s-2050s)
Evolution≈10^41 FLOPMuch longer-term if needed

Coefficient Giving Funding Strategy

Portfolio Approach

Research AreaFunding FocusKey RecipientsRationale
Technical alignment$100M+Anthropic, Redwood ResearchDirect work on making AI systems safer
AI governance$80M+Center for Security and Emerging Technology, policy fellowshipsInstitutional responses to AI development
Field building$50M+University programs, individual researchersGrowing research community
Compute governance$20M+Compute monitoring researchOversight of AI development resources

Grantmaking Philosophy

Key principles:

  • Hits-based giving: Expect most grants to have limited impact, few to be transformative
  • Long time horizons: Patient capital for 5-10 year research projects
  • Active partnership: Strategic guidance beyond just funding
  • Portfolio diversification: Multiple approaches given uncertainty

Notable funding decisions:

  • Anthropic investment: $580M to create safety-focused competitor to OpenAI
  • MIRI funding: Early support for foundational AI alignment research
  • Policy fellowships: Placing AI safety researchers in government positions

Current Views and Assessment

Karnofsky's AI Risk Timeline

Based on public statements and Coefficient Giving priorities from 2023-2024, Karnofsky's views reflect a combination of timeline estimates derived from technical forecasting and strategic assessments about field readiness and policy urgency:

Expert/SourceEstimateReasoning
Transformative AI (2022)15% by 2036, 50% by 2060Derived from the bio anchors framework developed with Ajeya Cotra, which estimates AI development timelines by comparing required computation to biological systems. This central estimate suggests transformative AI is more likely than not within this century, though substantial uncertainty remains around both shorter and longer timelines.
Field adequacy (2024)Still severely underfundedDespite directing over $100M toward AI safety and growing the field from approximately 20 to 400+ FTE researchers, Coefficient Giving continues aggressive hiring and grantmaking. This assessment reflects the belief that the scale of the challenge—ensuring safe development of transformative AI—far exceeds current resources and talent devoted to it.
Policy urgency (2024)High priorityCoefficient has significantly increased governance focus, funding policy research, placing fellows in government positions, and supporting regulatory frameworks. This shift recognizes that technical alignment work alone is insufficient—institutional and policy responses are critical to managing AI development trajectories and preventing racing dynamics.

Evolution of Views (2020-2024)

YearKey UpdateReasoning
2021"Most Important Century" seriesCrystallized long-term strategic thinking
2022Increased policy focusRecognition of need for governance alongside technical work
2023Anthropic model successValidation of safety-focused lab approach
2024Accelerated timelines concernShorter timelines than bio anchors suggested

Influence on AI Safety Field

Field Growth Metrics

Metric20152024Growth Factor
FTE researchers≈20≈40020x
Annual funding<$5M>$200M40x
University programs015+New category
Major organizations2-320+7x

Institutional Impact

Academic legitimacy:

  • Funding enabled AI safety courses at major universities
  • Supported tenure-track positions focused on alignment research
  • Created pathway for traditional CS researchers to enter field

Policy influence:

  • Funded experts now advising US AI Safety Institute
  • Supported research informing EU AI Act
  • Built relationships between AI safety community and policymakers

Key Uncertainties and Strategic Cruxes

Open Questions in Karnofsky's Framework

UncertaintyStakesCurrent Evidence
AI timeline accuracyEntire strategy timingMixed signals from recent capabilities
Technical tractabilityFunding allocation efficiencyEarly positive results but limited validation
Governance effectivenessPolicy investment valueUnclear institutional responsiveness
Anthropic successLarge investment justificationStrong early results but long-term unknown

Strategic Disagreements

Within EA community:

  • Some argue for longtermist focus beyond AI
  • Others prefer global health and development emphasis
  • Debate over concentration vs. diversification of funding

With AI safety researchers:

  • Tension between technical alignment focus and governance approaches
  • Disagreement over open vs. closed development funding
  • Questions about emphasis on capabilities research safety benefits

Public Communication and Influence

Cold Takes Blog Impact

Most influential posts:

  • "The Most Important Century" series (>100k views)
  • "AI Timelines: Where the Arguments Stand" (policy reference)
  • "Bio Anchors" explanation (research methodology)

Communication approach:

  • Transparent reasoning and uncertainty acknowledgment
  • Accessible explanations of complex topics
  • Regular updates as views evolve
  • Direct engagement with critics and alternative viewpoints

Media and Policy Engagement

PlatformReachImpact
Congressional testimonyDirect policy influenceInformed AI regulation debate
Academic conferencesResearch communityShaped university AI safety programs
EA Global talksMovement directionInfluenced thousands of career decisions
Podcast interviewsPublic understandingMainstream exposure for AI safety ideas

Current Priorities and Future Direction

2024-2026 Strategic Focus

Immediate priorities:

  1. Anthropic scaling: Supporting responsible development of powerful systems
  2. Governance acceleration: Policy research and implementation support
  3. Technical diversification: Funding multiple alignment research approaches
  4. International coordination: Supporting global AI safety cooperation

Emerging areas:

  • Compute governance infrastructure
  • AI evaluation methodologies
  • Corporate AI safety practices
  • Prediction market applications

Long-term Vision

Field development goals:

  • Self-sustaining research ecosystem independent of Coefficient Giving
  • Government funding matching or exceeding philanthropic support
  • Integration of safety research into mainstream AI development
  • International coordination mechanisms for AI governance

Critiques and Responses

Common Criticisms

CriticismKarnofsky's ResponseCounter-evidence
Over-concentration of powerFunding diversification, transparencyMultiple other major funders emerging
Field capture riskPortfolio approach, external evaluationContinued criticism tolerated and addressed
Timeline overconfidenceExplicit uncertainty, range estimatesRegular updating based on new evidence
Governance skepticismMeasured expectations, multiple approachesEarly policy wins demonstrate tractability

Ongoing Debates

Resource allocation:

  • Should Coefficient Giving fund more basic research vs. applied safety work?
  • Optimal balance between technical and governance approaches?
  • Geographic distribution of funding (US-centric concerns)

Strategic approach:

  • Speed vs. care in scaling funding
  • Competition vs. cooperation with AI labs
  • Public advocacy vs. behind-the-scenes influence

Sources & Resources

Primary Sources

TypeSourceDescription
BlogCold TakesKarnofsky's strategic thinking and analysis
OrganizationCoefficient GivingGrant database and reasoning
ResearchBio Anchors ReportTechnical forecasting methodology
TestimonyCongressional HearingPolicy positions and recommendations

Secondary Analysis

TypeSourceFocus
AcademicEA ResearchCritical analysis of funding decisions
JournalisticMIT Technology ReviewExternal perspective on influence
PolicyRAND CorporationGovernment research on philanthropic AI funding
  • Dario Amodei - CEO of Anthropic, major funding recipient
  • Paul Christiano - Technical alignment researcher, influenced Karnofsky's views
  • Nick Bostrom - Author of "Superintelligence," early influence on Coefficient AI focus
  • Eliezer Yudkowsky - MIRI founder, recipient of early Coefficient AI safety grants

References

RAND Corporation is a nonprofit research organization providing objective analysis and policy recommendations across a wide range of topics including national security, technology, governance, and emerging risks. It produces influential studies on AI policy, cybersecurity, and global governance challenges. RAND's work is frequently cited by governments and policymakers worldwide.

★★★★☆

Holden Karnofsky's 'Most Important Century' series argues that 21st-century AI development could trigger a productivity explosion leading to a galaxy-wide civilization far sooner than expected, making current decisions uniquely consequential for long-run human welfare. The series synthesizes arguments about AI timelines, transformative risk, and the moral weight of shaping humanity's long-term trajectory.

★★★☆☆
3EU AI Act – Official Resource Hubartificialintelligenceact.eu

The EU AI Act is the world's first comprehensive legal framework for artificial intelligence, establishing a risk-based classification system for AI applications. It imposes varying obligations on developers and deployers depending on the risk level of their AI systems, from minimal-risk to unacceptable-risk categories. The act sets precedents for global AI governance and compliance requirements.

MIT Technology Review is a major science and technology journalism outlet covering AI, biotechnology, climate, and emerging technologies. It publishes in-depth reporting, analysis, and magazine features on the societal implications of technology. The current title referencing 'Deepfake Coverage' does not match the general homepage content retrieved.

★★★★☆

This page appears to be a 404 error, meaning the original announcement about Open Philanthropy's investment in Anthropic is no longer accessible at this URL. The content that would have described this funding relationship between Anthropic and the effective altruism-aligned philanthropic organization is unavailable.

★★★★☆
6Congressional HearingUS Congress·Government

Congress.gov is the official U.S. government portal for tracking federal legislation, committee hearings, floor activities, and public laws. It provides access to bills, nominations, treaties, and congressional records for both the House and Senate. This resource serves as a primary reference for monitoring AI-related legislation and governance activities in the U.S. Congress.

★★★★★
7AI safety university programsCoefficient Giving

This URL originally pointed to an Open Philanthropy grant page for AI safety via market incentives, but now redirects to Coefficient Giving (the rebranded Open Philanthropy). The page reflects the organization's philanthropic focus areas including 'Navigating Transformative AI' to ensure AI is safe and well-governed.

★★★★☆

A layperson-friendly summary of Ajeya Cotra's 'Biological Anchors' framework for forecasting when transformative AI (specifically, AI that can automate all human activities driving scientific progress) might be developed. The method estimates training compute costs by anchoring to the human brain's scale, projecting when such training will become affordable. From nearly all modeled scenarios, it assigns high probability to transformative AI arriving this century.

★★★☆☆

Biographical profile of Ajeya Cotra, a researcher at Coefficient Giving (formerly Open Philanthropy) focused on forecasting AI's trajectory and ensuring its impact is positive. She previously led the technical AI safety program area and has worked on AI timelines, worldview diversification in budget allocation, and global catastrophic risk prioritization. She holds a B.S. in EECS from UC Berkeley.

★★★★☆

Cold Takes is Holden Karnofsky's (co-CEO of Open Philanthropy) personal blog exploring big-picture questions about AI, existential risk, effective altruism, and how to think about the most important challenges of our time. It features in-depth essays on AI timelines, transformative AI scenarios, and philanthropic strategy. The blog is notable for its 'Most Important Century' series arguing that we may be living at a uniquely pivotal moment in history.

★★★☆☆

GiveWell is a nonprofit charity evaluator that researches and recommends highly effective giving opportunities, focusing on evidence-based interventions with strong cost-effectiveness. It conducts in-depth analysis of charities to identify where donations can do the most good, primarily in global health and poverty. GiveWell exemplifies the effective altruism methodology of rigorous expected-value reasoning applied to philanthropic decisions.

A directory listing academic programs, courses, and educational resources focused on AI safety. The page serves as a centralized hub for individuals seeking formal or structured training in AI safety topics, ranging from technical alignment to governance.

Holden Karnofsky's Cold Takes post synthesizes arguments for why transformative AI may arrive sooner than the Biological Anchors framework suggested, reviewing both expert surveys and key lines of reasoning. It examines why some analysts believe we are on a faster trajectory toward powerful AI systems than mainstream estimates indicate.

★★★☆☆
14EA Forum Career PostsEA Forum·Blog post

The Effective Altruism Forum serves as a community hub for discussing careers, cause prioritization, and field-building within the EA and AI safety ecosystem. It hosts posts on career transitions into high-impact roles, including AI safety research, policy, and governance positions. The forum aggregates community thinking on how individuals can best contribute to reducing existential risks.

★★★☆☆

Open Philanthropy is a major philanthropic organization that funds work across global health, AI safety, biosecurity, and other cause areas. Their grants database provides transparency into which organizations and research directions receive funding. They are one of the largest funders of AI safety and existential risk research.

★★★★☆
16CSET: AI Market DynamicsCSET Georgetown

CSET (Center for Security and Emerging Technology) at Georgetown University is a policy research organization focused on the security implications of emerging technologies, particularly AI. It produces research on AI policy, workforce, geopolitics, and governance. The content could not be fully extracted, limiting detailed analysis.

★★★★☆

This Open Philanthropy report by Ajeya Cotra uses 'biological anchors' to estimate timelines for transformative AI, grounding forecasts in the computational resources required to match the human brain's training process. It synthesizes estimates across multiple hypotheses about the relevant scale of computation and projects probability distributions over when such compute might become affordable. The report became a foundational reference for AI timeline discussions in the EA and AI safety communities.

★★★★☆

Structured Data

13 facts·3 recordsView in FactBase →
Employed By
Anthropic
as of Jan 2025
Role / Title
Member of Technical Staff, Anthropic
as of Jan 2025
Birth Year
1981

All Facts

13
People
PropertyValueAs OfSource
Employed ByAnthropicJan 2025
2 earlier values
Jan 2014Open Philanthropy
Jan 2007GiveWell
Role / TitleMember of Technical Staff, AnthropicJan 2025
2 earlier values
Jan 2017Co-CEO, Open Philanthropy
Jan 2007Co-founder & Co-Executive Director, GiveWell
Biographical
PropertyValueAs OfSource
EducationBA in Social Studies, Harvard University
Notable ForCo-founder of GiveWell and Open Philanthropy; influential figure in effective altruism; author of 'Most Important Century' blog series on transformative AI
Social Media@HoldenKarnofsky
Wikipediahttps://en.wikipedia.org/wiki/Holden_Karnofsky
Birth Year1981
General
PropertyValueAs OfSource
Websitehttps://www.cold-takes.com
Other
PropertyValueAs OfSource
Board MemberRedwood Research

Career History

3
OrganizationTitleStartEnd
AnthropicMember of Technical Staff2025-01
Open PhilanthropyCo-CEO2014-012024
GiveWellCo-founder & Co-Executive Director2007-012014

Related Wiki Pages

Top Related Pages

Organizations

Machine Intelligence Research InstituteUS AI Safety Institute

Other

Toby OrdAnthropic StakeholdersRecoding America

Approaches

AI AlignmentEliciting Latent Knowledge (ELK)AI EvaluationCompute Monitoring

Concepts

International Coordination MechanismsLong-Timelines Technical WorldviewState Capacity and AI Governance

Risks

AI-Driven Concentration of PowerAI Development Racing Dynamics

Key Debates

Why Alignment Might Be HardAI Alignment Research AgendasOpen vs Closed Source AI

Analysis

Anthropic Valuation AnalysisAnthropic (Funder)

Policy

EU AI Act