Skip to content
Longterm Wiki
PC

Paul Christiano

Also known as: Paul

Pioneer of RLHF and AI alignment research; founder of Alignment Research Center (ARC); key theorist of iterated amplification and eliciting latent knowledge

Current Role
Head of AI Safety, US AI Safety Institute
Organization
US AI Safety Institute
Born
1992
Age ~34

Expert Positions12 topics

TopicViewEstimateConfidenceDateSource
AGI TimelinesMedium2035-2045low2023
P(doom)Significant10-20%medium2023
How Hard Is Alignment?Hard but tractable50%medium2023
Current Approaches ScaleUncertain40%medium2023ARC Research (2023)
Inner Alignment SolvabilityHard but tractableSolvable with sufficient investmentmedium2023
Likelihood of Deceptive AlignmentSignificant concern50%medium2023
P(AI X-Risk This Century)Moderate~20-50%medium2023
Would Misalignment Be Catastrophic?Uncertain, depends on scenario20-40% → catastrophiclow2023
P(AI Catastrophe)Significant10-20%medium2023Various posts (2022-2023)
Takeoff SpeedSlow5-15 yearsmedium2023
Will Advanced AI Be Deceptive?Possibly detectable40%medium2023
Will We Get Adequate Warning?Likely70%medium2023

Organization Roles2

Head of ARC / Advisor
Apr 2021 – present
AI ImpactsFounderCurrent
Co-founder
2014 – present

Education

PhD in Computer Science, UC Berkeley; BS in Mathematics, MIT

From wiki articleRead full article →

Overview

Paul Christiano is one of the most influential researchers in AI alignment, known for developing concrete, empirically testable approaches to the alignment problem. With a PhD in theoretical computer science from UC Berkeley, he has worked at OpenAI, DeepMind, and founded the Alignment Research Center (ARC).

Christiano pioneered the "prosaic alignment" approach—aligning AI without requiring exotic theoretical breakthroughs. His current risk assessment places ~10-20% probability on existential risk from AI this century, with AGI arrival in the 2030s-2040s. His work has directly influenced alignment research programs at major labs including OpenAI, Anthropic, and DeepMind.

Risk Assessment

Risk FactorChristiano's AssessmentEvidence/ReasoningComparison to Field
P(doom)≈10-20%Alignment tractable but challengingModerate (vs 50%+ doomers, <5% optimists)
AGI Timeline2030s-2040sGradual capability increaseMainstream range
Alignment DifficultyHard but tractableIterative progress possibleMore optimistic than MIRI
Coordination FeasibilityModerately optimisticLabs have incentives to cooperateMore optimistic than average

Links

Facts

10
People
Role / TitleHead of AI Safety, US AI Safety Institute
General
Websitehttps://paulfchristiano.com
Other
Board MemberRedwood Research
Biographical
EducationPhD in Computer Science, UC Berkeley; BS in Mathematics, MIT
Notable ForPioneer of RLHF and AI alignment research; founder of Alignment Research Center (ARC); key theorist of iterated amplification and eliciting latent knowledge
Social Media@paulfchristiano
Wikipediahttps://en.wikipedia.org/wiki/Paul_Christiano
Google Scholarhttps://scholar.google.com/citations?user=6gHkYDgAAAAJ
Birth Year1992