AI Safety Field Building Analysis
AI Safety Field Building Analysis
Comprehensive analysis of AI safety field-building showing growth from 400 to 1,100 FTEs (2022-2025) at 21-30% annual growth rates, with training programs achieving 37% career conversion at costs of $5,000-40,000 per career change. Identifies critical bottleneck: talent pipeline over-optimized for researchers while neglecting operations, policy, and organizational roles.
Quick Assessment
| Dimension | Assessment | Evidence |
|---|---|---|
| Field Size (2025) | 1,100 FTEs (600 technical, 500 non-technical) | AI Safety Field Growth Analysis 2025↗🔗 web★★★☆☆EA ForumAI Safety Field Growth Analysis 2025A useful reference for researchers and funders assessing the scale and growth rate of the AI safety field; numbers should be cross-checked against other workforce surveys as FTE estimates in this field vary significantly by methodology.Stephen McAleese (2025)77 karma · 13 commentsA longitudinal study tracking the growth of the AI safety field from 2010 to 2025, documenting expansion from approximately 400 to 1,100 full-time equivalent researchers across ...ai-safetyfield-buildinggovernancetechnical-safety+3Source ↗ |
| Annual Growth Rate | 21-30% since 2020 | Technical: 21% FTE growth; Non-technical: 30% |
| Total Philanthropic Funding | $110-130M/year (2024) | Overview of AI Safety Funding↗🔗 web★★★☆☆EA ForumAn Overview of the AI Safety Funding SituationUseful reference for understanding the financial infrastructure of AI safety research as of mid-2023, particularly relevant given the post-FTX collapse reshaping of the funding landscape.Stephen McAleese (2023)142 karma · 15 commentsA comprehensive survey of the AI safety funding landscape as of mid-2023, cataloging major philanthropic sources including Open Philanthropy, the FTX Future Fund, and the Long-T...ai-safetyfield-buildingcoordinationgovernance+1Source ↗ |
| Training Program Conversion | 37% work full-time in AI safety | BlueDot 2022 Cohort Analysis↗🔗 webBlueDot 2022 Cohort AnalysisUseful empirical reference for evaluating the effectiveness of AI safety education and talent pipeline programs; provides concrete outcome metrics for a structured cohort-based course intervention aimed at growing the AI safety field.BlueDot Impact's analysis of their 2022 AI Safety Fundamentals: Alignment course shows it increased the proportion of participants working on AI safety from 5% to 37% (342 parti...ai-safetyfield-buildingtraining-programsalignment+3Source ↗ |
| Cost per Career Change | $5,000-40,000 depending on program | ARENA lower-touch, MATS higher-touch |
| Key Bottleneck | Talent pipeline over-optimized for researchers | EA Forum analysis↗🔗 web★★★☆☆EA ForumAI Safety’s Talent Pipeline is Over-optimised for ResearchersA 2025 EA Forum post by Chris Clay offering a structural critique of AI safety community-building, relevant to anyone thinking about career pathways, talent strategy, or ecosystem coordination within AI safety.Chris Clay🔸 (2025)117 karma · 15 commentsThis EA Forum post argues that AI safety's talent pipeline is structurally biased toward producing researchers, despite leadership consensus that research is not the most neglec...ai-safetyfield-buildingcoordinationpolicy+3Source ↗ |
| Tractability | Medium-High | Programs show measurable outcomes |
Overview
Field-building focuses on growing the AI safety ecosystem rather than doing direct research or policy work. The theory is that by increasing the number and quality of people working on AI safety, we multiply the impact of all other interventions.
This is a meta-level or capacity-building intervention—it doesn't directly solve the technical or governance problems, but creates the infrastructure and talent pipeline that makes solving them possible.
The field has grown substantially: from approximately 400 full-time equivalents (FTEs) in 2022 to roughly 1,100 FTEs in 2025, with technical AI safety organizations growing at 24% annually and non-technical organizations at approximately 30% annually. However, this growth has created new challenges—the pipeline may be over-optimized for researchers while neglecting operations, policy, and other critical roles.
Theory of Change
Diagram (loading…)
flowchart TD FB[Field-Building Programs] --> TP[Talent Pipeline] FB --> KD[Knowledge Dissemination] FB --> CB[Community Building] FB --> FI[Funding Infrastructure] TP --> RC[Research Capacity] KD --> RC CB --> RC FI --> RC RC --> TR[Technical Research] RC --> GP[Governance & Policy] RC --> CI[Corporate Influence] TR --> RR[Reduced Risk] GP --> RR CI --> RR style FB fill:#e6f3ff style RC fill:#fff3e6 style RR fill:#e6ffe6
Key mechanisms:
- Talent pipeline: Train and recruit people into AI safety
- Knowledge dissemination: Spread ideas and frameworks
- Community building: Create support structures and networks
- Funding infrastructure: Direct resources to promising work
- Public awareness: Build broader support and understanding
Major Approaches
1. Education and Training Programs
Goal: Teach AI safety concepts and skills to potential contributors.
Training Program Comparison
| Program | Format | Duration | Scale | Cost/Participant | Placement Rate | Key Outcomes |
|---|---|---|---|---|---|---|
| MATS↗🔗 webMATS Research ProgramMATS is one of the primary talent pipelines into the AI safety field; wiki users interested in career transitions or field-building efforts should consider this a key institutional reference.MATS is an intensive fellowship program designed to help researchers transition into AI safety careers, offering structured mentorship from leading researchers, stipends, and co...ai-safetyalignmentfield-buildingeducational+4Source ↗ | Research mentorship | 3-4 months | 30-50/cohort | ≈$20,000-40,000 | 75% publish results | Alumni at Anthropic, OpenAI, DeepMind; founded Apollo Research, Timaeus |
| ARENA↗🔗 webARENA – AI Safety EducationARENA is a widely recommended starting point for technically skilled individuals looking to enter AI safety research, and its curriculum materials are frequently referenced in AI safety onboarding guides and reading lists.ARENA (Alignment Research Engineer Accelerator) is an educational program designed to train technical AI safety researchers by providing structured curriculum covering mechanist...ai-safetyalignmentinterpretabilitytechnical-safety+5Source ↗ | In-person bootcamp | 4-5 weeks | 20-30/cohort | ≈$5,000-15,000 | 8 confirmed FT positions (5.0 cohort) | Alumni at Apollo Research, METR, UK AISI |
| BlueDot Impact↗🔗 webBlueDot Impact – AI Safety Education & Field-BuildingBlueDot Impact is a key field-building organization in the AI safety ecosystem, relevant for anyone seeking structured entry points into AI safety or governance careers.BlueDot Impact is an organization focused on building the AI safety field through structured educational programs and courses. It offers cohort-based training programs designed ...ai-safetyfield-buildingtraining-programsgovernance+4Source ↗ | Online cohort-based | 8 weeks | 1,000+/year | ≈$440/student | 37% work FT in AI safety | 6,000+ trained since 2022; 75% completion rate |
| SPAR↗🔗 webSPAR - Research Program for AI RisksSPAR is a key entry-level program for those seeking to break into AI safety research; relevant for wiki users looking for mentorship opportunities or field-building initiatives in the AI safety community.SPAR (Supervised Program for Alignment Research) is a structured mentorship program that pairs aspiring researchers with experienced AI safety professionals to conduct research ...ai-safetyalignmentgovernancefield-building+4Source ↗ | Part-time remote | Varies | 50+/cohort | Low (volunteer mentors) | Research output focused | Connects aspiring researchers with professionals |
| AI Safety Camp | Project-based | 1-2 weeks | 20-40/camp | Varies | Project completion | Multiple camps globally |
Key Programs in Detail:
MATS (ML Alignment & Theory Scholars)↗🔗 webMATS Research ProgramMATS is one of the primary talent pipelines into the AI safety field; wiki users interested in career transitions or field-building efforts should consider this a key institutional reference.MATS is an intensive fellowship program designed to help researchers transition into AI safety careers, offering structured mentorship from leading researchers, stipends, and co...ai-safetyalignmentfield-buildingeducational+4Source ↗:
- Since 2021, has supported 298 scholars and 75 mentors
- Summer 2024: 1,220 applicants, 3-5% acceptance rate (comparable to MIT admissions)
- Spring 2024 Extension: 75% of scholars published results; 57% accepted to conferences
- Notable: Nina Panickssery's paper on steering Llama 2 won Outstanding Paper Award at ACL 2024
- Alumni include researchers at Anthropic, OpenAI, and Google DeepMind
- Received $23.6M in Coefficient Giving funding↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy Grant to MATS: Research AI Safety Research ExpensesThis grant record documents Open Philanthropy's financial support for MATS, a key AI safety training program; the scraped content appears to be from a different site, limiting available detail about the grant specifics.This page references an Open Philanthropy grant to MATS (Machine Learning Alignment Theory Scholars) program to cover AI safety research expenses. MATS is a field-building initi...ai-safetyfield-buildingtraining-programsgovernance+2Source ↗ for general support
ARENA (Alignment Research Engineer Accelerator)↗🔗 webARENA – AI Safety EducationARENA is a widely recommended starting point for technically skilled individuals looking to enter AI safety research, and its curriculum materials are frequently referenced in AI safety onboarding guides and reading lists.ARENA (Alignment Research Engineer Accelerator) is an educational program designed to train technical AI safety researchers by providing structured curriculum covering mechanist...ai-safetyalignmentinterpretabilitytechnical-safety+5Source ↗:
- Run 2-3 bootcamps per year, each 4-5 weeks, based at LISA in London
- ARENA 5.0↗✏️ blog★★★☆☆LessWrongARENA 5.0 Impact ReportARENA is a recurring in-person upskilling program for AI safety; this impact report documents outcomes from the fifth cohort and is useful for those evaluating talent pipeline and field-building initiatives in the AI safety ecosystem.JScriven, JamesH, James Fox (2025)25 karma · 0 commentsARENA 5.0 is a 4-week intensive in-person program that upskills technically talented individuals for AI safety work, covering mechanistic interpretability, reinforcement learnin...ai-safetyfield-buildingtraining-programsinterpretability+4Source ↗: 8 participants confirmed full-time AI safety positions post-program
- Participants rate exercise enjoyment 8.7/10, LISA location value 9.6/10
- Alumni quote: "ARENA was the most useful thing that could happen to someone with a mathematical background who wants to enter technical AI safety research"
- Claims to be among most cost-effective technical AI safety training programs
BlueDot Impact↗🔗 webBlueDot Impact – AI Safety Education & Field-BuildingBlueDot Impact is a key field-building organization in the AI safety ecosystem, relevant for anyone seeking structured entry points into AI safety or governance careers.BlueDot Impact is an organization focused on building the AI safety field through structured educational programs and courses. It offers cohort-based training programs designed ...ai-safetyfield-buildingtraining-programsgovernance+4Source ↗ (formerly AI Safety Fundamentals):
- Trained 6,000+ professionals worldwide since 2022
- 2022 cohort analysis↗🔗 webBlueDot 2022 Cohort AnalysisUseful empirical reference for evaluating the effectiveness of AI safety education and talent pipeline programs; provides concrete outcome metrics for a structured cohort-based course intervention aimed at growing the AI safety field.BlueDot Impact's analysis of their 2022 AI Safety Fundamentals: Alignment course shows it increased the proportion of participants working on AI safety from 5% to 37% (342 parti...ai-safetyfield-buildingtraining-programsalignment+3Source ↗: 123 alumni (37% of 342) now work full-time on AI safety
- 20 alumni would not be working on AI safety were it not for the course (counterfactual impact)
- 75% completion rate (vs. 20% for typical Coursera courses)
- Raised $34M total funding, including $25M in 2025
- Alumni at Anthropic, Google DeepMind, UK AI Security Institute
Theory of change: Train people in AI safety → some pursue careers → net increase in research capacity
Effectiveness considerations:
- High leverage: One good researcher can contribute for decades
- Measurable conversion: BlueDot shows 37% career conversion; ARENA shows 8+ direct placements per cohort
- Counterfactual question: BlueDot estimates 20 counterfactual career changes from 2022 cohort
- Quality vs. quantity: More selective programs (MATS, ARENA) show higher placement rates
Cost Per Career Change Estimates
Training programs vary significantly in their cost-effectiveness at converting participants into AI safety careers. Different program models—from high-touch research mentorships to scalable online courses—represent different trade-offs between cost per participant and career conversion rate.
| Expert/Source | Estimate | Reasoning |
|---|---|---|
| ARENA (successful cases) | $1,000-15,000 | ARENA represents the lower bound for intensive programs, achieving direct program costs per successful career change through its efficient 4-5 week bootcamp format. The program's in-person structure at LISA combined with focused technical curriculum allows for cost-effective training, with ARENA 5.0 placing 8 participants in full-time AI safety positions. The cost includes venue, materials, and instructor time but benefits from concentrated delivery and high placement rates among participants who complete the program. |
| MATS | $10,000-40,000 | MATS represents a higher-touch research mentorship model with significantly higher costs per career change, reflecting its 3-4 month duration and personalized 1-on-1 mentorship structure. The program's selectivity (3-5% acceptance rate) and focus on research output—with 75% of Spring 2024 scholars publishing results—justifies higher per-participant investment. Costs include mentor compensation, scholar stipends, and program infrastructure, with the model optimized for producing research-ready talent rather than maximizing conversion volume. |
| BlueDot Impact | $140-2,000 | BlueDot Impact achieves the lowest cost per career change through its scalable online cohort model, training 1,000+ participants annually at approximately $140 per student. The 37% career conversion rate from the 2022 cohort (123 of 342 alumni working full-time in AI safety) yields an estimated $1,200-2,000 cost per successful career change when accounting for program overhead. The model sacrifices depth for scale but maintains 75% completion rates—far higher than typical MOOCs—through cohort-based structure and volunteer facilitators. |
Who's doing this:
- ARENA (Redwood Research / independent)
- MATS (independent, Lightcone funding)
- BlueDot Impact
- Various university courses and programs
2. Public Communication and Awareness
Goal: Increase general understanding of AI risk and build support for safety efforts.
Approaches:
Popular Media:
- Podcasts (Lex Fridman, Dwarkesh Patel, 80K Hours)
- Books (Superintelligence, The Alignment Problem, The Precipice)
- Documentaries and videos
- News articles and op-eds
- Social media presence
High-Level Engagement:
- Statement on AI Risk (May 2023): Geoffrey Hinton, Yoshua Bengio, Demis Hassabis, Sam Altman, Dario Amodei signed
- "Mitigating the risk of extinction from AI should be a global priority"
- Raised public and elite awareness
- Expert testimony to governments
- Academic conferences and workshops
- Industry events and presentations
Accessible Explanations:
- Robert Miles YouTube channel
- AI Safety memes and infographics
- Explainer articles
- University lectures and courses
Theory of change: Awareness → political will for governance + cultural shift toward safety + talent recruitment
Effectiveness:
- Uncertain impact on x-risk: Unclear if awareness translates to action
- Possible downsides:
- AI hype and race dynamics
- Association with less credible narratives
- Backlash and polarization
- Possible upsides:
- Political support for regulation
- Recruitment to field
- Cultural shift in labs
Who's doing this:
- Individual communicators (Miles, Yudkowsky, Christiano, etc.)
- Organizations (CAIS, FLI)
- Journalists covering AI
- Academics doing public engagement
3. Funding and Grantmaking
Goal: Direct resources to high-impact work and people.
AI Safety Funding Landscape (2024)
| Funding Source | Amount (2024) | % of Total | Key Recipients |
|---|---|---|---|
| Coefficient Giving | ≈$63.6M | 49% | CAIS ($8.5M), Redwood ($6.2M), MIRI ($4.1M) |
| Individual Donors (e.g., Jaan Tallinn) | ≈$20M | 15% | Various orgs and researchers |
| Government Funding | ≈$32.4M | 25% | AI Safety Institutes, university research |
| Corporate External Investment | ≈$8.2M | 6% | Frontier Model Forum AI Safety Fund |
| Academic Endowments | ≈$6.8M | 5% | University centers |
| Total Philanthropic | $110-130M | 100% | — |
Source: Overview of AI Safety Funding Situation↗🔗 web★★★☆☆EA ForumAn Overview of the AI Safety Funding SituationUseful reference for understanding the financial infrastructure of AI safety research as of mid-2023, particularly relevant given the post-FTX collapse reshaping of the funding landscape.Stephen McAleese (2023)142 karma · 15 commentsA comprehensive survey of the AI safety funding landscape as of mid-2023, cataloging major philanthropic sources including Open Philanthropy, the FTX Future Fund, and the Long-T...ai-safetyfield-buildingcoordinationgovernance+1Source ↗
Note: This excludes internal corporate safety research budgets, estimated at greater than $500M annually across major AI labs. Total ecosystem funding including corporate is approximately $600-650M/year.
Context: Philanthropic funding for climate risk mitigation was approximately $9-15 billion in 2023—roughly 20x philanthropic AI safety funding. With over $189 billion invested in AI projected for 2024, safety funding remains less than 2% of total AI investment.
Major Funders:
Coefficient Giving↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy grants databaseOpen Philanthropy is one of the most influential funders in AI safety; their grants database is a useful reference for understanding which organizations and research directions receive major philanthropic support.Open Philanthropy is a major philanthropic organization that funds work across global health, AI safety, biosecurity, and other cause areas. Their grants database provides trans...ai-safetyexistential-riskgovernancecoordination+3Source ↗:
- Largest AI safety funder (≈$50-65M/year to technical AI safety)
- 2025 Technical AI Safety RFP↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy Request for Proposals: Technical AI Safety ResearchThis Open Philanthropy RFP is a key funding opportunity document that shaped the direction of technical AI safety research by publicly identifying priority areas; useful context for understanding how philanthropic funding influences the field.Open Philanthropy issued a request for proposals seeking technical AI safety research projects, signaling funding priorities and research directions the organization considers m...ai-safetytechnical-safetyinterpretabilityscalable-oversight+4Source ↗: Expected to spend ≈$40M over 5 months
- Key 2024-25 grants: MATS ($23.6M), CAIS ($8.5M), Redwood Research ($6.2M)
- Self-assessment: "Rate of spending was too slow" in 2024; committed to expanding support
- Supporting work on AI safety since 2015
AI Safety Fund (Frontier Model Forum)↗🔗 web★★★☆☆Frontier Model ForumAI Safety Fund (AISF) – Frontier Model ForumThis page describes an industry-funded grantmaking initiative; useful for understanding how major AI labs are collectively funding external safety research and who the grantees are.The AI Safety Fund (AISF) is a $10 million+ collaborative initiative launched in October 2023 by Anthropic, Google, Microsoft, and OpenAI (via the Frontier Model Forum) along wi...ai-safetygovernanceevaluationfield-building+5Source ↗:
- $10M+ collaborative initiative established October 2023
- Founding members: Anthropic, Google, Microsoft, OpenAI
- Philanthropic partners: Patrick J. McGovern Foundation, Packard Foundation, Schmidt Sciences, Jaan Tallinn
Survival and Flourishing Fund (SFF):
- ≈$30-50M/year
- Broad AI safety focus
- Supports unconventional projects
- Smaller grants, more experimental
Effective Altruism Funds (Long-Term Future Fund):
- ≈$10-20M/year to AI safety
- Small to medium grants
- Individual researchers and projects
- Lower bar for experimental work
Grantmaking Strategies:
Hits-based giving:
- Accept high failure rate for potential breakthroughs
- Fund unconventional approaches
- Support early-stage ideas
Ecosystem development:
- Fund infrastructure (ARENA, MATS, etc.)
- Support conferences and gatherings
- Build community spaces
Diversification:
- Support multiple approaches
- Don't cluster too heavily
- Hedge uncertainty
Theory of change: Capital → enables people and orgs to work on AI safety → research and policy progress
Bottlenecks:
- Talent exceeds funding for roles, but not for orgs: Plenty of aspiring researchers but not enough organizations to hire them↗🔗 web★★★☆☆EA ForumAI Safety’s Talent Pipeline is Over-optimised for ResearchersA 2025 EA Forum post by Chris Clay offering a structural critique of AI safety community-building, relevant to anyone thinking about career pathways, talent strategy, or ecosystem coordination within AI safety.Chris Clay🔸 (2025)117 karma · 15 commentsThis EA Forum post argues that AI safety's talent pipeline is structurally biased toward producing researchers, despite leadership consensus that research is not the most neglec...ai-safetyfield-buildingcoordinationpolicy+3Source ↗
- Grantmaker capacity: Coefficient Giving struggled to make qualified senior hires↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy: Progress in 2024 and Plans for 2025Open Philanthropy is one of the largest funders in the AI safety space; their annual progress reports are useful for understanding funding priorities, institutional strategy, and which organizations and research directions receive major philanthropic backing.Open Philanthropy reviews its 2024 philanthropic activities and outlines priorities for 2025, with emphasis on AI safety research funding, strategic partnerships, and grants spa...ai-safetyexistential-riskfield-buildinggovernance+2Source ↗ for technical AI safety grantmaking
- Competition with labs: AI Safety Institutes and external research struggle to compete on compensation with frontier labs
Who should consider this:
- Program officers at foundations
- Individual donors with wealth
- Fund managers
- Requires: wealth or institutional position + good judgment + network
4. Community Building and Support
Goal: Create infrastructure that supports AI safety work.
Activities:
Gatherings and Conferences:
- EA Global (AI safety track)
- AI Safety conferences
- Workshops and retreats
- Local meetups
- Online forums (Alignment Forum, LessWrong, Discord servers)
Career Support:
- 80,000 Hours career advising
- Mentorship programs
- Job boards and hiring pipelines
- Introductions and networking
Research Infrastructure:
- Alignment Forum (discussion platform)
- ArXiv overlays and aggregation
- Compute access programs
- Shared datasets and benchmarks
Emotional and Social Support:
- Community spaces
- Mental health resources
- Peer support for difficult work
- Social events
Theory of change: Supportive community → people stay in field longer → more cumulative impact + better mental health
Challenges:
- Insularity: Echo chambers and groupthink
- Barrier to entry: Can feel cliquish to newcomers
- Time investment: Social events vs. object-level work
- Ideological narrowness: Lack of diversity in perspectives
Who's doing this:
- CEA (Centre for Effective Altruism)
- Local EA groups
- Lightcone Infrastructure (LessWrong, Alignment Forum)
- Individual organizers
5. Academic Field Building
Goal: Establish AI safety as legitimate academic field.
University Centers and Programs:
| Institution | Center/Program | Focus | Status |
|---|---|---|---|
| UC Berkeley | CHAI↗🔗 webCenter for Human-Compatible AICHAI is one of the leading academic institutions focused on AI alignment research, founded by Stuart Russell (author of 'Human Compatible'); its homepage provides an overview of ongoing projects, researchers, and publications central to the field.CHAI is a UC Berkeley research center dedicated to reorienting AI development toward systems that are provably beneficial and aligned with human values. It conducts technical an...ai-safetyalignmenttechnical-safetygovernance+3Source ↗ (Center for Human-Compatible AI) | Foundational alignment research | Active |
| Oxford | Future of Humanity Institute | Existential risk research | Closed 2024 |
| MIT | AI Safety Initiative | Technical safety, governance | Growing |
| Stanford | HAI (Human-Centered AI) | Broad AI policy, some safety | Active |
| Carnegie Mellon | AI Safety Research | Technical safety | Active |
| Cambridge | LCFI, CSER | Existential risk, policy | Active |
Key Developments (2024-2025):
- FHI closure at Oxford marks significant shift in academic landscape
- Growing number of PhD programs with explicit AI safety focus
- NSF and other agencies beginning to fund safety research specifically
- Coefficient Giving funding university-based safety research↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy funding university-based safety researchThis grants database is a key reference for understanding the institutional funding landscape of AI safety research; useful for researchers seeking funding context, field historians, and those mapping the ecosystem of organizations working on AI safety.A searchable database of Open Philanthropy grants related to AI safety, providing transparency into one of the field's largest funders. It documents funding awarded to universit...ai-safetyalignmentgovernancefield-building+4Source ↗ including Ohio State
Academic Incentives:
- Tenure-track positions in AI safety emerging
- PhD programs with safety focus
- Grants for safety research (NSF, etc.)
- Prestigious publication venues (NeurIPS safety track, ICLR)
- Academic conferences (AI Safety research conferences)
Curriculum Development:
- AI safety courses at major universities
- 80,000 Hours technical AI safety upskilling resources↗🔗 web★★★☆☆80,000 Hours80,000 Hours technical AI safety upskilling resourcesPublished by 80,000 Hours in June 2025, this resource is aimed at career-changers and students seeking structured pathways into technical AI safety research; useful as a starting point for newcomers to the field.A curated guide from 80,000 Hours providing resources for individuals looking to develop technical skills relevant to AI safety research. It aggregates learning materials, cours...ai-safetytechnical-safetyfield-buildingalignment+4Source ↗
- Integration into CS curriculum slowly increasing
Challenges:
- Slow timelines: Academic careers are 5-10 year investments
- Misaligned incentives: Publish or perish vs. impact
- Capabilities research: Universities also advance capabilities
- Brain drain: Best people leave for industry/nonprofits (frontier labs pay 2-5x academic salaries)
Benefits:
- Legitimacy: Academic credibility helps policy
- Training: PhD pipeline
- Long-term research: Can work on harder problems
- Geographic distribution: Not just SF/Bay Area
Theory of change: Academic legitimacy → more talent + more funding + political influence → field growth
Field Growth Statistics
The AI safety field has grown substantially since 2020, with acceleration around 2023 coinciding with increased public attention following ChatGPT's release.
Field Size Over Time
| Year | Technical AI Safety FTEs | Non-Technical AI Safety FTEs | Total FTEs | Organizations |
|---|---|---|---|---|
| 2015 | ≈50 | ≈20 | ≈70 | ≈15 |
| 2020 | ≈150 | ≈50 | ≈200 | ≈30 |
| 2022 | ≈300 | ≈100 | ≈400 | ≈50 |
| 2024 | ≈500 | ≈400 | ≈900 | ≈65 |
| 2025 | ≈600-645 | ≈500 | ≈1,100 | ≈70 |
Source: AI Safety Field Growth Analysis 2025↗🔗 web★★★☆☆EA ForumAI Safety Field Growth Analysis 2025A useful reference for researchers and funders assessing the scale and growth rate of the AI safety field; numbers should be cross-checked against other workforce surveys as FTE estimates in this field vary significantly by methodology.Stephen McAleese (2025)77 karma · 13 commentsA longitudinal study tracking the growth of the AI safety field from 2010 to 2025, documenting expansion from approximately 400 to 1,100 full-time equivalent researchers across ...ai-safetyfield-buildinggovernancetechnical-safety+3Source ↗
Growth rates:
- Technical AI safety organizations: 24% annual growth
- Technical AI safety FTEs: 21% annual growth
- Non-technical AI safety: approximately 30% annual growth (accelerating since 2023)
Top research areas by FTEs:
- Miscellaneous technical safety (scalable oversight, adversarial robustness, jailbreaks)
- LLM safety
- Interpretability
Methodology note: These estimates may undercount people working on AI safety since many work at organizations that don't explicitly brand themselves as AI safety organizations, particularly in technical safety in academia.
What Needs to Be True
For field-building to be high impact:
- Talent is bottleneck: More people actually means more progress (vs. "too many cooks")
- Sufficient time: Field-building is multi-year investment; need time before critical period
- Quality maintained: Growth doesn't dilute quality or focus
- Absorptive capacity: Ecosystem can integrate new people
- Right people: Recruiting those with high potential for contribution
- Complementarity: New people enable work that wouldn't happen otherwise
Key Bottlenecks and Challenges
The AI safety field faces several structural challenges that limit the effectiveness of field-building efforts:
Pipeline Over-Optimization for Researchers
According to analysis on the EA Forum↗🔗 web★★★☆☆EA ForumAI Safety’s Talent Pipeline is Over-optimised for ResearchersA 2025 EA Forum post by Chris Clay offering a structural critique of AI safety community-building, relevant to anyone thinking about career pathways, talent strategy, or ecosystem coordination within AI safety.Chris Clay🔸 (2025)117 karma · 15 commentsThis EA Forum post argues that AI safety's talent pipeline is structurally biased toward producing researchers, despite leadership consensus that research is not the most neglec...ai-safetyfield-buildingcoordinationpolicy+3Source ↗, the AI safety talent pipeline is over-optimized for researchers:
- The majority of AI safety talent pipelines are optimized for selecting and producing researchers
- Research is not the most neglected talent type in AI safety
- This leads to research-specific talent being over-represented in the community
- Supporting programs strongly select for research skills, missing other crucial roles
Neglected roles: Operations, program management, communications, policy implementation, organizational leadership.
Scaling Gap
There's a massive gap between awareness-level training and the expertise required for selective research fellowships:
- BlueDot plans to train 100,000 people in AI safety fundamentals over 4.5 years
- But few programs bridge from introductory courses to elite research fellowships
- Need scalable programs for the "missing middle"
Organizational Infrastructure Deficit
- Not enough talented founders are building AI safety organizations
- Catalyze's pilot program↗✏️ blog★★★☆☆EA ForumCatalyze's pilot programThis job posting doubles as an overview of Catalyze Impact's incubation model for AI safety organizations, useful for understanding field-building infrastructure efforts within the EA/AI safety ecosystem.Catalyze Impact, Alexandra Bos, Mick (2025)20 karma · 0 commentsCatalyze Impact is a nonprofit incubator addressing organizational bottlenecks in the AI safety ecosystem by connecting founders with co-founders, mentors, and funding. Their pi...ai-safetyfield-buildingcoordinationcommunity+2Source ↗ incubated 11 organizations, with participants reporting the program accelerated progress by an average of 11 months
- Open positions often don't exist because organizations haven't been founded
Compensation Competition
AI Safety Institutes and external research struggle to compete with frontier AI companies:
- Frontier companies offer substantially higher compensation packages
- AISIs must appeal to researchers' desire for public service and impact
- Some approaches: joint university appointments, research sabbaticals, rotating fellowships
Risks and Considerations
Dilution Risk
- Too many people with insufficient expertise
- "Alignment washing" - superficial engagement
- Noise drowns out signal
Mitigation: Selective programs, emphasis on quality, mentorship
Information Hazards
- Publicly discussing AI capabilities could accelerate them
- Spreading awareness of potential attacks
- Attracting bad actors
Mitigation: Careful communication, expert judgment on what to share
Race Dynamics
- Public attention accelerates AI development
- Creates FOMO (fear of missing out)
- Geopolitical competition
Mitigation: Frame carefully, emphasize cooperation, private engagement
Community Problems
- Groupthink and echo chambers
- Lack of ideological diversity
- Social dynamics override epistemic rigor
- Cult-like dynamics
Mitigation: Encourage disagreement, diverse perspectives, epistemic humility
Estimated Impact by Worldview
Long Timelines (10+ years)
Impact: Very High
- Time for field-building to compound
- Training pays off over decades
- Can build robust institutions
- Best time to invest in human capital
Short Timelines (3-5 years)
Impact: Low-Medium
- Insufficient time for new people to become experts
- Better to leverage existing talent
- Exception: rapid deployment of already-skilled people
Optimism About Field Growth
Impact: High
- Every good researcher counts
- Ecosystem effects are strong
- More perspectives improve solutions
Pessimism About Field Growth
Impact: Low
- Talent bottleneck is overstated
- Coordination costs dominate
- Focus on existing excellent people
Who Should Consider This
Strong fit if you:
- Enjoy teaching, mentoring, organizing
- Good at operations and logistics
- Strong communication skills
- Can evaluate talent and potential
- Patient with long timelines
- Value community and culture
Specific roles:
- Program manager: Run training programs (ARENA, MATS, etc.)
- Grantmaker: Evaluate and fund projects
- Educator: Teach courses, create content
- Community organizer: Events, spaces, support
- Communicator: Explain AI safety to various audiences
Backgrounds:
- Education / pedagogy
- Program management
- Operations
- Communications
- Community organizing
- Content creation
Entry paths:
- Staff role at training program
- Local group organizer → full-time
- Teaching assistant → program lead
- Communications role
- Grantmaking entry programs
Less good fit if:
- Prefer direct object-level work
- Impatient with meta-level interventions
- Don't enjoy working with people
- Want immediate measurable impact
Key Organizations
Training Programs
- ARENA (Redwood / independent)
- MATS (independent)
- BlueDot Impact (running AGI Safety Fundamentals)
- AI Safety Camp
Community Organizations
- Centre for Effective Altruism (CEA)
- EAG conferences
- University group support
- Community health
- Lightcone Infrastructure
- LessWrong, Alignment Forum
- Conferences and events
- Office spaces
Funding Organizations
- Coefficient Giving (largest funder)
- Survival and Flourishing Fund
- EA Funds - Long-Term Future Fund
- Founders Pledge
Academic Centers
- CHAI (UC Berkeley)
- Various university groups
Communication
- Individual content creators
- Center for AI Safety (CAIS) (public advocacy)
- Journalists and media
Career Considerations
Pros
- Leveraged impact: Enable many others
- People-focused: Work with smart, motivated people
- Varied work: Teaching, organizing, strategy
- Lower barrier: Don't need research-level technical skills
- Rewarding: See people grow and succeed
Cons
- Hard to measure: Impact is indirect and delayed
- Meta-level: One step removed from object-level problem
- Uncertain: May not produce expected talent
- Community dependent: Success depends on others
- Burnout risk: Emotionally demanding
Compensation
- Program staff: $10-100K
- Directors: $100-150K
- Grantmakers: $80-150K
- Community organizers: $40-80K (often part-time)
Note: Field-building often pays less than technical research but more than pure volunteering
Skills Development
- Program management
- Teaching and mentoring
- Evaluation and judgment
- Operations
- Communication
Complementary Interventions
Field-building enables and amplifies:
- Technical research: Creates researcher pipeline
- Governance: Trains policy experts
- Corporate influence: Provides talent to labs
- All interventions: Increases capacity across the board
Open Questions
Key Questions
- ?Is AI safety talent-constrained or idea-constrained?Talent-constrained
We have more ideas than people to execute them. Good researchers are bottleneck. Field-building is critical.
→ Invest heavily in training and recruitment
Confidence: mediumIdea-constrainedWe don't know what to work on. More people without better ideas doesn't help. Need conceptual breakthroughs first.
→ Focus on research, not growth; be selective about field-building
Confidence: medium - ?Should we prioritize growth or quality in field-building?Growth - quantity has quality of its own
Bigger field attracts more talent, resources, attention. Can't predict who will contribute most. Inclusive approach.
→ Lower barriers, scale programs, broad recruitment
Confidence: lowQuality - excellence is rare and crucialOne excellent researcher worth 100 mediocre ones. Dilution risks real. Selectivity maintains standards.
→ Highly selective programs, mentorship-heavy, focus on top talent
Confidence: medium
Getting Started
If you want to contribute to field-building:
-
Understand the field first:
- Learn AI safety yourself
- Engage with community
- Understand current state
-
Identify your niche:
- Teaching? → Develop curriculum, TA for programs
- Organizing? → Start local group, help with events
- Funding? → Learn grantmaking, advise donors
- Communication? → Write, make videos, explain concepts
-
Start small:
- Volunteer for existing programs
- Organize local reading group
- Create content
- Help with events
-
Build track record:
- Demonstrate impact
- Get feedback
- Iterate and improve
-
Scale up:
- Apply for staff roles
- Launch new programs
- Seek funding for initiatives
Resources:
- CEA community-building resources
- 80,000 Hours on field-building
- Alignment Forum posts on field growth
- MATS/ARENA/BlueDot as examples
Sources & Further Reading
Field Growth and Statistics
- AI Safety Field Growth Analysis 2025↗🔗 web★★★☆☆EA ForumAI Safety Field Growth Analysis 2025A useful reference for researchers and funders assessing the scale and growth rate of the AI safety field; numbers should be cross-checked against other workforce surveys as FTE estimates in this field vary significantly by methodology.Stephen McAleese (2025)77 karma · 13 commentsA longitudinal study tracking the growth of the AI safety field from 2010 to 2025, documenting expansion from approximately 400 to 1,100 full-time equivalent researchers across ...ai-safetyfield-buildinggovernancetechnical-safety+3Source ↗ — Comprehensive dataset of technical and non-technical AI safety organizations and FTEs
- AI Safety Field Growth Analysis 2025 (LessWrong)↗🔗 web★★★☆☆LessWrongAI Safety Field Growth Analysis 2025An updated version of a 2022 baseline estimate by Stephen McAleese; useful for understanding the scale and growth trajectory of the AI safety field, but relies on author estimates rather than comprehensive verified data.Stephen McAleese (2025)30 karma · 14 commentsA quantitative analysis tracking growth of the AI safety field from 2010–2025, estimating total FTEs grew from ~400 in 2022 to ~1,100 in 2025 across 119 organizations. Technical...ai-safetyfield-buildinggovernancepolicy+4Source ↗ — Cross-post with additional discussion
Funding
- An Overview of the AI Safety Funding Situation↗🔗 web★★★☆☆EA ForumAn Overview of the AI Safety Funding SituationUseful reference for understanding the financial infrastructure of AI safety research as of mid-2023, particularly relevant given the post-FTX collapse reshaping of the funding landscape.Stephen McAleese (2023)142 karma · 15 commentsA comprehensive survey of the AI safety funding landscape as of mid-2023, cataloging major philanthropic sources including Open Philanthropy, the FTX Future Fund, and the Long-T...ai-safetyfield-buildingcoordinationgovernance+1Source ↗ — Detailed breakdown of philanthropic funding sources
- Coefficient Giving: Our Progress in 2024 and Plans for 2025↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy: Progress in 2024 and Plans for 2025Open Philanthropy is one of the largest funders in the AI safety space; their annual progress reports are useful for understanding funding priorities, institutional strategy, and which organizations and research directions receive major philanthropic backing.Open Philanthropy reviews its 2024 philanthropic activities and outlines priorities for 2025, with emphasis on AI safety research funding, strategic partnerships, and grants spa...ai-safetyexistential-riskfield-buildinggovernance+2Source ↗ — Self-assessment of AI safety grantmaking
- Coefficient Giving Technical AI Safety RFP↗🔗 web★★★★☆Coefficient GivingOpen Philanthropy Request for Proposals: Technical AI Safety ResearchThis Open Philanthropy RFP is a key funding opportunity document that shaped the direction of technical AI safety research by publicly identifying priority areas; useful context for understanding how philanthropic funding influences the field.Open Philanthropy issued a request for proposals seeking technical AI safety research projects, signaling funding priorities and research directions the organization considers m...ai-safetytechnical-safetyinterpretabilityscalable-oversight+4Source ↗ — 2025 request for proposals ($10M available)
- AI Safety and Security Need More Funders↗🔗 web★★★★☆Coefficient GivingAI Safety and Security Need More FundersPublished by Coefficient Giving, a philanthropic advisory organization; useful for funders and researchers interested in understanding the funding landscape and gaps in AI safety and security.This research piece from Coefficient Giving argues that AI safety and security research is significantly underfunded relative to the risks involved, and makes the case for phila...ai-safetyexistential-riskfield-buildingcoordination+3Source ↗ — Analysis of funding gaps
Training Programs
- MATS Program↗🔗 webMATS Research ProgramMATS is one of the primary talent pipelines into the AI safety field; wiki users interested in career transitions or field-building efforts should consider this a key institutional reference.MATS is an intensive fellowship program designed to help researchers transition into AI safety careers, offering structured mentorship from leading researchers, stipends, and co...ai-safetyalignmentfield-buildingeducational+4Source ↗ — ML Alignment & Theory Scholars official site
- MATS Spring 2024 Extension Retrospective↗✏️ blog★★★☆☆LessWrongMATS Spring 2024 Extension RetrospectiveMATS (ML Alignment Theory Scholars) is a prominent AI safety training program; this retrospective is useful for those interested in field-building, program design, or evaluating the effectiveness of AI safety researcher pipelines.HenningB, Matthew Wearden, Cameron Holmes et al. (2025)27 karma · 1 commentsA retrospective on the MATS (ML Alignment Theory Scholars) Spring 2024 Extension program, reviewing outcomes, lessons learned, and the effectiveness of training researchers for ...ai-safetyfield-buildingtraining-programsalignment+3Source ↗ — Detailed outcomes data
- ARENA 5.0 Impact Report↗✏️ blog★★★☆☆LessWrongARENA 5.0 Impact ReportARENA is a recurring in-person upskilling program for AI safety; this impact report documents outcomes from the fifth cohort and is useful for those evaluating talent pipeline and field-building initiatives in the AI safety ecosystem.JScriven, JamesH, James Fox (2025)25 karma · 0 commentsARENA 5.0 is a 4-week intensive in-person program that upskills technically talented individuals for AI safety work, covering mechanistic interpretability, reinforcement learnin...ai-safetyfield-buildingtraining-programsinterpretability+4Source ↗ — Program outcomes and effectiveness
- ARENA 4.0 Impact Report↗🔗 web★★★☆☆LessWrongARENA 4.0 Impact ReportARENA (Alignment Research Engineer Accelerator) is a prominent technical training program in the AI safety ecosystem; this impact report is useful for evaluators and funders assessing field-building interventions.Chloe Li, JamesH, James Fox (2024)45 karma · 3 commentsThis report documents the outcomes and impact of ARENA 4.0, a technical AI safety training program designed to upskill researchers in alignment-relevant skills such as mechanist...ai-safetyfield-buildingtraining-programstechnical-safety+4Source ↗ — Earlier cohort data
- BlueDot Impact: 2022 AI Alignment Course Impact↗🔗 webBlueDot 2022 Cohort AnalysisUseful empirical reference for evaluating the effectiveness of AI safety education and talent pipeline programs; provides concrete outcome metrics for a structured cohort-based course intervention aimed at growing the AI safety field.BlueDot Impact's analysis of their 2022 AI Safety Fundamentals: Alignment course shows it increased the proportion of participants working on AI safety from 5% to 37% (342 parti...ai-safetyfield-buildingtraining-programsalignment+3Source ↗ — Detailed analysis showing 37% career conversion
Talent Pipeline
- AI Safety's Talent Pipeline is Over-optimised for Researchers↗🔗 web★★★☆☆EA ForumAI Safety’s Talent Pipeline is Over-optimised for ResearchersA 2025 EA Forum post by Chris Clay offering a structural critique of AI safety community-building, relevant to anyone thinking about career pathways, talent strategy, or ecosystem coordination within AI safety.Chris Clay🔸 (2025)117 karma · 15 commentsThis EA Forum post argues that AI safety's talent pipeline is structurally biased toward producing researchers, despite leadership consensus that research is not the most neglec...ai-safetyfield-buildingcoordinationpolicy+3Source ↗ — Key critique of current pipeline structure
- Widening AI Safety's Talent Pipeline↗✏️ blog★★★☆☆EA ForumWidening AI Safety's Talent PipelineRelevant to AI safety community-builders and funders evaluating scalable training interventions; provides empirical data from a novel part-time cohort program aimed at broadening the technical AI safety talent pipeline beyond traditional elite fellowship pathways.RubenCastaing, Nelson_GC, danwil (2025)21 karma · 0 commentsThis report introduces the Technical Alignment Research Accelerator (TARA), a 14-week part-time program designed to fill the gap between introductory AI safety awareness and eli...ai-safetyalignmentfield-buildingeducation+3Source ↗ — Proposals for improvement
- 80,000 Hours: AI Safety Technical Research Career Review↗🔗 web★★★☆☆80,000 HoursAI safety technical research | Career review | 80,000 HoursA practical career-guidance resource from 80,000 Hours aimed at individuals considering entering AI safety research; useful for understanding field entry points and skill requirements rather than technical research content itself.80,000 Hours provides a comprehensive career guide for technical AI safety research, covering empirical and theoretical paths, entry requirements, key organizations, and strateg...ai-safetyalignmenttechnical-safetyexistential-risk+4Source ↗ — Career guidance
- 80,000 Hours: Updates to Our Research About AI Risk and Careers↗🔗 web★★★☆☆80,000 Hours80,000 Hours: Updates to Our Research About AI Risk and CareersA 2024 update from 80,000 Hours, a career-guidance organization focused on high-impact careers, revising their guidance on AI risk and how individuals can best contribute to AI safety efforts through career choices.80,000 Hours updates its research and recommendations regarding AI risk and career paths in AI safety, reflecting evolving views on the urgency and tractability of AI-related ex...ai-safetyexistential-riskgovernancefield-building+3Source ↗ — 2024 strategic updates
Industry Assessment
- FLI AI Safety Index 2024↗🔗 web★★★☆☆Future of Life InstituteFuture of Life Institute: AI Safety Index 2024A high-profile civil society audit of leading AI labs' safety practices, useful for understanding how external organizations assess and compare industry safety commitments; complements internal lab safety cards and government evaluations.The Future of Life Institute's AI Safety Index 2024 systematically evaluates six leading AI companies—including OpenAI, Google DeepMind, Anthropic, Meta, xAI, and Mistral—across...ai-safetyevaluationgovernancepolicy+4Source ↗ — Assessment of AI company safety practices
- AI Safety Index Winter 2025↗🔗 web★★★☆☆Future of Life InstituteAI Safety Index Winter 2025A structured industry-wide safety benchmarking report from FLI; useful for governance discussions and tracking whether leading AI labs are meeting their stated safety commitments over successive index editions.The Future of Life Institute evaluated eight major AI companies across 35 safety indicators, finding widespread deficiencies in risk management and existential safety practices....ai-safetygovernanceevaluationexistential-risk+4Source ↗ — Updated industry assessment
- CAIS 2024 Impact Report↗🔗 web★★★★☆Center for AI SafetyCAIS 2024 Impact ReportThis is CAIS's annual impact report for 2024, useful for understanding the organization's scope, priorities, and contributions to the AI safety ecosystem; relevant for those tracking institutional activity in the field.The Center for AI Safety (CAIS) 2024 Impact Report summarizes the organization's activities, accomplishments, and reach over the year, covering its research, educational program...ai-safetyfield-buildinggovernanceexistential-risk+4Source ↗ — Center for AI Safety annual report
International Coordination
- International AI Safety Report 2025↗🔗 webInternational AI Safety Report 2025This is the first major intergovernmental-style scientific report on AI safety, often compared to the IPCC; highly relevant for understanding the international policy landscape and current scientific consensus on AI risk.A landmark international scientific assessment co-authored by 96 experts from 30 countries, providing a comprehensive overview of general-purpose AI capabilities, risks, and ris...ai-safetygovernancecapabilitiesevaluation+6Source ↗ — Report by 96 AI experts on global safety landscape
- The Global Landscape of AI Safety Institutes↗🔗 webThe Global Landscape of AI Safety InstitutesA useful reference for understanding how national AI Safety Institutes are structured, what they do, and how the international landscape of AI governance institutions is evolving post-Bletchley.This article provides a comprehensive overview of AI Safety Institutes (AISIs) as a novel global governance model, cataloguing existing institutes worldwide and analyzing their ...ai-safetygovernancepolicyevaluation+3Source ↗ — Overview of government AI safety efforts