AI Safety Knowledge Base
A structured reference covering risks, technical approaches, governance, organizations, and key people shaping the future of AI safety.
Explore by topic
AI Safety
Governance
Recently updated
View all →Anthropic Valuation Analysis
Valuation analysis updated for Series G (Feb 2026). Anthropic raised $30B at $380B post-money with $14B run-rate revenue, yielding ~27x multiple—no...
Anthropic (Funder)
Comprehensive model of EA-aligned philanthropic capital at Anthropic. At $380B valuation (Series G, Feb 2026, $30B raised): $27-76B risk-adjusted E...
Anthropic IPO
Anthropic is actively preparing for a potential 2026 IPO with concrete steps like hiring Wilson Sonsini and conducting bank discussions, though tim...
Longterm Wiki
A self-referential documentation page describing the Longterm Wiki platform itself—a strategic intelligence tool with ~550 pages, crux mapping of ~...
Rogue AI Scenarios
Analysis of five scenarios for agentic AI takeover-by-accident—sandbox escape, training signal corruption, correlated policy failure, delegation ch...
AI Acceleration Tradeoff Model
Quantitative framework for evaluating how changes to AI development speed affect existential risk and long-term value. Models the marginal impact o...
Goodfire
Goodfire is a well-funded AI interpretability startup valued at $1.25B (Feb 2026) developing mechanistic interpretability tools like Ember API to m...
MAIM (Mutually Assured AI Malfunction)
MAIM (Mutually Assured AI Malfunction) is a deterrence framework introduced in the 2025 paper 'Superintelligence Strategy' by Dan Hendrycks (CAIS),...
641 pages · Continuously updated