The National Institute of Standards and Technology's role in developing AI standards, risk management frameworks, and safety guidelines for the United States.
AI EvaluationApproachAI EvaluationComprehensive overview of AI evaluation methods spanning dangerous capability assessment, safety properties, and deception detection, with categorized frameworks from industry (Anthropic Constituti...Quality: 72/100Cooperative AIApproachCooperative AICooperative AI research addresses multi-agent coordination failures through game theory and mechanism design, with ~$1-20M/year investment primarily at DeepMind and academic groups. The field remai...Quality: 55/100
Policy
AI Safety Institutes (AISIs)PolicyAI Safety Institutes (AISIs)Analysis of government AI Safety Institutes finding they've achieved rapid institutional growth (UK: 0→100+ staff in 18 months) and secured pre-deployment access to frontier models, but face critic...Quality: 69/100
Concepts
Agentic AICapabilityAgentic AIAnalysis of agentic AI capabilities and deployment challenges, documenting industry forecasts (40% of enterprise apps by 2026, $199B market by 2034) alongside implementation difficulties (40%+ proj...Quality: 68/100Government Orgs OverviewGovernment Orgs OverviewOverview of national AI Safety Institutes (UK, US, and 11+ countries as of 2026) and intergovernmental bodies, covering budgets, mandates, and key dynamics like political vulnerability and lab rela...Quality: 41/100AI Standards DevelopmentConceptAI Standards DevelopmentComprehensive analysis of AI standards bodies (ISO/IEC, IEEE, NIST, CEN-CENELEC) showing how voluntary technical standards become de facto requirements through regulatory integration, particularly ...Quality: 69/100
Risks
Emergent CapabilitiesRiskEmergent CapabilitiesEmergent capabilities—abilities appearing suddenly at scale without explicit training—pose high unpredictability risks. Wei et al. documented 137 emergent abilities; recent models show step-functio...Quality: 61/100
Organizations
Alignment Research CenterOrganizationAlignment Research CenterComprehensive reference page on ARC (Alignment Research Center), covering its evolution from a dual theory/evals organization to ARC Theory (3 permanent researchers) plus the METR spin-out (Decembe...Quality: 57/100Frontier Model ForumOrganizationFrontier Model ForumThe Frontier Model Forum represents the AI industry's primary self-governance initiative for frontier AI safety, establishing frameworks and funding research, but faces fundamental criticisms about...Quality: 58/100US AI Safety InstituteOrganizationUS AI Safety InstituteThe US AI Safety Institute (AISI), established November 2023 within NIST with $10M budget (FY2025 request $82.7M), conducted pre-deployment evaluations of frontier models through MOUs with OpenAI a...Quality: 91/100Global Partnership on Artificial Intelligence (GPAI)OrganizationGlobal Partnership on Artificial Intelligence (GPAI)GPAI represents the first major multilateral AI governance initiative but operates as a non-binding policy laboratory with limited enforcement power and structural coordination challenges. While pr...Quality: 50/100METROrganizationMETRMETR conducts pre-deployment dangerous capability evaluations for frontier AI labs (OpenAI, Anthropic, Google DeepMind), testing autonomous replication, cybersecurity, CBRN, and manipulation capabi...Quality: 66/100Centre for Long-Term ResilienceOrganizationCentre for Long-Term ResilienceThe Centre for Long-Term Resilience is a UK-based think tank that has demonstrated concrete policy influence on AI and biosecurity risks, including contributing to the UK's AI Strategy and Biologic...Quality: 63/100
Other
Richard MallahPersonRichard MallahExecutive Director of CARMA (Center for AI Risk Management & Alignment), incubated by the Future of Life Foundation. Former Principal AI Safety Strategist at FLI since 2014. Focus areas include AI ...Percy LiangPersonPercy LiangStanford professor of Computer Science. Director of the Center for Research on Foundation Models (CRFM). Creator of the HELM evaluation framework.