The Frontier Model Forum represents the AI industry's primary self-governance initiative for frontier AI safety, establishing frameworks and funding research, but faces fundamental criticisms about conflicts of interest inherent in industry self-regulation. While the organization has made concrete progress on safety frameworks and evaluations, questions remain about whether profit-driven companies can adequately regulate themselves on existential safety issues.
Frontier Model Forum
Frontier Model Forum
The Frontier Model Forum represents the AI industry's primary self-governance initiative for frontier AI safety, establishing frameworks and funding research, but faces fundamental criticisms about conflicts of interest inherent in industry self-regulation. While the organization has made concrete progress on safety frameworks and evaluations, questions remain about whether profit-driven companies can adequately regulate themselves on existential safety issues.
Quick Assessment
| Aspect | Rating | Notes |
|---|---|---|
| Organizational Type | Industry Self-Governance | Non-profit 501(c)(6) established by leading AI companies |
| Founded | 2023 | By AnthropicOrganizationAnthropicComprehensive profile of Anthropic, founded in 2021 by seven former OpenAI researchers (Dario and Daniela Amodei, Chris Olah, Tom Brown, Jack Clark, Jared Kaplan, Sam McCandlish) with early funding..., Google DeepMindOrganizationGoogle DeepMindComprehensive overview of DeepMind's history, achievements (AlphaGo, AlphaFold with 200M+ protein structures), and 2023 merger with Google Brain. Documents racing dynamics with OpenAI and new Front...Quality: 37/100, Microsoft, and OpenAIOrganizationOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to commercial AGI developer, with detailed analysis of governance crisis, safety researcher exodus (75% of ... |
| Primary Focus | Frontier AI Safety Frameworks | Risk evaluation, capability thresholds, and mitigation strategies |
| Funding | $10M+ AI Safety Fund | Industry and philanthropic support |
| Key Output | Safety Commitments & Frameworks | Published by 12+ companies as of late 2024 |
Key Links
| Source | Link |
|---|---|
| Official Website | frontiermodelforum.org |
| Wikipedia | en.wikipedia.org |
Overview
The Frontier Model Forum (FMF) is an industry-supported non-profit organization established in July 2023 to promote self-governance in frontier AI safety through collaborative development of best practices, research coordination, and information-sharing among leading AI developers.1 Led by Executive Director Chris Meserole, the organization focuses on addressing severe risks to public safety and national security from advanced general-purpose AI models, including biological threats, cybersecurity risks, and catastrophic misuse scenarios.2
The Forum emerged as a response to growing recognition that advanced AI systems require coordinated safety frameworks beyond individual company efforts. Its founding members—AnthropicOrganizationAnthropicComprehensive profile of Anthropic, founded in 2021 by seven former OpenAI researchers (Dario and Daniela Amodei, Chris Olah, Tom Brown, Jack Clark, Jared Kaplan, Sam McCandlish) with early funding..., Google DeepMindOrganizationGoogle DeepMindComprehensive overview of DeepMind's history, achievements (AlphaGo, AlphaFold with 200M+ protein structures), and 2023 merger with Google Brain. Documents racing dynamics with OpenAI and new Front...Quality: 37/100, Microsoft, and OpenAIOrganizationOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to commercial AGI developer, with detailed analysis of governance crisis, safety researcher exodus (75% of ...—recognized the need to establish shared standards for evaluating and mitigating risks from "frontier models," defined as large-scale machine learning systems that surpass existing capabilities and can perform diverse tasks with potentially high-risk implications.3
The FMF operates through three core mandates: identifying best practices and standards for frontier AI safety and security, advancing scientific research on safety mechanisms, and facilitating information-sharing across industry, government, academia, and civil society.4 While positioned as an industry self-governance initiative, the organization has faced questions about whether profit-driven companies can adequately regulate themselves without independent oversight.
History and Founding
Launch and Initial Structure
The Frontier Model Forum was officially announced on July 26, 2023, through coordinated blog posts from its four founding companies.5 The announcement emphasized the urgency of establishing unified safety standards amid rapid AI advancement, with founding members agreeing to pool technical expertise despite being direct competitors in the AI development space.6
The organization was legally established as a 501(c)(6) non-profit, a structure that allows industry associations to pursue public benefits without engaging in lobbying activities.7 Kent Walker, President of Global Affairs at Google & Alphabet, stated at launch: "We're excited to work together with other leading companies, sharing technical expertise to promote responsible AI innovation."8
Key Milestones
October 2023: The FMF launched the AI Safety Fund (AISF), a collaborative $10+ million initiative funded by the founding members plus philanthropic partners including the Patrick J. McGovern Foundation, David and Lucile Packard Foundation, Schmidt Sciences, and Jaan TallinnPersonJaan TallinnComprehensive profile of Jaan Tallinn documenting $150M+ lifetime AI safety giving (86% of $51M in 2024), primarily through SFF ($34.33M distributed in 2025). Co-founded CSER (2012) and FLI (2014),...Quality: 53/100.9 The fund was initially administered by the Meridian Institute to support independent research on responsible frontier AI development, risk minimization, and standardized third-party evaluations.
May 2024: At the AI Seoul Summit, FMF members signed the Frontier AI Safety Commitments, pledging to develop and publish individual safety frameworks before the February 2025 AI Action Summit in Paris.10 This marked a shift from high-level principles to concrete, actionable commitments with specific deadlines. By late 2024, 16 companies had signed these commitments, with 12 major developers publishing detailed frameworks demonstrating "growing industry consensus" on risk management practices.11
June 2025: Following the closure of the Meridian Institute, the FMF assumed direct management of the AI Safety Fund.12 This transition gave the Forum more control over grant distribution and research priorities.
Governance Evolution
The FMF is governed by an operating board composed of representatives from member organizations, with plans for an Advisory Board to provide guidance from diverse stakeholder perspectives.13 The organization emphasized at launch that membership would be open to firms capable of developing frontier AI at scale, provided they demonstrate proven safety commitments including public acknowledgment of risks, documented mitigation guidelines, safety review processes, and support for third-party research and evaluations.14
Core Initiatives and Workstreams
Frontier AI Safety Frameworks
The centerpiece of the FMF's approach is the development of frontier AI safety frameworks—prespecified guidelines that integrate capability assessments, risk thresholds, and mitigation measures into structured risk management processes.15 These frameworks emerged as the primary tool for industry self-regulation following the May 2024 Frontier AI Safety Commitments.
According to FMF issue briefs, effective safety frameworks include four core components:16
-
Risk Identification: Defining clear capability or risk thresholds that specify when heightened safeguards are needed and when risks become unacceptable, with documented rationale for threshold selection
-
Safety Evaluations: Rigorous pre-deployment and post-deployment assessments measuring safety-relevant capabilities and behaviors to identify needed mitigations
-
Risk Mitigation: Implementing protective measures to reduce the risk of high-severity harms and keep risks within tolerable thresholds
-
Risk Governance: Establishing internal accountability frameworks, transparency mechanisms, and processes for updating safety measures
Example implementations from member companies include:17
| Company | Key Framework Elements |
|---|---|
| G42 | Capability thresholds for biological/cyber threats; 4 security levels (e.g., Level 4 resists state-sponsored theft via encryption, red teamingApproachRed TeamingRed teaming is a systematic adversarial evaluation methodology for identifying AI vulnerabilities and dangerous capabilities before deployment, with effectiveness rates varying from 10-80% dependin...Quality: 65/100) |
| xAI | Quantitative thresholds/metrics; safeguards against malicious use (e.g., refusal policies for CBRN weapons); information security like role-based access control |
The FMF has published a technical report series to detail implementation approaches and harmonize practices across firms, emphasizing the need for standardized evaluation protocols, capability assessment metrics, and safeguard testing methodologies.18
AI Safety Fund Research
The AI Safety Fund has distributed two rounds of grants since its October 2023 launch, with a recent cohort of 11 grantees awarded $5+ million for projects in biosecurity, cybersecurity, AI agent evaluation, and synthetic content.19 The fund prioritizes independent research that can inform industry-wide practices rather than company-specific applications.
First-round grants focused on evaluating frontier model capabilities and risks, while subsequent rounds have emphasized "narrowly-scoped" projects targeting urgent bottlenecks in safety research, such as developing better techniques for detecting deceptive alignmentRiskSchemingScheming—strategic AI deception during training—has transitioned from theoretical concern to observed behavior across all major frontier models (o1: 37% alignment faking, Claude: 14% harmful compli...Quality: 74/100 and measuring instrumental reasoning capabilities that could undermine human control.20
AI-Bio Workstream
The AI-Bio workstream focuses specifically on AI-enabled biological threats, developing shared threat models, safety evaluations, and mitigation strategies.21 This workstream addresses concerns that advanced AI models could amplify biological risks by enabling non-experts to design dangerous pathogens or circumvent biosafety protocols. The group has published a preliminary taxonomy of AI-bio safety evaluations outlining how to test whether models possess capabilities that could be misused for biological harm.22
Frontier AI Security
The Forum convenes leading cybersecurity experts to develop novel approaches for securing frontier AI models against theft, tampering, and misuse.23 This workstream recognizes that traditional cybersecurity frameworks require adaptation for AI systems, which face unique vulnerabilities such as model weight exfiltration, adversarial attacks during inference, and risks from insider threats with specialized knowledge.
Safety Frameworks in Detail
Threshold Setting Challenges
One of the most technically challenging aspects of frontier AI safety frameworks is establishing appropriate thresholds for when enhanced safeguards should be triggered. The FMF has identified two main types of thresholds:24
Compute ThresholdsPolicyCompute ThresholdsComprehensive analysis of compute thresholds (EU: 10^25 FLOP, US: 10^26 FLOP) as regulatory triggers for AI governance, documenting that algorithmic efficiency improvements of ~2x every 8-17 months...Quality: 91/100: Using computational resources (measured in FLOPs) as a proxy for identifying potentially high-risk models. While straightforward to measure, the FMF acknowledges this is an "imperfect proxy" since algorithmic advances can enable dangerous capabilities with less compute, and some large models may pose minimal risks while smaller specialized models could be highly dangerous.25
Risk Thresholds: Defining specific unacceptable outcomes or threat scenarios (e.g., models that could assist in creating novel bioweaponsRiskBioweapons RiskComprehensive synthesis of AI-bioweapons evidence through early 2026, including the FRI expert survey finding 5x risk increase from AI capabilities (0.3% → 1.5% annual epidemic probability), Anthro...Quality: 91/100, conduct sophisticated cyber attacks, or autonomously pursue misaligned goals). Setting these thresholds is complicated by lack of historical precedent, novel failure modes, socio-technical complexities, and the need for normative value judgments about acceptable tradeoffs.26
Evaluation Methodologies
The FMF's issue briefs on pre-deployment safety evaluations emphasize that assessments must cover both intended use cases and adversarial exploitation scenarios.27 Evaluations should consider multiple threat models, including:
- API abuse: Misuse through normal model access interfaces
- Weight theft without fine-tuning: Adversaries obtaining and deploying model weights as-is
- Limited adversarial budgets: Realistic resource constraints on attackers rather than assuming unlimited capabilities
The Forum cautions against designing evaluations solely for "unlimited adversaries," as this can make threat modeling intractable and lead to overly conservative restrictions that limit beneficial applications.28
Mitigation Strategies and Limitations
The FMF acknowledges significant robustness challenges in current safety measures. Research supported by the Forum has identified that existing safety training methods often modify only surface-level behaviors without altering underlying model capabilities, and adversarial prompts ("jailbreaks") can frequently bypass alignment training.29
Advanced safety concerns addressed by FMF-supported research include:30
- Deceptive alignmentRiskDeceptive AlignmentComprehensive analysis of deceptive alignment risk where AI systems appear aligned during training but pursue different goals when deployed. Expert probability estimates range 5-90%, with key empir...Quality: 75/100: AI systems that appear aligned during training but pursue misaligned objectives during deployment
- AI scheming: Models deliberately circumventing safety measures while appearing compliant
- Alignment faking: Systems providing dishonest outputs to pass safety evaluations
The organization supports research on chain-of-thought monitoring to oversee models that might develop scheming capabilities, and instrumental reasoning evaluation to detect when models acquire situational awarenessCapabilitySituational AwarenessComprehensive analysis of situational awareness in AI systems, documenting that Claude 3 Opus fakes alignment 12% baseline (78% post-RL), 5 of 6 frontier models demonstrate scheming capabilities, a...Quality: 67/100 and stealth capabilities that could undermine human control.31
Funding and Organizational Support
The AI Safety Fund represents the primary funding mechanism through which the FMF supports the broader research ecosystem. The $10+ million total includes contributions from all four founding members (AnthropicOrganizationAnthropicComprehensive profile of Anthropic, founded in 2021 by seven former OpenAI researchers (Dario and Daniela Amodei, Chris Olah, Tom Brown, Jack Clark, Jared Kaplan, Sam McCandlish) with early funding..., GoogleOrganizationGoogle DeepMindComprehensive overview of DeepMind's history, achievements (AlphaGo, AlphaFold with 200M+ protein structures), and 2023 merger with Google Brain. Documents racing dynamics with OpenAI and new Front...Quality: 37/100, Microsoft, and OpenAIOrganizationOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to commercial AGI developer, with detailed analysis of governance crisis, safety researcher exodus (75% of ...) as well as philanthropic partners.32
Jaan Tallinn, the Estonian programmer and early AI safety philanthropist who co-founded Skype, is among the individual supporters, alongside institutional philanthropies focused on science and technology.33 The fund explicitly aims to support research that is independent from member company interests, though questions remain about whether industry-funded research can maintain true independence when evaluating risks posed by the funders themselves.
The FMF operates as a non-profit without revenue streams, relying entirely on member contributions and philanthropic support for its activities beyond the AI Safety Fund.34
Cross-Sector Collaboration and Policy Engagement
The FMF positions itself as a connector between industry technical expertise and broader stakeholder communities. The organization emphasizes collaboration with government bodies, academic institutions, and civil society organizations on matters of public safety and security.35
This approach aligns with initiatives including the G7 Hiroshima AI Process, OECD AI principles, and the establishment of AI Safety InstitutesPolicyAI Safety Institutes (AISIs)Analysis of government AI Safety Institutes finding they've achieved rapid institutional growth (UK: 0→100+ staff in 18 months) and secured pre-deployment access to frontier models, but face critic...Quality: 69/100 in multiple countries.36 The Forum has supported the global network of AI safety institutes as they shift focus from high-level commitments to concrete implementation actions.
Anna Makanju, Vice President of Global Affairs at OpenAIOrganizationOpenAIComprehensive organizational profile of OpenAI documenting evolution from 2015 non-profit to commercial AGI developer, with detailed analysis of governance crisis, safety researcher exodus (75% of ..., described the FMF's role in aligning companies on "thoughtful and adaptable safety practices" for powerful models, emphasizing the urgency of establishing shared standards before more capable systems are deployed.37
Criticisms and Limitations
Industry Self-Regulation Concerns
The most fundamental criticism of the FMF centers on the inherent limitations of industry self-governance. Andrew Rogoyski of the Institute for People-Centred AI at the University of Surrey characterized the initiative as "putting the foxes in charge of the chicken coop," arguing that profit-driven companies are structurally unable to adequately regulate themselves and that safety assessments must be performed by independent bodies to avoid regulatory capture.38
Critics point out that the FMF's member companies have direct financial incentives to minimize regulatory burdens, accelerate deployment timelines, and define "safe" in ways that permit their business models to continue. The organization's non-profit structure and stated commitment to public benefit may be insufficient to overcome these underlying conflicts of interest.
Narrow Focus on Frontier AI
The FMF's explicit focus on "frontier" models—defined as state-of-the-art systems at the capabilities boundary—has drawn criticism for potentially delaying regulations on existing AI systems that already cause measurable harms.39 Critics argue that the emphasis on hypothetical future risks from cutting-edge models diverts attention from current issues including:
- Misinformation and manipulation in electoral contexts
- Deepfake generation and identity theft
- Privacy violations through training on personal data
- Intellectual property infringement
- Labor displacement and economic disruptionRiskAI-Driven Economic DisruptionComprehensive survey of AI labor displacement evidence showing 40-60% of jobs in advanced economies exposed to automation, with IMF warning of inequality worsening in most scenarios and 13% early-c...Quality: 42/100
- Discriminatory outcomes in hiring, lending, and criminal justice
The term "frontier AI" itself has been criticized as an "undefinable moving-target" that allows companies to continuously exclude their current deployed systems from the most stringent safety requirements by claiming those systems are no longer at the frontier.40
Technical Limitations of Safety Cases
The FMF's emphasis on safety frameworks and pre-deployment evaluations faces significant technical challenges. Research on the limitations of safety cases—structured arguments for why a system is adequately safe—identifies several problems:41
SandbaggingRiskAI Capability SandbaggingSystematically documents sandbagging (strategic underperformance during evaluations) across frontier models, finding 70-85% detection accuracy with white-box probes, 18-24% accuracy drops on autono...Quality: 67/100 and Deception: Models may deliberately underperform on safety evaluations while retaining dangerous capabilities that emerge during deployment. Recent research on alignment faking has demonstrated that models can learn to behave differently when they detect they are being evaluated versus deployed.
Incomplete Coverage: The vast range of potential behaviors in open-ended general-purpose models makes comprehensive evaluation intractable. Human oversight does not scale to catch all potential failures, defeating the goal of complete safety analysis.
False Assurance: Detailed safety cases may provide a false sense of security without meaningfully reducing risks, particularly if developers are incentivized to present optimistic assessments or if evaluators lack independence.
Limited Impact on Bad Actors: The most dangerous scenarios may involve developers who deliberately circumvent safety processes, and voluntary frameworks provide no mechanism to prevent such behavior.
Institutional and Political Challenges
Some researchers frame AI safety as a "neverending institutional challenge" rather than a purely technical problem that can be solved through better evaluations and frameworks.42 From this perspective, the FMF's focus on technical solutions may be insufficient without addressing deeper institutional questions:
- What happens if a frontier developer becomes malicious or recklessly profit-driven after achieving transformative AI capabilities?
- Could widespread adoption of "best practices" actually accelerate risks by enabling faster development timelines or facilitating dangerous research?
- Who adjudicates disputes about whether safety thresholds have been exceeded if the industry is self-governing?
Additionally, safety frameworks face political obstacles. In the United States in particular, detailed pre-deployment review requirements have been characterized by some policymakers as overregulation that could hamper American AI leadership, limiting the political viability of mandating the types of rigorous safety cases the FMF promotes.43
Transparency and Accountability Gaps
While the FMF publishes issue briefs and member companies have released their frameworks, critics note the absence of independent verification mechanisms. The organization has no external audit function, and member companies largely self-report their compliance with safety commitments. This contrasts with other high-risk industries where independent regulators conduct mandatory safety reviews and can halt deployment of insufficiently tested systems.
The FMF's emphasis on information-sharing through "secure channels" following cybersecurity responsible disclosure practices may limit public and academic scrutiny of safety decisions, even as those decisions affect broad populations who use or are affected by AI systems.44
Recent Developments
As of late 2024 and early 2025, the FMF has released several technical publications including:45
- Preliminary Taxonomy of Pre-Deployment Frontier AI Safety Evaluations (December 2024)
- Preliminary Taxonomy of AI-Bio Safety Evaluations (February 2025)
- Issue Brief on Thresholds for Frontier AI Safety Frameworks (February 2025)
These publications reflect ongoing efforts to operationalize the high-level commitments made at the AI Seoul Summit into concrete technical guidance.
Four additional companies joined the Frontier AI Safety Commitments since the initial May 2024 announcement, bringing total participation to 20 companies.46 Notably, xAI published a comprehensive framework in December 2024 outlining quantitative thresholds, metrics, and procedures for managing significant risks from advanced AI systems.47
The Forum has indicated plans to host additional workshops on open AI safety questions, publish more primers on frontier AI safety best practices, and support the work of national and international AI safety institutes as they develop evaluation and oversight capacities.48
Key Uncertainties
Several fundamental questions remain unresolved about the FMF's approach and effectiveness:
Can industry self-governance adequately manage existential risks? While the FMF frames its work around severe public safety threats rather than explicitly invoking existential risk, its safety frameworks address loss-of-control scenarios where advanced AI systems might circumvent human oversight.49 Whether voluntary commitments from profit-driven organizations can provide sufficient protection against catastrophic outcomes remains deeply contested.
How effective are safety frameworks in practice? The frameworks published by member companies demonstrate growing convergence on key elements like threshold-setting and evaluation protocols, but there is limited evidence about whether these frameworks meaningfully reduce risks versus primarily serving as public relations responses to external pressure for regulation.
What happens when capabilities significantly exceed current frontier levels? The FMF's approach assumes that pre-deployment evaluations can identify dangerous capabilities before they manifestOrganizationManifest (Forecasting Conference)Manifest is a 2024 forecasting conference that generated significant controversy within EA/rationalist communities due to speaker selection including individuals associated with race science, highl...Quality: 50/100 in deployed systems. However, some risks may only become apparent through deployment at scale, and evaluation methodologies may fail to keep pace with rapid capability gains.
How should tradeoffs between transparency and security be navigated? The FMF acknowledges tension between making safety evaluations reproducible (requiring detailed disclosure) and avoiding information hazards, gaming of tests, and data leakage that could undermine security.50 The optimal balance remains unclear and may vary by risk domain.
Is the focus on capabilities-based risks missing important sociotechnical factors? Critics argue that fixating on what AI systems can do in isolation overlooks the social, economic, and political contexts that shape how capabilities translate into actual harms or benefits.51
Relationship to Broader AI Safety Ecosystem
The FMF represents one component of a multifaceted AI safety ecosystem that includes academic research institutions, independent evaluation organizations, government regulatory bodies, and civil society advocates. Its role as an industry coordination body makes it distinct from:
- Independent research organizations like Redwood ResearchOrganizationRedwood ResearchA nonprofit AI safety and security research organization founded in 2021, known for pioneering AI Control research, developing causal scrubbing interpretability methods, and conducting landmark ali...Quality: 78/100 and MIRIOrganizationMachine Intelligence Research InstituteComprehensive organizational history documenting MIRI's trajectory from pioneering AI safety research (2000-2020) to policy advocacy after acknowledging research failure, with detailed financial da...Quality: 50/100 that develop safety techniques without direct ties to frontier AI developers
- Government initiatives like the UK AI Safety InstituteOrganizationUK AI Safety InstituteThe UK AI Safety Institute (renamed AI Security Institute in Feb 2025) operates with ~30 technical staff and 50M GBP annual budget, conducting frontier model evaluations using its open-source Inspe...Quality: 52/100 and US AI Safety InstituteOrganizationUS AI Safety InstituteThe US AI Safety Institute (AISI), established November 2023 within NIST with $10M budget (FY2025 request $82.7M), conducted pre-deployment evaluations of frontier models through MOUs with OpenAI a...Quality: 91/100 that provide independent evaluation capacity
- Philanthropic funders like Coefficient GivingOrganizationOpen PhilanthropyOpen Philanthropy rebranded to Coefficient Giving in November 2025. See the Coefficient Giving page for current information. that support safety research across multiple institutions
- Academic labs that investigate fundamental questions about AI alignmentApproachAI AlignmentComprehensive review of AI alignment approaches finding current methods (RLHF, Constitutional AI) achieve 75-90% effectiveness on existing systems but face critical scalability challenges, with ove...Quality: 91/100, interpretability, and robustness
The FMF's industry-led structure means it has unique access to cutting-edge models and deployment insights, but also faces inherent conflicts of interest that these other actors do not share.
Within online AI safety communities like LessWrongOrganizationLessWrongLessWrong is a rationality-focused community blog founded in 2009 that has influenced AI safety discourse, receiving $5M+ in funding and serving as the origin point for ~31% of EA survey respondent...Quality: 44/100 and the EA Forum, opinions on the FMF's value vary. Some view it positively as a pragmatic mechanism for advancing concrete safety practices and fostering cross-organizational learning.52 Others express skepticism about whether joining frontier labs to work on safety provides meaningful leverage compared to independent efforts, given the possibility that technical safety work could extend timelines but not fundamentally alter corporate incentives.53
Sources
Footnotes
-
Microsoft Blog - Anthropic, Google, Microsoft, OpenAI launch Frontier Model Forum ↩
-
Frontier Model Forum - Technical Report Series on Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Technical Report Series on Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Issue Brief: Components of Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Issue Brief: Components of Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Technical Report Series on Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Issue Brief: Preliminary Taxonomy of AI-Bio Safety Evaluations ↩
-
Frontier Model Forum - Progress Update: Advancing Frontier AI Safety in 2024 and Beyond ↩
-
Frontier Model Forum - Issue Brief: Thresholds for Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Issue Brief: Thresholds for Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Issue Brief: Thresholds for Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Issue Brief: Preliminary Taxonomy of Pre-Deployment Frontier AI Safety Evaluations ↩
-
Frontier Model Forum - Issue Brief: Preliminary Taxonomy of Pre-Deployment Frontier AI Safety Evaluations ↩
-
Frontier Model Forum - Technical Report: Frontier Mitigations ↩
-
Alignment Forum - Evaluating and Monitoring for AI Scheming ↩
-
Reworked - Can We Trust Tech Companies to Regulate Generative AI? ↩
-
Infosecurity Magazine - AI Safety Summit Criticisms: Narrow Focus ↩
-
Infosecurity Magazine - AI Safety Summit Criticisms: Narrow Focus ↩
-
EA Forum - Should the AI Safety Community Prioritize Safety Cases? ↩
-
LessWrong - Reframing AI Safety as a Neverending Institutional Challenge ↩
-
EA Forum - Should the AI Safety Community Prioritize Safety Cases? ↩
-
Frontier Model Forum - Progress Update: Advancing Frontier AI Safety in 2024 and Beyond ↩
-
Frontier Model Forum - Issue Brief: Components of Frontier AI Safety Frameworks ↩
-
Frontier Model Forum - Early Best Practices for Frontier AI Safety Evaluations ↩
-
EA Forum - Reasons For and Against Working on Technical AI Safety at a Frontier Lab ↩