Skip to content
Longterm Wiki
Back

Apollo Research - AI Safety Organization

web
apollo-research.ai·apollo-research.ai/

Apollo Research is a key organization in applied AI safety evaluations; their empirical work on scheming and deceptive alignment is frequently cited in discussions of inner alignment and situational awareness risks in frontier models.

Metadata

Importance: 72/100homepage

Summary

Apollo Research is an AI safety organization focused on evaluating advanced AI models for dangerous capabilities, deceptive alignment, and situational awareness. They conduct empirical research into model evaluations, particularly around scheming and deceptive behaviors in frontier AI systems. Their work aims to inform deployment decisions and safety standards for advanced AI.

Key Points

  • Specializes in advanced AI evaluations, particularly testing for deceptive alignment and scheming behaviors in frontier models
  • Conducts empirical research on situational awareness and whether AI models can recognize and exploit their deployment context
  • Works with major AI labs to assess dangerous capabilities before model deployment
  • Focuses on the practical, near-term risks of AI systems that may pursue hidden goals or misrepresent their reasoning
  • Produces public research reports on model evaluations relevant to existential and catastrophic risk scenarios

Cited by 1 page

PageTypeQuality
Deceptive AlignmentRisk75.0
Resource ID: 8b8a890e2ea44a2d | Stable ID: NGY0Y2QyOW