Skip to content
Longterm Wiki
Back

Credibility Rating

4/5
High(4)

High quality. Established institution or organization with editorial oversight and accountability.

Rating inherited from publication venue: OpenAI

This July 2023 announcement marked a major institutional commitment by OpenAI to long-term alignment research; the Superalignment team was later restructured following the departures of Sutskever and Leike in 2024.

Metadata

Importance: 78/100blog postprimary source

Summary

OpenAI announces the formation of its Superalignment team, co-led by Ilya Sutskever and Jan Leike, with a goal of solving the core technical challenges of superintelligence alignment within four years. The team will dedicate 20% of OpenAI's secured compute to developing scalable oversight, automated interpretability, robustness testing, and adversarial testing methods to align AI systems smarter than humans.

Key Points

  • OpenAI commits 20% of its secured compute over four years to the Superalignment team focused on superintelligence alignment.
  • Current alignment techniques like RLHF rely on human supervision and will not scale to superintelligent systems.
  • Core research pillars: scalable oversight, generalization, automated interpretability, robustness, and adversarial testing.
  • The team's primary goal is to build a roughly human-level automated alignment researcher that can iteratively help align superintelligence.
  • OpenAI believes superintelligence could arrive this decade, making solving alignment an urgent priority.

Cited by 1 page

PageTypeQuality
Jan LeikePerson27.0

2 FactBase facts citing this source

EntityPropertyValueAs Of
Jan LeikeRole / TitleCo-lead, SuperalignmentJul 2023
Jan LeikeEmployed By1LcLlMGLbwJan 2021

Cached Content Preview

HTTP 200Fetched Mar 20, 20268 KB
Introducing Superalignment \| OpenAI

July 5, 2023

# Introducing Superalignment

###### We need scientific and technical breakthroughs to steer and control AI systems much smarter than us. To solve this problem within four years, we’re starting a new team, co-led by Ilya Sutskever and Jan Leike, and dedicating 20% of the compute we’ve secured to date to this effort. We’re looking for excellent ML researchers and engineers to join us.

* * *

Superintelligence will be the most impactful technology humanity has ever invented, and could help us solve many of the world’s most important problems. But the vast power of superintelligence could also be very dangerous, and could lead to the disempowerment of humanity or even human extinction.

While superintelligence[A](https://openai.com/index/introducing-superalignment/#citation-bottom-A) seems far off now, we believe it could arrive this decade.

Managing these risks will require, among [other things⁠](https://openai.com/index/how-should-ai-systems-behave/), [new institutions for governance⁠](https://openai.com/index/governance-of-superintelligence/) and solving the problem of superintelligence alignment:

_How do we ensure AI systems much smarter than humans follow human intent?_

Currently, we don't have a solution for steering or controlling a potentially superintelligent AI, and preventing it from going rogue. Our current techniques for aligning AI, such as [reinforcement learning from human feedback⁠](https://openai.com/index/instruction-following/), rely on humans’ ability to supervise AI. But humans won’t be able to reliably supervise AI systems much smarter than us,[B](https://openai.com/index/introducing-superalignment/#citation-bottom-B) and so our current alignment techniques will not scale to superintelligence. We need new scientific and technical breakthroughs.

#### Our approach

Our goal is to build a roughly human-level [automated alignment researcher⁠](https://openai.com/blog/our-approach-to-alignment-research/). We can then use vast amounts of compute to scale our efforts, and iteratively align superintelligence.To align the first automated alignment researcher, we will need to 1) develop a scalable training method, 2) validate the resulting model, and 3) stress test our entire alignment pipeline:

1. To provide a training signal on tasks that are difficult for humans to evaluate, we can leverage AI systems to [assist evaluation of other AI systems⁠](https://openai.com/research/critiques/) ( _scalable oversight)._ In addition, we want to understand and control how our models generalize our oversight to tasks we can’t supervise ( _generalization)_.
2. To validate the alignment of our systems, we [automate search for problematic behavior⁠(opens in a new window)](https://www.deepmind.com/blog/red-teaming-language-models-with-language-models) ( _robustness)_ and problematic internals ( [_automated interpretability_ ⁠](https://openai.com/research/language-models-can-explain-neurons-in-lan

... (truncated, 8 KB total)
Resource ID: 41000216ddbfc99d | Stable ID: M2QxZTJjOW