Skip to content
Longterm Wiki

Corrigibility

Scalable Oversightactive

Research on building AI systems that allow themselves to be corrected, modified, or shut down by human operators.

Organizations
3
Key Papers
2
Grants
2
Total Funding
$105K
First Proposed: 2015 (Soares et al., MIRI)
Cluster: Scalable Oversight
Parent Area: Scalable Oversight

Tags

corrigibilityshutdownsafety-research

Grants2

NameRecipientAmountFunderDate
AI Alignment Awards — Shutdown Problem ContestAI Alignment Awards$75KCoefficient Giving2022-09
Building towards a "Limited Agent Foundations" thesis on mild optimization and corrigibilityAlex Turner$30KLong-Term Future Fund (LTFF)2019-04

Funding by Funder

FunderGrantsTotal Amount
Coefficient Giving1$75K
Long-Term Future Fund (LTFF)1$30K