All Publications
OpenAI
Company BlogHigh(4)
GPT developer, leading AI lab
Credibility Rating
4/5
High(4)High quality. Established institution or organization with editorial oversight and accountability.
104
Resources
117
Citing pages
1
Tracked domains
Tracked Domains
openai.com
Resources (104)
104 resources
| Authors | Summary | ||||
|---|---|---|---|---|---|
| OpenAI Official Homepage | web | — | — | S | 24 |
| OpenAI Preparedness Framework | web | — | — | S | 19 |
| OpenAI: Model Behavior | paper | Rakshith Purushothaman | 2025-01-01 | S | 15 |
| OpenAI Safety Updates | web | — | — | S | 13 |
| GPT-4 System Card | web | — | — | S | 8 |
| Weak-to-strong generalization | web | — | — | S | 7 |
| OpenAI Preparedness Framework | web | — | — | S | 6 |
| Preparedness Framework | web | — | — | S | 6 |
| GPT-4 Technical Report and Research Overview | web | — | — | S | 5 |
| announced December 2024 | web | — | — | S | 5 |
| Safety & responsibility | web | — | — | S | 4 |
| Introducing Superalignment | web | — | — | S | 4 |
| OpenAI Superalignment Fast Grants | web | — | — | S | 4 |
| 2025 OpenAI-Anthropic joint evaluation | web | — | — | S | 4 |
| SWE-bench Verified - OpenAI | web | — | — | S | 4 |
| OpenAI - How We Think About Safety Alignment | web | — | — | S | 3 |
| OpenAI's alignment research | web | — | — | S | 3 |
| Learning to Reason with LLMs: OpenAI o1 | web | — | — | S | 3 |
| OpenAI CoT Monitoring | web | — | — | S | 3 |
| OpenAI Usage Policies | web | — | — | S | 3 |
| OpenAI: Preparedness Framework Version 2 | web | — | — | S | 3 |
| Deliberative alignment: reasoning enables safer language models | web | — | — | S | 3 |
| Extracting Concepts from GPT-4 | web | — | — | S | 3 |
| OpenAI on detection limits | web | — | — | S | 2 |
| Learning to summarize with human feedback | web | — | — | S | 2 |
Rows per page:
Page 1 of 5
Citing Pages (117)
AI Accident Risk CruxesAgentic AIAGI DevelopmentAGI TimelineAI-Assisted AlignmentAI AlignmentAlignment EvaluationsApollo ResearchAlignment Research CenterAuthentication CollapseBioweapons RiskAI Uplift Assessment ModelCenter for AI SafetyCapabilities-to-Safety Pipeline ModelCapability ElicitationAI Capability Threshold ModelAutonomous CodingCompute ConcentrationAI-Driven Concentration of PowerConstitutional AICorporate AI Safety ResponsesCorrigibility FailureAI Risk Critical Uncertainties ModelCyberweapons RiskDangerous Capability EvaluationsDeceptive AlignmentDeep Learning Revolution EraAI Safety Defense in Depth ModelAI DisinformationElicit (AI Research Tool)Emergent CapabilitiesEpistemic SycophancyEU AI ActEval Saturation & The Evals GapAI EvaluationsEvals-Based Deployment GatesAI EvaluationGoal Misgeneralization Probability ModelAI Governance and PolicyHeavy Scaffolding / Agentic SystemsAI-Human Hybrid SystemsInstrumental ConvergenceInstrumental Convergence FrameworkIs Interpretability Sufficient for Safety?AI Safety Intervention Effectiveness MatrixAI Safety Intervention PortfolioJan LeikeAI Knowledge MonopolyAI Lab Safety CultureLarge Language ModelsLarge Language ModelsAI Value Lock-inLong-Horizon Autonomous TasksMesa-OptimizationMesa-Optimization Risk AnalysisMetaculusMETRMinimal ScaffoldingAI Misuse Risk CruxesThird-Party Model AuditingMultipolar Trap Dynamics ModelOpenAIOpenAI FoundationOptimistic Alignment WorldviewPaul ChristianoShould We Pause AI Development?Persuasion and Social ManipulationProcess SupervisionAI ProliferationAI Proliferation Risk ModelAI Development Racing DynamicsRacing Dynamics Impact ModelReasoning and PlanningRed TeamingReducing Hallucinations in AI-Generated Wiki ContentAI Alignment Research AgendasReward HackingReward Hacking Taxonomy and Severity ModelAI Risk Activation Timeline ModelAI Risk Cascade Pathways ModelAI Risk Interaction Network ModelRLHFSafety-Capability Tradeoff ModelAI Safety CasesAI Safety Research Allocation ModelAI Safety Research Value ModelAI Safety Researcher Gap ModelSam AltmanAI Capability SandbaggingSandboxing / ContainmentScalable Eval ApproachesScalable OversightIs Scaling All You Need?AI Scaling LawsSchemingScheming & Deception DetectionScheming Likelihood AssessmentSelf-Improvement and Recursive EnhancementSharp Left TurnSituational AwarenessSleeper Agent DetectionAI Safety Solution CruxesSparse Autoencoders (SAEs)AI Model SteganographySuperintelligenceSycophancyAI Safety Technical Pathway DecompositionTechnical AI Safety ResearchCompute ThresholdsTool Use and Computer UseTreacherous TurnVoluntary AI Safety CommitmentsAI Risk Warning Signs ModelWeak-to-Strong GeneralizationWhy Alignment Might Be HardWorld Models + PlanningWorldview-Intervention Mapping
Publication ID:
openai