Longterm Wiki
Updated 2026-01-29HistoryData
Page StatusContent
Edited 2 weeks ago1.6k words1 backlinks
37
QualityDraft
28
ImportancePeripheral
10
Structure10/15
1804300%15%
Updated every 3 weeksDue in 6 days
Summary

Conjecture is a 30-40 person London-based AI safety org founded 2021, pursuing Cognitive Emulation (CoEm) - building interpretable AI from ground-up rather than aligning LLMs - with $30M+ Series A funding. Founded by Connor Leahy (EleutherAI), they face high uncertainty about CoEm competitiveness (3-5 year timeline) and commercial pressure risks.

Issues1
QualityRated 37 but structure suggests 67 (underrated by 30 points)

Conjecture

Safety Org

Conjecture

Conjecture is a 30-40 person London-based AI safety org founded 2021, pursuing Cognitive Emulation (CoEm) - building interpretable AI from ground-up rather than aligning LLMs - with $30M+ Series A funding. Founded by Connor Leahy (EleutherAI), they face high uncertainty about CoEm competitiveness (3-5 year timeline) and commercial pressure risks.

TypeSafety Org
Founded2022
LocationLondon, UK
Related
People
Connor Leahy
Safety Agendas
InterpretabilityProsaic Alignment
Organizations
AnthropicRedwood ResearchUK AI Safety Institute
1.6k words ยท 1 backlinks
Safety Org

Conjecture

Conjecture is a 30-40 person London-based AI safety org founded 2021, pursuing Cognitive Emulation (CoEm) - building interpretable AI from ground-up rather than aligning LLMs - with $30M+ Series A funding. Founded by Connor Leahy (EleutherAI), they face high uncertainty about CoEm competitiveness (3-5 year timeline) and commercial pressure risks.

TypeSafety Org
Founded2022
LocationLondon, UK
Related
People
Connor Leahy
Safety Agendas
InterpretabilityProsaic Alignment
Organizations
AnthropicRedwood ResearchUK AI Safety Institute
1.6k words ยท 1 backlinks

Overview

Conjecture is an AI safety research organization founded in 2021 by Connor Leahy and a team of researchers concerned about existential risks from advanced AI. The organization pursues a distinctive technical approach centered on "Cognitive Emulation" (CoEm) - building interpretable AI systems based on human cognition principles rather than aligning existing large language models.

Based in London with a team of 30-40 researchers, Conjecture raised over $10M in Series A funding in 2023. Their research agenda emphasizes mechanistic interpretability and understanding neural network internals, representing a fundamental alternative to mainstream prosaic alignment approaches pursued by organizations like Anthropic and OpenAI.

AspectAssessmentEvidenceSource
Technical InnovationHighNovel CoEm research agendaConjecture Blogโ†—
Funding SecurityStrong$30M+ Series A (2023)TechCrunch Reportsโ†—
Research OutputModerateSelective publication strategyResearch Publicationsโ†—
InfluenceGrowingEuropean AI policy engagementUK AISIโ†—

Risk Assessment

Risk CategorySeverityLikelihoodTimelineTrend
CoEm UncompetitiveHighModerate3-5 yearsUncertain
Commercial Pressure CompromiseMediumHigh2-3 yearsWorsening
Research InsularityLowModerateOngoingStable
Funding SustainabilityMediumLow5+ yearsImproving

Founding and Evolution

Origins (2021)

Conjecture emerged from the EleutherAI collective, an open-source AI research group that successfully recreated GPT-3 as open-source models (GPT-J, GPT-NeoX). Key founding factors:

FactorImpactDetails
EleutherAI ExperienceHighDemonstrated capability replication feasibility
Safety ConcernsHighRecognition of risks from capability proliferation
European GapMediumLimited AI safety ecosystem outside Bay Area
Funding AvailabilityMediumGrowing investor interest in AI safety

Philosophical Evolution: The transition from EleutherAI's "democratize AI" mission to Conjecture's safety-focused approach represents a significant shift in thinking about AI development and publication strategies.

Funding Trajectory

YearFunding StageAmountImpact
2021SeedUndisclosedInitial team of โ‰ˆ15 researchers
2023Series A$30M+Scaled to 30-40 researchers
2024OperatingOngoingSustained research operations

Cognitive Emulation (CoEm) Research Agenda

Core Philosophy

Conjecture's signature approach contrasts sharply with mainstream AI development:

ApproachPhilosophyMethodsEvaluation
Prosaic AlignmentTrain powerful LLMs, align post-hocRLHF, Constitutional AIBehavioral testing
Cognitive EmulationBuild interpretable systems from ground upHuman cognition principlesMechanistic understanding

Key Research Components

Mechanistic Interpretability

  • Circuit discovery in neural networks
  • Feature attribution and visualization
  • Scaling interpretability to larger models
  • Interpretability research collaboration

Architecture Design

  • Modular systems for better control
  • Interpretability-first design choices
  • Trading capabilities for understanding
  • Novel training methodologies

Model Organisms

  • Smaller, interpretable test systems
  • Alignment property verification
  • Deception detection research
  • Goal representation analysis

Key Personnel

Leadership Team

Connor Leahy
CEO and Co-founder
EleutherAI, autodidact ML researcher
Sid Black
Co-founder
EleutherAI technical researcher
Gabriel Alfour
CTO
Former Tezos CTO, systems engineering

Connor Leahy Profile

AspectDetails
BackgroundEleutherAI collective member, GPT-J contributor
EvolutionFrom open-source advocacy to safety-focused research
Public RoleActive AI policy engagement, podcast appearances
ViewsShort AI timelines, high P(doom), interpretability-necessary

Timeline Estimates: Leahy has consistently expressed short AI timeline views, suggesting AGI within years rather than decades.

Research Focus Areas

Mechanistic Interpretability

Research AreaStatusKey Questions
Circuit AnalysisActiveHow do transformers implement reasoning?
Feature ExtractionOngoingWhat representations emerge in training?
Scaling MethodsDevelopmentCan interpretability scale to AGI-level systems?
Goal DetectionEarlyHow can we detect goal-directedness mechanistically?

Comparative Advantages

OrganizationPrimary FocusInterpretability Approach
ConjectureCoEm, ground-up interpretabilityDesign-time interpretability
AnthropicFrontier models + interpretabilityPost-hoc analysis of LLMs
ARCTheoretical alignmentEvaluation and ELK research
RedwoodAI controlInterpretability for control

Strategic Position

Theory of Change

Conjecture's pathway to AI safety impact:

  1. Develop scalable interpretability techniques for powerful AI systems
  2. Demonstrate CoEm viability as competitive alternative to black-box scaling
  3. Influence field direction toward interpretability-first development
  4. Inform governance with technical feasibility insights
  5. Build safe systems using CoEm principles if successful

European AI Safety Hub

RoleImpactExamples
Geographic DiversityHighAlternative to Bay Area concentration
Policy EngagementGrowingUK AISI consultation
Talent DevelopmentModerateEuropean researcher recruitment
Community BuildingEarlyWorkshops and collaborations

Challenges and Criticisms

Technical Feasibility

ChallengeSeverityStatus
CoEm CompetitivenessHighUnresolved - early stage
Interpretability ScalingHighActive research question
Human Cognition ComplexityMediumOngoing investigation
Timeline AlignmentHighCritical if AGI timelines short

Organizational Tensions

Commercial Pressure vs Safety Mission

  • VC funding creates return expectations
  • Potential future deployment pressure
  • Comparison to Anthropic's commercialization path

Publication Strategy Criticism

  • Shift from EleutherAI's radical openness
  • Selective research sharing decisions
  • Balance between transparency and safety

Current Research Outputs

Published Work

TypeFocusImpact
Technical PapersInterpretability methodsResearch community
Blog PostsCoEm explanationsPublic understanding
Policy ContributionsTechnical feasibilityGovernance decisions
Open Source ToolsInterpretability softwareResearch ecosystem

Research Questions

Key Questions

  • ?Can CoEm produce AI systems competitive with scaled LLMs?
  • ?Is mechanistic interpretability sufficient for AGI safety verification?
  • ?How will commercial pressures affect Conjecture's research direction?
  • ?What role should interpretability play in AI governance frameworks?
  • ?Can cognitive emulation bridge neuroscience and AI safety research?
  • ?How does CoEm relate to other alignment approaches like Constitutional AI?

Timeline and Risk Estimates

Leadership Risk Assessments

Conjecture's leadership has articulated clear views on AI timelines and safety approaches, which fundamentally motivate their Cognitive Emulation research agenda and organizational strategy:

Expert/SourceEstimateReasoning
Connor LeahyAGI: 2-10 yearsLeahy has consistently expressed short AI timeline views across multiple public statements and podcasts from 2023-2024, suggesting transformative AI systems could emerge within years rather than decades. These short timelines create urgency for developing interpretability-first approaches before AGI arrives.
Connor LeahyP(doom): High without major changesLeahy has expressed significant concern about the default trajectory of AI development in 2023 statements, arguing that prosaic alignment approaches pursued by frontier labs are insufficient to ensure safety. This pessimism about conventional alignment motivates Conjecture's alternative CoEm approach.
Conjecture ResearchProsaic alignment: InsufficientThe organization's core research direction reflects a fundamental assessment that post-hoc alignment of large language models through techniques like RLHF and Constitutional AI cannot provide adequate safety guarantees. This view, maintained since founding, drives their pursuit of interpretability-first system design.
OrganizationInterpretability: Necessary for safetyConjecture's founding premise holds that mechanistic interpretability is not merely useful but necessary for AI safety verification. This fundamental research assumption distinguishes them from organizations pursuing behavioral safety approaches and shapes their entire technical agenda.

Future Scenarios

Research Trajectory Projections

TimelineOptimisticRealisticPessimistic
2-3 yearsCoEm demonstrations, policy influenceContinued interpretability advancesCommercial pressure compromises
3-5 yearsCompetitive interpretable systemsMixed results, partial successResearch agenda stagnates
5+ yearsField adoption of CoEm principlesPortfolio contribution to safetyMarginalized approach

Critical Dependencies

FactorImportanceUncertainty
Technical FeasibilityCriticalHigh - unproven at scale
Funding ContinuityHighMedium - VC expectations
AGI TimelineCriticalHigh - if very short, insufficient time
Field ReceptivityMediumMedium - depends on results

Relationships and Collaborations

Within AI Safety Ecosystem

OrganizationRelationshipCollaboration Type
AnthropicFriendly competitionInterpretability research sharing
ARCComplementaryDifferent technical approaches
MIRIAligned concernsSkepticism of prosaic alignment
Academic LabsCollaborativeInterpretability technique development

Policy and Governance

UK Engagement

  • UK AI Safety Institute consultation
  • Technical feasibility assessments
  • European AI Act discussions

International Influence

  • Growing presence in global AI safety discussions
  • Alternative perspective to US-dominated discourse
  • Technical grounding for governance approaches

Sources & Resources

Primary Sources

TypeSourceDescription
Official WebsiteConjecture.devโ†—Research updates, team information
Research PapersGoogle Scholarโ†—Technical publications
Blog PostsConjecture Blogโ†—Research explanations, philosophy
InterviewsConnor Leahy Talksโ†—Leadership perspectives

Secondary Analysis

TypeSourceFocus
AI Safety AnalysisLessWrong Postsโ†—Community discussion
Technical ReviewsAlignment Forumโ†—Research evaluation
Policy ReportsGovAI Analysisโ†—Governance implications
Funding NewsTechCrunch Coverageโ†—Business developments

Related Resources

TopicInternal LinksExternal Resources
InterpretabilityTechnical InterpretabilityAnthropic Interpretabilityโ†—
Alignment ApproachesWhy Alignment is HardAI Alignment Forumโ†—
European AI PolicyUK AISIEU AI Officeโ†—
Related OrgsSafety OrganizationsAI Safety Communityโ†—

Related Pages

Top Related Pages

People

Dario Amodei

Labs

Apollo ResearchFAR AI

Safety Research

Prosaic Alignment

Approaches

Scheming & Deception DetectionSparse Autoencoders (SAEs)

Analysis

Model Organisms of MisalignmentCapability-Alignment Race Model

Organizations

Redwood Research

Concepts

AnthropicOpenAIMachine Intelligence Research InstituteAlignment Research CenterUK AI Safety Institute

Risks

Deceptive Alignment

Key Debates

AI Alignment Research AgendasTechnical AI Safety Research

Historical

Deep Learning Revolution EraMainstream Era

Transition Model

Interpretability Coverage