Skip to content

Conjecture

📋Page Status
Page Type:ContentStyle Guide →Standard knowledge base article
Quality:37 (Draft)⚠️
Importance:28 (Peripheral)
Last edited:2026-01-29 (3 days ago)
Words:1.6k
Structure:
📊 18📈 0🔗 35📚 015%Score: 10/15
LLM Summary:Conjecture is a 30-40 person London-based AI safety org founded 2021, pursuing Cognitive Emulation (CoEm) - building interpretable AI from ground-up rather than aligning LLMs - with $30M+ Series A funding. Founded by Connor Leahy (EleutherAI), they face high uncertainty about CoEm competitiveness (3-5 year timeline) and commercial pressure risks.
Issues (1):
  • QualityRated 37 but structure suggests 67 (underrated by 30 points)
Research Lab

Conjecture

Importance28

Conjecture is an AI safety research organization founded in 2021 by Connor Leahy and a team of researchers concerned about existential risks from advanced AI. The organization pursues a distinctive technical approach centered on “Cognitive Emulation” (CoEm) - building interpretable AI systems based on human cognition principles rather than aligning existing large language models.

Based in London with a team of 30-40 researchers, Conjecture raised over $10M in Series A funding in 2023. Their research agenda emphasizes mechanistic interpretability and understanding neural network internals, representing a fundamental alternative to mainstream prosaic alignment approaches pursued by organizations like Anthropic and OpenAI.

AspectAssessmentEvidenceSource
Technical InnovationHighNovel CoEm research agendaConjecture Blog
Funding SecurityStrong$30M+ Series A (2023)TechCrunch Reports
Research OutputModerateSelective publication strategyResearch Publications
InfluenceGrowingEuropean AI policy engagementUK AISI
Risk CategorySeverityLikelihoodTimelineTrend
CoEm UncompetitiveHighModerate3-5 yearsUncertain
Commercial Pressure CompromiseMediumHigh2-3 yearsWorsening
Research InsularityLowModerateOngoingStable
Funding SustainabilityMediumLow5+ yearsImproving

Conjecture emerged from the EleutherAI collective, an open-source AI research group that successfully recreated GPT-3 as open-source models (GPT-J, GPT-NeoX). Key founding factors:

FactorImpactDetails
EleutherAI ExperienceHighDemonstrated capability replication feasibility
Safety ConcernsHighRecognition of risks from capability proliferation
European GapMediumLimited AI safety ecosystem outside Bay Area
Funding AvailabilityMediumGrowing investor interest in AI safety

Philosophical Evolution: The transition from EleutherAI’s “democratize AI” mission to Conjecture’s safety-focused approach represents a significant shift in thinking about AI development and publication strategies.

YearFunding StageAmountImpact
2021SeedUndisclosedInitial team of ≈15 researchers
2023Series A$30M+Scaled to 30-40 researchers
2024OperatingOngoingSustained research operations

Cognitive Emulation (CoEm) Research Agenda

Section titled “Cognitive Emulation (CoEm) Research Agenda”

Conjecture’s signature approach contrasts sharply with mainstream AI development:

ApproachPhilosophyMethodsEvaluation
Prosaic AlignmentTrain powerful LLMs, align post-hocRLHF, Constitutional AIBehavioral testing
Cognitive EmulationBuild interpretable systems from ground upHuman cognition principlesMechanistic understanding

Mechanistic Interpretability

  • Circuit discovery in neural networks
  • Feature attribution and visualization
  • Scaling interpretability to larger models
  • Interpretability research collaboration

Architecture Design

  • Modular systems for better control
  • Interpretability-first design choices
  • Trading capabilities for understanding
  • Novel training methodologies

Model Organisms

  • Smaller, interpretable test systems
  • Alignment property verification
  • Deception detection research
  • Goal representation analysis
Leadership Team
CL
Connor Leahy
CEO and Co-founder
SB
Sid Black
Co-founder
GA
Gabriel Alfour
CTO
AspectDetails
BackgroundEleutherAI collective member, GPT-J contributor
EvolutionFrom open-source advocacy to safety-focused research
Public RoleActive AI policy engagement, podcast appearances
ViewsShort AI timelines, high P(doom), interpretability-necessary

Timeline Estimates: Leahy has consistently expressed short AI timeline views, suggesting AGI within years rather than decades.

Research AreaStatusKey Questions
Circuit AnalysisActiveHow do transformers implement reasoning?
Feature ExtractionOngoingWhat representations emerge in training?
Scaling MethodsDevelopmentCan interpretability scale to AGI-level systems?
Goal DetectionEarlyHow can we detect goal-directedness mechanistically?
OrganizationPrimary FocusInterpretability Approach
ConjectureCoEm, ground-up interpretabilityDesign-time interpretability
AnthropicFrontier models + interpretabilityPost-hoc analysis of LLMs
ARCTheoretical alignmentEvaluation and ELK research
RedwoodAI controlInterpretability for control

Conjecture’s pathway to AI safety impact:

  1. Develop scalable interpretability techniques for powerful AI systems
  2. Demonstrate CoEm viability as competitive alternative to black-box scaling
  3. Influence field direction toward interpretability-first development
  4. Inform governance with technical feasibility insights
  5. Build safe systems using CoEm principles if successful
RoleImpactExamples
Geographic DiversityHighAlternative to Bay Area concentration
Policy EngagementGrowingUK AISI consultation
Talent DevelopmentModerateEuropean researcher recruitment
Community BuildingEarlyWorkshops and collaborations
ChallengeSeverityStatus
CoEm CompetitivenessHighUnresolved - early stage
Interpretability ScalingHighActive research question
Human Cognition ComplexityMediumOngoing investigation
Timeline AlignmentHighCritical if AGI timelines short

Commercial Pressure vs Safety Mission

  • VC funding creates return expectations
  • Potential future deployment pressure
  • Comparison to Anthropic’s commercialization path

Publication Strategy Criticism

  • Shift from EleutherAI’s radical openness
  • Selective research sharing decisions
  • Balance between transparency and safety
TypeFocusImpact
Technical PapersInterpretability methodsResearch community
Blog PostsCoEm explanationsPublic understanding
Policy ContributionsTechnical feasibilityGovernance decisions
Open Source ToolsInterpretability softwareResearch ecosystem
Key Questions (6)
  • Can CoEm produce AI systems competitive with scaled LLMs?
  • Is mechanistic interpretability sufficient for AGI safety verification?
  • How will commercial pressures affect Conjecture's research direction?
  • What role should interpretability play in AI governance frameworks?
  • Can cognitive emulation bridge neuroscience and AI safety research?
  • How does CoEm relate to other alignment approaches like Constitutional AI?

Conjecture’s leadership has articulated clear views on AI timelines and safety approaches, which fundamentally motivate their Cognitive Emulation research agenda and organizational strategy:

Expert/SourceEstimateReasoning
Connor LeahyAGI: 2-10 yearsLeahy has consistently expressed short AI timeline views across multiple public statements and podcasts from 2023-2024, suggesting transformative AI systems could emerge within years rather than decades. These short timelines create urgency for developing interpretability-first approaches before AGI arrives.
Connor LeahyP(doom): High without major changesLeahy has expressed significant concern about the default trajectory of AI development in 2023 statements, arguing that prosaic alignment approaches pursued by frontier labs are insufficient to ensure safety. This pessimism about conventional alignment motivates Conjecture’s alternative CoEm approach.
Conjecture ResearchProsaic alignment: InsufficientThe organization’s core research direction reflects a fundamental assessment that post-hoc alignment of large language models through techniques like RLHF and Constitutional AI cannot provide adequate safety guarantees. This view, maintained since founding, drives their pursuit of interpretability-first system design.
OrganizationInterpretability: Necessary for safetyConjecture’s founding premise holds that mechanistic interpretability is not merely useful but necessary for AI safety verification. This fundamental research assumption distinguishes them from organizations pursuing behavioral safety approaches and shapes their entire technical agenda.
TimelineOptimisticRealisticPessimistic
2-3 yearsCoEm demonstrations, policy influenceContinued interpretability advancesCommercial pressure compromises
3-5 yearsCompetitive interpretable systemsMixed results, partial successResearch agenda stagnates
5+ yearsField adoption of CoEm principlesPortfolio contribution to safetyMarginalized approach
FactorImportanceUncertainty
Technical FeasibilityCriticalHigh - unproven at scale
Funding ContinuityHighMedium - VC expectations
AGI TimelineCriticalHigh - if very short, insufficient time
Field ReceptivityMediumMedium - depends on results
OrganizationRelationshipCollaboration Type
AnthropicFriendly competitionInterpretability research sharing
ARCComplementaryDifferent technical approaches
MIRIAligned concernsSkepticism of prosaic alignment
Academic LabsCollaborativeInterpretability technique development

UK Engagement

  • UK AI Safety Institute consultation
  • Technical feasibility assessments
  • European AI Act discussions

International Influence

  • Growing presence in global AI safety discussions
  • Alternative perspective to US-dominated discourse
  • Technical grounding for governance approaches
TypeSourceDescription
Official WebsiteConjecture.devResearch updates, team information
Research PapersGoogle ScholarTechnical publications
Blog PostsConjecture BlogResearch explanations, philosophy
InterviewsConnor Leahy TalksLeadership perspectives
TypeSourceFocus
AI Safety AnalysisLessWrong PostsCommunity discussion
Technical ReviewsAlignment ForumResearch evaluation
Policy ReportsGovAI AnalysisGovernance implications
Funding NewsTechCrunch CoverageBusiness developments
TopicInternal LinksExternal Resources
InterpretabilityTechnical InterpretabilityAnthropic Interpretability
Alignment ApproachesWhy Alignment is HardAI Alignment Forum
European AI PolicyUK AISIEU AI Office
Related OrgsSafety OrganizationsAI Safety Community