QualityGoodQuality: 63/100Human-assigned rating of overall page quality, considering depth, accuracy, and completeness.Structure suggests 100
12
ImportancePeripheralImportance: 12/100How central this topic is to AI safety. Higher scores mean greater relevance to understanding or mitigating AI risk.
15
Structure15/15Structure: 15/15Automated score based on measurable content features.Word count2/2Tables3/3Diagrams2/2Internal links2/2Citations3/3Prose ratio2/2Overview section1/1
24TablesData tables in the page7DiagramsCharts and visual diagrams24Internal LinksLinks to other wiki pages0FootnotesFootnote citations [^N] with sources13External LinksMarkdown links to outside URLs%4%Bullet RatioPercentage of content in bullet lists
A self-referential documentation page describing the Longterm Wiki platform itself—a strategic intelligence tool with ~550 pages, crux mapping of ~50 uncertainties, and quality scoring across 6 dimensions. Features include entity cross-linking, interactive causal diagrams, and structured YAML databases tracking expert positions on key AI safety cruxes.
Issues2
QualityRated 63 but structure suggests 100 (underrated by 37 points)
A self-referential documentation page describing the Longterm Wiki platform itself—a strategic intelligence tool with ~550 pages, crux mapping of ~50 uncertainties, and quality scoring across 6 dimensions. Features include entity cross-linking, interactive causal diagrams, and structured YAML databases tracking expert positions on key AI safety cruxes.
QURIOrganizationQURI (Quantified Uncertainty Research Institute)QURI develops Squiggle (probabilistic programming language with native distribution types), SquiggleAI (Claude-powered model generation producing 100-500 line models), Metaforecast (aggregating 2,1...Quality: 48/100 (Quantified Uncertainty Research Institute)
The Longterm Wiki is a strategic intelligence platform for AI safety prioritization. Unlike general encyclopedias or community wikis, it serves as a decision-support tool for funders, researchers, and policymakers asking: "Where should the next marginal dollar or researcher-hour go?"
Loading diagram...
The project addresses four problems in the AI safety field:
Problem
How the Wiki Addresses It
Fragmented knowledge
Consolidated, cross-linked knowledge base with ≈550 pages
Unclear cruxes
Explicit mapping of key uncertainties and expert disagreements
Poor prioritization legibility
Worldview → intervention mapping showing how assumptions lead to priorities
The wiki is deliberately opinionated about importance and uncertainty—it rates content quality, tracks expert positions on cruxes, and makes prioritization implications explicit. This distinguishes it from neutral reference works like Wikipedia or discussion platforms like LessWrong.
Content is editorially curated rather than community-contributed, ensuring consistency and quality control. Each page goes through a grading pipeline that scores novelty, rigor, actionability, and completeness.
Content Architecture
The wiki has four interconnected layers of content:
Priority rankings, robust interventions, high-VOI research
Derived from above
Major Sections
Section
Content
Page Count
Example Pages
Knowledge Base
Risks, interventions, organizations, people
≈350
Deceptive AlignmentRiskDeceptive AlignmentComprehensive analysis of deceptive alignment risk where AI systems appear aligned during training but pursue different goals when deployed. Expert probability estimates range 5-90%, with key empir...Quality: 75/100, AI Safety InstitutesPolicyAI Safety Institutes (AISIs)Analysis of government AI Safety Institutes finding they've achieved rapid institutional growth (UK: 0→100+ staff in 18 months) and secured pre-deployment access to frontier models, but face critic...Quality: 69/100
AI Transition Model
Comprehensive factor network with outcomes and scenarios
npm run crux -- --help # Show all domains
npm run crux -- validate # Run all validators
npm run crux -- analyze # Analysis and reporting
npm run crux -- fix # Auto-fix common issues
npm run crux -- content # Page management
npm run crux -- generate # Content generation
Compare interpretabilitySafety AgendaInterpretabilityMechanistic interpretability has extracted 34M+ interpretable features from Claude 3 Sonnet with 90% automated labeling accuracy and demonstrated 75-85% success in causal validation, though less th...Quality: 66/100 vs governanceParameterGovernance (Civ. Competence)This is a placeholder page with no actual content - only component imports that would render data from elsewhere in the system. Cannot assess importance or quality without the underlying content. approaches
Crux identification
Crux mapping shows which uncertainties matter most
Which assumptions drive different funding priorities?
Expert landscape
Expert profiles with positions
Who believes what about deceptive alignment?
Gap analysis
Quality scores reveal under-developed areas
Which important topics lack quality coverage?
For Researchers
Use Case
Wiki Feature
Example
Literature synthesis
Consolidated coverage with citations
Find all sources on a specific risk
Gap identification
Coverage analysis, importance vs quality
What important topics need more research?
Position mapping
Disagreement visualization
Where do Yudkowsky and Christiano diverge?
Model building
Causal diagrams as starting points
Use wiki models as research scaffolding
For Policymakers
Use Case
Wiki Feature
Example
Risk taxonomy
Structured hierarchy with assessments
Navigate from high-level categories to specific risks
Intentional limitation, links to broader resources
Early stage
Incomplete coverage
Active development, prioritized expansion
No real-time data
Static forecasts
Links to MetaforecastProjectMetaforecastMetaforecast is a forecast aggregation platform combining 2,100+ questions from 10+ sources (Metaculus, Manifold, Polymarket, etc.) with daily updates via automated scraping. Created by QURI, it pr...Quality: 35/100 for live data
Relationship to QURI Ecosystem
Loading diagram...
Tool
Relationship to Longterm Wiki
SquiggleProjectSquiggleSquiggle is a domain-specific probabilistic programming language optimized for intuition-driven estimation rather than data-driven inference, developed by QURI and adopted primarily in the EA commu...Quality: 41/100
SquiggleAIProjectSquiggleAISquiggleAI is an LLM tool (primarily Claude Sonnet 4.5) that generates probabilistic Squiggle models from natural language, using ~20K tokens of cached documentation to produce 100-500 line models ...Quality: 37/100
LW models could be converted to executable Squiggle estimates
MetaforecastProjectMetaforecastMetaforecast is a forecast aggregation platform combining 2,100+ questions from 10+ sources (Metaculus, Manifold, Polymarket, etc.) with daily updates via automated scraping. Created by QURI, it pr...Quality: 35/100
LW links to relevant forecasts as evidence for claims
Squiggle Hub
Potential future integration for interactive models embedded in pages
QURI (Quantified Uncertainty Research Institute)OrganizationQURI (Quantified Uncertainty Research Institute)QURI develops Squiggle (probabilistic programming language with native distribution types), SquiggleAI (Claude-powered model generation producing 100-500 line models), Metaforecast (aggregating 2,1...Quality: 48/100SquiggleProjectSquiggleSquiggle is a domain-specific probabilistic programming language optimized for intuition-driven estimation rather than data-driven inference, developed by QURI and adopted primarily in the EA commu...Quality: 41/100MetaforecastProjectMetaforecastMetaforecast is a forecast aggregation platform combining 2,100+ questions from 10+ sources (Metaculus, Manifold, Polymarket, etc.) with daily updates via automated scraping. Created by QURI, it pr...Quality: 35/100SquiggleAIProjectSquiggleAISquiggleAI is an LLM tool (primarily Claude Sonnet 4.5) that generates probabilistic Squiggle models from natural language, using ~20K tokens of cached documentation to produce 100-500 line models ...Quality: 37/100About This WikiAbout This WikiTechnical documentation for the Longterm Wiki platform covering content architecture (~550 MDX pages, ~100 entities), quality scoring system (6 dimensions on 0-10 scale), data layer (YAML databases...Quality: 55/100VisionVisionInternal planning document outlining a 2-person-year (3-4 people over 6-8 months) scope for building LongtermWiki as a strategic AI safety prioritization platform, with ~170 planned pages across fo...Quality: 51/100