Longterm Wiki

Suggested Pages

100 pages the wiki should add, ranked by priority. Priority is based on how often the topic is mentioned across existing pages and its importance to AI safety coverage. Sourced from the Feb 2026 gap analysis.

100 of 100 results
Why
100AI Governanceconcept183Mentioned on 183 pages. Sub-topics exist (compute governance, etc.) but no umbrella page.
99Reinforcement Learningconcept81Mentioned on 81 pages. Underpins RLHF, reward modeling, and alignment methods — no standalone page.
98Retrieval-Augmented Generation (RAG)concept404Referenced on 404 pages — most-mentioned concept without a page.
97GPT-4capability193Central reference point for frontier capabilities, mentioned on 193 pages.
96Claude (Model Family)capability186Anthropic's flagship model, mentioned on 186 pages, no standalone page.
95Frontier Model (Concept)concept161The concept of 'frontier model' is used on 161 pages with no definition page.
94Training Dataconcept111Fundamental topic (curation, bias, consent, copyright) — 111 page mentions.
93Safety Evaluationsresponse106Referenced on 106 pages; evals are how labs demonstrate safety.
92Misalignment Potentialai-transition-model110110 EntityLinks to this transition-model factor, no page exists.
91Civilizational Competenceai-transition-model104104 EntityLinks to this transition-model factor, no page exists.
90Model Evaluation (Methodology)response55Referenced on 55 pages. Capability evals, dangerous-capability evals, eval science — no methodology page.
89Fine-Tuningconcept89Key technique for adapting models; safety implications of open fine-tuning.
88Gemini (Google DeepMind)capability88Google's frontier model family, mentioned on 88 pages.
87Llama (Meta)capability82Most widely used open-weights model. 82 page mentions.
86DeepSeekorganization71Chinese frontier lab (R1, V3) — changed compute-efficiency assumptions globally.
85GPT-5 / Next-Gen OpenAIcapability67Frequently referenced as next capability milestone, 67 page mentions.
84Transformer Architectureconcept65The architecture underlying all frontier models; no explainer page.
83Jailbreaking & Prompt Injectionrisk81Primary attack vector against deployed LLMs. Covers direct jailbreaks + indirect injection in tool-use contexts.
82Multimodal AIcapability49Vision/audio models have distinct safety challenges. 49 mentions.
81Training Runs & Compute Costconcept47Economics of training — cost, duration, environmental impact. 47 mentions.
80Foundation Model (Concept)concept46Distinct from 'frontier model' — the general category. 46 mentions.
79Hallucinationrisk41Most user-visible AI failure mode. 41 mentions, no dedicated page.
78AI Chips & Hardwareconcept39GPU/TPU/custom silicon — hardware is a key governance lever. 39 mentions.
77Semiconductor Industryconcept35Supply chain chokepoints (TSMC, ASML). 35 mentions.
76Grok (xAI)capability33xAI's model. 33 mentions; no entity or page.
75Embeddings & Vector Searchconcept33How models represent knowledge. 33 mentions.
74AI Incidents Databaseincidents32Only 2 incident pages exist. Need a comprehensive tracker.
73Benchmarks & Leaderboardsconcept28How capabilities are measured; gaming and limitations. 28 mentions.
72Mistral AIorganization27Leading European frontier lab. Important for EU AI Act context.
71DPO & RLHF Alternativesresponse27DPO, IPO, KTO — alternatives to RLHF for alignment. 27 mentions.
70Transition Turbulenceai-transition-model2626 EntityLinks to this transition-model factor, no page.
69Synthetic Dataconcept24Self-play and synthetic training data — model collapse risk. 24 mentions.
68Model Weights (Security & Access)concept40Referenced on 40 pages. Weight theft, open release decisions, proliferation risk — no standalone page.
67Pre-Trainingconcept21The initial training phase. Distinct safety considerations from fine-tuning.
66Knowledge Distillationconcept20Compressing large models; safety properties may not transfer. 20 mentions.
65Post-Training (RLHF, Safety)concept20Where safety alignment happens in practice. 20 mentions.
64Content Provenance & C2PAresponse19Technical countermeasure to deepfakes and AI content. 19 mentions.
63Misuse Potentialai-transition-model1818 EntityLinks to this transition-model factor, no page.
62AI Watermarkingresponse18SynthID, text watermarks — detection of AI-generated content.
61Data Annotation & AI Laborconcept18Ghost work, RLHF annotators, labor conditions. 18 mentions.
60Intelligence Explosionconcept17Core AI safety concept (Good, Bostrom). 17 mentions, no page.
59Voice Cloningrisk17Fraud, impersonation, consent issues. 17 mentions.
58Model Cards & Documentationresponse17Standard disclosure format for AI models. 17 mentions.
57Context Windowsconcept17Key capability dimension (4k to 1M+) with safety implications.
56TSMCorganization16Single point of failure for advanced chips. 16 mentions.
55Epoch AIorganization1616 EntityLinks to this org — key data source for AI trends.
54Open Weightsconcept16Distinct from 'open source' — weights-only release model.
53Attention Mechanismconcept14Core transformer component. 14 mentions.
52Capability Overhangconcept13When existing hardware can run much more capable models. 13 mentions.
51Test-Time Compute & Reasoningcapability11o1/o3/R1 inference-scaling paradigm — changes safety assumptions.
50Chinchilla Scalingconcept11Compute-optimal training. 11 mentions.
49Hugging Faceorganization10Central hub for open-weights models and datasets.
48Knowledge Graphs for AIconcept10Structured knowledge + LLMs. 10 mentions.
47Alignment Taxconcept9Cost of making models safe vs. capable. Key policy concept.
46Image & Video Generationcapability16Diffusion models, DALL-E, Midjourney, Sora — architecture and safety issues.
45AI Auditingresponse9Third-party safety audits. Emerging profession.
44Data Poisoningrisk8Supply-chain attack on training data. Distinct from adversarial examples.
43Brain Emulationconcept8Whole brain emulation as alternative path to AGI. 8 mentions.
42Algorithmic Biasrisk66 dangling EntityLinks. Needs entity + page.
41Model Collapserisk7Training on AI-generated data degrades quality. Emerging research.
40AI Consciousness & Moral Statusconcept7Sentience, moral patienthood, digital minds. 7 mentions.
39AI Technical Standardsresponse38Referenced on 38 pages. ISO, NIST, IEEE frameworks — how standards interact with regulation.
38Chain-of-Thought Reasoningconcept6Prompting technique that elicits reasoning. Safety implications.
37Function Calling & Tool Usecapability6Agentic capability — models invoking APIs. Security concerns.
36Differential Privacyresponse6Mathematical privacy guarantees for training data.
35Regulatory Arbitragerisk55 dangling EntityLinks. Companies choosing least-regulated jurisdictions.
34AI Liability & Legal Frameworksresponse5Who pays when AI causes harm? Foundational governance question.
33NVIDIAorganization44 dangling EntityLinks. Dominant AI chip supplier.
32Compliance Costsconcept44 dangling EntityLinks. Cost of regulation for AI companies.
31Chinese AI Ecosystemconcept4Baidu, Alibaba, Tencent, ByteDance — different safety norms.
30Reward Modelingresponse3Positive framing of reward specification. Complements reward-hacking page.
29Model Merging & Weight Manipulationrisk3Open-source technique to combine or modify model capabilities.
28AI Supply Chainconcept3End-to-end: data, compute, training, deployment. Chokepoints.
27Post-Deployment Monitoringresponse2Runtime safety monitoring. Most safety work is pre-deployment.
26Federated Learningconcept2Privacy-preserving training across distributed data.
25AI Energy & Environmental Impactconcept2Data center power, water use, carbon footprint of training.
24Compute Governance Trackingmetric2Are compute thresholds actually enforced? No tracking page.
23Foundation Model Commoditizationmodel2Pricing collapse changes lab safety incentives.
22In-Context Learningconcept2How LLMs learn from prompts. Safety implications for elicitation.
21AI-Enabled Scientific Fraudrisk2Paper mills, fabricated data, fake peer reviews.
20Speculative Decodingconcept1Inference optimization affecting deployment safety properties.
19Shane LeggresearcherDeepMind co-founder. Entity exists, no page.
18Nate SoaresresearcherMIRI Executive Director. Entity exists, no page.
17Beth BarnesresearcherFounded METR (Model Evaluation & Threat Research). Entity exists, no page.
16Gary MarcusresearcherProminent AI critic and public commentator. Entity exists, no page.
15Ian HogarthresearcherChair of UK AI Safety Institute. Entity exists, no page.
14Buck ShlegerisresearcherCEO of Redwood Research. Entity exists, no page.
13Elizabeth KellyresearcherDirector of US AI Safety Institute. Entity exists, no page.
12ARC EvaluationsorganizationEntity exists (arc-evals), no page. Key eval org.
11Redwood ResearchorganizationEntity 'redwood' exists separately from redwood-research page — may need merge or separate page.
10Homomorphic Encryption for AIconcept1Privacy-preserving inference. Niche but growing.
9Tokenizationconcept3How text becomes model input. Affects multilingual safety.
8Deepfake Detectionresponse3Technical countermeasure to deepfakes. Detection arms race.
7AI Copyright & Fair Useconcept3Training data rights, output ownership. Active litigation.
6Catastrophic Forgettingconcept2Models lose capabilities during fine-tuning. Safety implications.
5Mixture of Expertsconcept2Architecture used by GPT-4, Mixtral. Efficiency vs. safety.
4AI Labor Displacement (Empirical)metric1Tracking actual job impacts as of 2026.
3Red-Teaming-as-a-Serviceresponse1Commercial red-teaming offerings and effectiveness.
2Continual Learningconcept1Models that learn after deployment. Safety of ongoing adaptation.
1AI Military & Intelligence Applicationsconcept1Beyond autonomous weapons — broader military AI use.