Scientific Knowledge Corruption
Scientific Knowledge Corruption
Documents AI-enabled scientific fraud with evidence that 2-20% of submissions are from paper mills (field-dependent), 300,000+ fake papers exist, and detection tools are losing an arms race against AI generation. Paper mill output doubles every 1.5 years vs. retractions every 3.5 years. Projects 2027-2030 scenarios ranging from controlled degradation (40% probability) to epistemic collapse (20% probability) affecting medical treatments and policy decisions. Wiley/Hindawi scandal resulted in 11,300+ retractions and $35-40M losses.
Quick Assessment
| Dimension | Assessment | Evidence |
|---|---|---|
| Current Scale | 2-20% of published papers potentially fraudulent | PNAS 2025: estimates vary by field; 32,786 papers flagged in Problematic Paper Screener |
| Growth Rate | Doubling every 1.5 years | Paper mill output doubling; retractions doubling only every 3.5 years |
| Detection Gap | 75% of paper mill products never retracted | Only 25-28% of suspected paper mill papers ever retracted |
| AI Detection Accuracy | 14-22% of papers show AI involvement | Science 2024: 22.5% in CS; 14% in biomedicine |
| Publisher Impact | $35-40M lost by single publisher | Wiley lost revenue after retracting 11,300+ Hindawi papers |
| Medical Impact | 11% of meta-analyses change conclusions | PubMed 2025: 51% of reviews potentially affected |
| Trend | Deteriorating rapidly | "Could have more than half of studies fraudulent within a decade" |
Overview
Scientific knowledge corruption represents the systematic degradation of research integrity through AI-enabled fraud, fake publications, and data fabrication. According to PNAS research (2025), paper mill output is doubling every 1.5 years while retractions double only every 3.5 years. Northwestern University researcher Reese Richardson warns: "You can see a scenario in a decade or less where you could have more than half of [studies being published] each year being fraudulent."
This isn't a future threat—it's already happening. Current estimates suggest 2-20% of journal submissions come from paper mills depending on field, with over 300,000 fake papers already in the literature. The Retraction Watch database now contains over 63,000 retractions, with 2023 marking a record high of over 10,000 retractions. AI tools are rapidly industrializing fraud production, creating an arms race between detection and generation that detection appears to be losing.
The implications extend far beyond academia: corrupted medical research could lead to harmful treatments, while fabricated policy research could undermine evidence-based governance and public trust in science itself.
Scientific Corruption Cascade
Diagram (loading…)
flowchart TD AI[AI Text and Image Generation] --> PM[Paper Mills Scale Up] PM --> FP[Flood of Fake Papers] FP --> OD[Overwhelmed Detection] FP --> MA[Corrupted Meta-Analyses] MA --> CG[Unreliable Clinical Guidelines] MA --> PD[Flawed Policy Decisions] CG --> PT[Patient Harm] PD --> RM[Resource Misallocation] OD --> TC[Trust Collapse] TC --> RS[Research Slowdown] style AI fill:#ffcccc style PM fill:#ffcccc style FP fill:#ffcccc style PT fill:#ff9999 style TC fill:#ff9999 style CG fill:#ffddcc style PD fill:#ffddcc
Risk Assessment
| Factor | Assessment | Evidence | Timeline |
|---|---|---|---|
| Current Prevalence | High | 300,000+ fake papers identified | Already present |
| Growth Rate | Accelerating | Paper mill adoption of AI tools | 2024-2026 |
| Detection Capacity | Insufficient | Detection tools lag behind AI generation | Worsening |
| Impact Severity | Severe | Medical/policy decisions at risk | 2025-2030 |
| Trend Direction | Deteriorating | Arms race favors fraudsters | Next 5 years |
Responses That Address This Risk
| Response | Mechanism | Effectiveness |
|---|---|---|
| AI Content Authentication | Cryptographic provenance for research outputs | Medium-High (if adopted) |
| AI-Era Epistemic Security | Systematic protection of knowledge infrastructure | Medium |
| AI-Era Epistemic Infrastructure | Strengthening scientific institutions | Medium |
| Mandatory data sharing | Enables replication and fraud detection | Medium (easy to circumvent) |
| Preregistration requirements | Reduces p-hacking and selective reporting | Low-Medium |
| COPE United2Act | Publisher collaboration on paper mill detection | Early stage |
Current Evidence & Scale
Documented Fraud Levels
| Metric | Current State | Source |
|---|---|---|
| Paper mill submissions | 2-20% of submissions by field | PNAS 2025, Byrne & Christopher (2020)↗📄 paper★★★★★Nature (peer-reviewed)Byrne & Christopher (2020)Relevant to AI safety knowledge bases as a cautionary example of how incentive structures can corrupt research integrity; useful background for thinking about evaluation reliability and governance of AI research publication norms.This Nature article examines the problem of paper mills—organizations that produce fraudulent academic papers for sale—and their impact on scientific integrity and the replicati...scientific-integritypaper-millsreplication-crisisevaluation+1Source ↗ |
| Estimated fake papers | 300,000+ in literature | Cabanac et al. (2022)↗📄 paper★★★☆☆arXivCabanac et al. (2022)A machine learning paper on spatio-temporal traffic prediction using graph neural networks; relevant to AI safety as an example of complex ML system design in safety-critical transportation domains.Guangyin Jin, Fuxian Li, Jinlei Zhang et al. (2022)85 citationsThis paper proposes Auto-DSTSGN, an automated dilated spatio-temporal synchronous graph network for traffic prediction in intelligent transportation systems. The key innovation ...capabilitieseconomicscientific-integritypaper-mills+1Source ↗ |
| Image manipulation | 3.8% of biomedical papers | Bik et al. (2016)↗🔗 webBik et al. (2016) - The Prevalence of Inappropriate Image Duplication in Biomedical Research PublicationsRelevant to AI safety discussions around benchmarking and evaluation integrity; demonstrates systemic vulnerabilities in scientific publishing that parallel concerns about trustworthiness of AI research results and capability evaluations.This large-scale study screened over 20,000 papers across 40 scientific journals and found that 3.8% contained problematic figures with inappropriate image duplication, at least...scientific-integrityreplication-crisisevaluationresearch-misconduct+1Source ↗ |
| Total retractions (2024) | 63,000+ in database | Retraction Watch Database |
| Retractions in 2023 | 10,000+ papers (record high) | Chemistry World |
| AI-assisted content (CS) | 22.5% of abstracts | Science 2024 |
Major Paper Mill Incidents (2023-2025)
| Incident | Scale | Impact | Source |
|---|---|---|---|
| Wiley/Hindawi scandal | 11,300+ papers retracted | $35-40M revenue loss; 19 journals closed | Retraction Watch |
| Europe's largest paper mill | 1,500+ suspect articles | 380 journals affected; Ukraine/Russia/Kazakhstan authors | Science 2024 |
| ARDA India network | 86 journals (up from 14) | 6x growth 2018-2024 | GIJN Investigation |
| PLOS One editor collusion | 49 papers retracted | 0.25% of editors handled 30% of retractions | PNAS 2025 |
| Tortured phrases corpus | 42,500+ papers flagged | Single phrase indicator | Problematic Paper Screener |
AI-Enabled Fraud Detection
| Type | Detection Rate | Challenge |
|---|---|---|
| Tortured phrases | 863,000+ papers flagged | Problematic Paper Screener↗🔗 webProblematic Paper ScreenerTangentially relevant to AI safety: reliable scientific literature is foundational to evidence-based safety research, and tools that identify fraudulent papers help ensure the integrity of the knowledge base AI safety researchers rely on.A web-based tool designed to help researchers identify potentially problematic or fraudulent academic papers, supporting research integrity efforts. It screens papers for indica...evaluationresearch-integrityscientific-integritypaper-mills+3Source ↗ |
| Synthetic images | Growing undetected rate | AI-generated images improving rapidly |
| ChatGPT content | ≈1% of ArXiv submissions | Detection tools unreliable↗📄 paper★★★★★Nature (peer-reviewed)Detection tools unreliableNature article reporting on research showing that AI detection tools are unreliable at identifying AI-generated academic content, highlighting risks to research integrity and the need for improved detection methods.Aditya Vempaty, Bhavya Kailkhura, Pramod K. Varshney (2018)A preprint study found that AI chatbots like ChatGPT can generate research paper abstracts that are convincing enough to fool scientists into believing they are human-written. T...scientific-integritypaper-millsreplication-crisisSource ↗ |
| Fake peer reviews | Unknown scale | Recently discovered at major venues |
Attack Vectors & Mechanisms
Vector 1: Industrialized Paper Mills
Traditional paper mills produce 400-2,000 papers annually. AI-enhanced mills could scale to hundreds of thousands:
| Stage | Traditional | AI-Enhanced |
|---|---|---|
| Text generation | Human ghostwriters | GPT-4/Claude automated |
| Data fabrication | Manual creation | Synthetic datasets |
| Image creation | Photoshop manipulation | Diffusion model generation |
| Citation networks | Manual cross-referencing | Automated citation webs |
Evidence: Paper mills now advertise "AI-powered research services" openly.
Vector 2: Review Process Compromise
| Component | Attack Method | Detection Rate |
|---|---|---|
| Peer review | AI-generated reviews | Unknown (recently discovered) |
| Editorial assessment | Overwhelm with volume | Limited editorial capacity |
| Post-publication review | Fake comments/endorsements | Minimal monitoring |
Vector 3: Preprint Flooding
Preprint servers↗📄 paper★★★☆☆arXivShlegeris et al. (2024)An arxiv preprint by Shlegeris et al. (2024) that likely presents original research relevant to AI safety; arxiv serves as a primary venue for circulating cutting-edge safety research before peer review.monitoringcontainmentdefense-in-depthscientific-integrity+1Source ↗ have minimal review processes, making them vulnerable:
- ArXiv: ~200,000 papers/year, minimal screening
- medRxiv: Medical preprints, used by media/policymakers
- bioRxiv: Biology preprints, influence grant funding
Attack scenario: AI generates 10,000+ fake preprints monthly, drowning real research.
Consequences by Sector
Medical Research Impact
| Risk | Mechanism | Examples |
|---|---|---|
| Ineffective treatments adopted | Fake efficacy studies | Ivermectin COVID studies included fabricated data |
| Drug approval delays | Fake negative studies | Could delay life-saving treatments |
| Clinical guideline corruption | Meta-analyses of fake papers | WHO/CDC guidelines based on literature reviews |
| Patient harm | Treatments based on fake safety data | Direct medical interventions |
Quantified Impact on Medical Evidence
| Metric | Finding | Source |
|---|---|---|
| Meta-analyses with retracted studies | 61 systematic reviews identified | PubMed 2025 |
| Statistical significance changes | 11% of meta-analyses changed after removing retracted studies | PubMed 2025 |
| Reviews with substantially affected findings | 51% likely to change if retracted trials removed | Peer Review Congress |
| Retraction timing | 74% of retractions occur after citation in systematic reviews | PubMed 2025 |
| Affected primary outcomes | 40% of corrupted meta-analyses involved primary outcomes | PubMed 2025 |
Policy & Governance
| Domain | Vulnerability | Potential Impact |
|---|---|---|
| Environmental policy | Climate studies fabricated | Delayed/misdirected climate action |
| Economic policy | Fake impact assessments | Poor resource allocation |
| Education policy | Fabricated intervention studies | Ineffective educational reforms |
| Healthcare policy | Corrupted epidemiological data | Public health failures |
Research Ecosystem
| Impact | Current Trend | Projected 2027 | Source |
|---|---|---|---|
| Research productivity | 10% time waste on fake replication | 30-50% time waste | Expert estimates |
| Funding misallocation | Investigation costs ≈$525K per case | Wiley lost $35-40M in single incident | PLOS Medicine |
| Career advancement | Citation gaming via paper mills | Merit evaluation unreliable | COPE |
| Scientific trust | Declining public confidence | Potential epistemic collapse | Expert consensus |
| Publication volume affected | 10-13% of submissions flagged by Wiley | Could exceed 50% within decade | Retraction Watch |
Detection & Defense Status
Current Detection Tools
| Tool | Capability | Limitations |
|---|---|---|
| Problematic Paper Screener↗🔗 webProblematic Paper ScreenerTangentially relevant to AI safety: reliable scientific literature is foundational to evidence-based safety research, and tools that identify fraudulent papers help ensure the integrity of the knowledge base AI safety researchers rely on.A web-based tool designed to help researchers identify potentially problematic or fraudulent academic papers, supporting research integrity efforts. It screens papers for indica...evaluationresearch-integrityscientific-integritypaper-mills+3Source ↗ | Tortured phrase detection | Arms race; AI improving |
| ImageTwin↗🔗 webImageTwinThis site is currently a placeholder ('coming soon') page with no substantive content; it may be intended as an image duplicate detection tool relevant to research integrity, but cannot be evaluated until launched.ImageTwin appears to be a service related to image detection or verification, likely for identifying duplicate or manipulated images in scientific publications, but the website ...scientific-integrityevaluationred-teamingSource ↗ | Image duplication detection | Limited to exact/near-exact matches |
| Statcheck↗🔗 webStatcheck: Automated Statistical Error CheckerRelevant to AI safety insofar as rigorous empirical standards and reproducibility matter for ML research; Statcheck represents automated oversight tooling for scientific claims, paralleling concerns about verifying AI research results.Statcheck is a free online tool that automatically checks statistical results in research papers for inconsistencies and errors, such as mismatches between reported test statist...scientific-integrityreplication-crisisevaluationresearch-methods+1Source ↗ | Statistical inconsistency detection | Only catches simple errors |
| AI detection tools | Content authenticity | High false positive rates |
Detection Effectiveness
| Method | Success Rate | Challenge | Source |
|---|---|---|---|
| AI text detection (pure AI) | 91-100% accuracy | Degrades with paraphrasing | Frontiers 2024 |
| AI text detection (modified) | 30-50% accuracy | Human editing defeats detection | SAGE 2025 |
| False positive rate (AI detectors) | 1.3% (AI); 5% (humans) | Risk of flagging legitimate work | PMC 2025 |
| Paper mill pre-screening (Wiley) | 10-13% flagged | 600-1,000 papers/month rejected | Retraction Watch |
| Eventual retraction rate | 25-28% of paper mill papers | 72-75% of fake papers remain in literature | PNAS 2025 |
| Peer review fraud detection | 5-15% detection rate | Declining with volume increases | Byrne & Christopher (2020)↗📄 paper★★★★★Nature (peer-reviewed)Byrne & Christopher (2020)Relevant to AI safety knowledge bases as a cautionary example of how incentive structures can corrupt research integrity; useful background for thinking about evaluation reliability and governance of AI research publication norms.This Nature article examines the problem of paper mills—organizations that produce fraudulent academic papers for sale—and their impact on scientific integrity and the replicati...scientific-integritypaper-millsreplication-crisisevaluation+1Source ↗ |
Institutional Responses
| Organization | Response | Status | Source |
|---|---|---|---|
| COPE↗🔗 webCOPE (Committee on Publication Ethics)COPE is relevant to AI safety researchers concerned with research integrity, responsible AI publication norms, and governance of AI-generated content in scholarly literature; its AI-in-publishing guidance may inform norms around AI-assisted research.COPE is a global membership organization promoting ethical standards in scholarly publishing, providing guidance, education, and leadership on issues like research integrity, ed...governancescientific-integritypolicycoordination+2Source ↗ + STM | United2Act initiative; 5 working groups | Launched 2024; ongoing | COPE |
| Retraction Watch↗🔗 webRetraction WatchRetraction Watch is tangentially relevant to AI safety as it documents systemic failures in scientific integrity, which affects the reliability of research AI systems are trained on or that informs AI policy and benchmarking decisions.Retraction Watch is a blog and database that tracks retractions of scientific papers and other issues of research integrity, providing transparency about errors, fraud, and misc...evaluationscientific-integrityreplication-crisispaper-mills+3Source ↗ | Database of 63,000+ retractions; now owned by Crossref | Active monitoring | Crossref |
| STM Integrity Hub | Paper Mill Checker Tool; Duplicate Submission Detection | MVP launched June 2024 | COPE |
| Wiley | 6-tool screening system; 600-1,000 rejections/month | Active since 2024 | Retraction Watch |
| Funding agencies | Data sharing requirements | Easy to circumvent | Various |
Current Trajectory & Projections
2024-2025: Detection Arms Race
- AI detection tools deployment vs. improved AI generation
- Paper mills adopt GPT-4/Claude for content generation
- First major scandals of AI-generated paper acceptance
2025-2027: Scale Transition
- Fraud production scales from thousands to hundreds of thousands annually
- Detection systems overwhelmed
- Research communities begin fragmenting into "trusted" networks
2027-2030: Potential Collapse Scenarios
| Scenario | Probability | Characteristics |
|---|---|---|
| Controlled degradation | 40% | Gradual decline, institutional adaptation |
| Bifurcated system | 35% | "High-trust" vs. "open" research tiers |
| Epistemic collapse | 20% | Public loses confidence in scientific literature |
| Successful defense | 5% | Detection keeps pace with generation |
Key Uncertainties & Research Gaps
Key Questions
- ?What is the true current rate of AI-generated content in scientific literature?
- ?Can detection methods fundamentally keep pace with AI generation, or is this an unwinnable arms race?
- ?At what point does corruption become so pervasive that scientific literature becomes unreliable for policy?
- ?How will different fields (medicine vs. social science) be differentially affected?
- ?What threshold of corruption would trigger institutional collapse vs. adaptation?
- ?Can blockchain/cryptographic methods provide solutions for research integrity?
- ?How will this interact with existing problems like the replication crisis?
Critical Research Needs
| Research Area | Priority | Current Gap |
|---|---|---|
| Baseline measurement | High | Unknown true fraud rates |
| Detection technology | High | Fundamental limitations unclear |
| Institutional resilience | Medium | Adaptation capacity unknown |
| Cross-field variation | Medium | Differential impact modeling |
| Public trust dynamics | Medium | Tipping point identification |
Related Risks & Interactions
This risk intersects with several other epistemic risks:
- Epistemic collapse: Scientific corruption could trigger broader epistemic system failure
- Expertise atrophy: Researchers may lose skills if AI does the work
- Trust cascade: Scientific fraud could undermine trust in all expertise
Sources & Resources
Research Organizations
| Organization | Focus | Key Resource |
|---|---|---|
| Retraction Watch↗🔗 webRetraction WatchRetraction Watch is tangentially relevant to AI safety as it documents systemic failures in scientific integrity, which affects the reliability of research AI systems are trained on or that informs AI policy and benchmarking decisions.Retraction Watch is a blog and database that tracks retractions of scientific papers and other issues of research integrity, providing transparency about errors, fraud, and misc...evaluationscientific-integrityreplication-crisispaper-mills+3Source ↗ | Fraud monitoring | Database of 38,000+ retractions↗🔗 webRetraction Watch DatabaseRelevant to AI safety primarily as a tool for auditing training data quality and scientific reliability; highlights systemic issues in peer review that could affect AI systems trained on scientific literature.The Retraction Watch Database is a searchable repository tracking retractions, expressions of concern, and corrections across scientific literature. It documents the reasons beh...evaluationscientific-integrityreplication-crisisdataset+2Source ↗ |
| Committee on Publication Ethics↗🔗 webCOPE (Committee on Publication Ethics)COPE is relevant to AI safety researchers concerned with research integrity, responsible AI publication norms, and governance of AI-generated content in scholarly literature; its AI-in-publishing guidance may inform norms around AI-assisted research.COPE is a global membership organization promoting ethical standards in scholarly publishing, providing guidance, education, and leadership on issues like research integrity, ed...governancescientific-integritypolicycoordination+2Source ↗ | Publishing ethics | Fraud detection guidelines↗🔗 webFraud detection guidelinesCOPE guidance is tangentially relevant to AI safety through its role in preserving the integrity of scientific literature that AI research depends on; most directly useful for those concerned with publication ethics, replication, and fraud in research contexts.The Committee on Publication Ethics (COPE) provides comprehensive guidelines for editors, authors, and reviewers on handling research misconduct, fraud detection, and ethical pu...scientific-integrityevaluationgovernancepolicy+1Source ↗ |
| For Better Science↗🔗 webFor Better ScienceTangentially relevant to AI safety as unreliable scientific literature affects training data quality and benchmark integrity; primarily a biomedical misconduct watchdog blog with limited direct AI safety content.For Better Science is an investigative blog by Leonid Schneider that exposes scientific misconduct, data fraud, paper mills, and integrity failures across biomedical and life sc...evaluationscientific-integrityreplication-crisispaper-mills+2Source ↗ | Fraud investigation | Independent fraud research |
| PubPeer↗🔗 webPubPeer - Search publications and join the conversation.PubPeer is tangentially relevant to AI safety via its role in maintaining scientific integrity; ML researchers can use it to flag or investigate concerns about published AI/ML papers, supporting reliable empirical foundations for the field.PubPeer is an online platform enabling post-publication peer review, allowing researchers to comment on and critique published scientific papers anonymously. It has become a maj...evaluationscientific-integrityreplication-crisispaper-mills+2Source ↗ | Post-publication review | Community-driven quality control |
Key Academic Research
| Study | Findings | Source |
|---|---|---|
| Fanelli (2009) | 2% scientists admit fabrication | PLOS ONE↗🔗 webPLOS ONEA peer-reviewed journal article from PLOS ONE that may contain research relevant to AI safety, depending on its specific topic and methodology.6 citations · PLOS ONEscientific-integritypaper-millsreplication-crisisSource ↗ |
| Cabanac et al. (2022) | 300,000+ fake papers estimated | arXiv↗📄 paper★★★☆☆arXivCabanac et al. (2022)A machine learning paper on spatio-temporal traffic prediction using graph neural networks; relevant to AI safety as an example of complex ML system design in safety-critical transportation domains.Guangyin Jin, Fuxian Li, Jinlei Zhang et al. (2022)85 citationsThis paper proposes Auto-DSTSGN, an automated dilated spatio-temporal synchronous graph network for traffic prediction in intelligent transportation systems. The key innovation ...capabilitieseconomicscientific-integritypaper-mills+1Source ↗ |
| Ioannidis (2005) | "Why Most Research Findings Are False" | PLOS Medicine↗🔗 webWhy Most Published Research Findings Are FalseHighly relevant to AI safety researchers evaluating empirical claims about model capabilities or alignment techniques, as the statistical biases described apply directly to ML benchmarking and safety evaluation literature.Sebastian Lobentanzer (2020)John Ioannidis's landmark 2005 paper demonstrates mathematically that the majority of published research findings are likely false positives, due to low statistical power, publi...evaluationscientific-integrityreplication-crisisresearch-methods+2Source ↗ |
| Bik et al. (2016) | 3.8% image manipulation rate | mBio↗🔗 webBik et al. (2016) - The Prevalence of Inappropriate Image Duplication in Biomedical Research PublicationsRelevant to AI safety discussions around benchmarking and evaluation integrity; demonstrates systemic vulnerabilities in scientific publishing that parallel concerns about trustworthiness of AI research results and capability evaluations.This large-scale study screened over 20,000 papers across 40 scientific journals and found that 3.8% contained problematic figures with inappropriate image duplication, at least...scientific-integrityreplication-crisisevaluationresearch-misconduct+1Source ↗ |
Detection & Monitoring Tools
| Tool | Function | Access |
|---|---|---|
| Problematic Paper Screener↗🔗 webProblematic Paper ScreenerTangentially relevant to AI safety: reliable scientific literature is foundational to evidence-based safety research, and tools that identify fraudulent papers help ensure the integrity of the knowledge base AI safety researchers rely on.A web-based tool designed to help researchers identify potentially problematic or fraudulent academic papers, supporting research integrity efforts. It screens papers for indica...evaluationresearch-integrityscientific-integritypaper-mills+3Source ↗ | Tortured phrase detection | Public database |
| ImageTwin↗🔗 webImageTwinThis site is currently a placeholder ('coming soon') page with no substantive content; it may be intended as an image duplicate detection tool relevant to research integrity, but cannot be evaluated until launched.ImageTwin appears to be a service related to image detection or verification, likely for identifying duplicate or manipulated images in scientific publications, but the website ...scientific-integrityevaluationred-teamingSource ↗ | Image duplication | Web interface |
| Statcheck↗🔗 webStatcheck: Automated Statistical Error CheckerRelevant to AI safety insofar as rigorous empirical standards and reproducibility matter for ML research; Statcheck represents automated oversight tooling for scientific claims, paralleling concerns about verifying AI research results.Statcheck is a free online tool that automatically checks statistical results in research papers for inconsistencies and errors, such as mismatches between reported test statist...scientific-integrityreplication-crisisevaluationresearch-methods+1Source ↗ | Statistical consistency | R package |
| Crossref Event Data↗🔗 webCrossref Event DataCrossref Event Data is a reference tool for monitoring scholarly citation activity; it is tangentially relevant to AI safety primarily through its use in assessing scientific integrity and detecting paper mills or citation manipulation in AI research literature.Crossref Event Data is a service that tracks and aggregates online activity and discussions around scholarly content, collecting data on how research is referenced, shared, and ...scientific-integrityevaluationdatasetcoordination+1Source ↗ | Citation monitoring | API access |
Policy & Guidelines
| Resource | Organization | Focus |
|---|---|---|
| COPE Guidelines↗🔗 webFraud detection guidelinesCOPE guidance is tangentially relevant to AI safety through its role in preserving the integrity of scientific literature that AI research depends on; most directly useful for those concerned with publication ethics, replication, and fraud in research contexts.The Committee on Publication Ethics (COPE) provides comprehensive guidelines for editors, authors, and reviewers on handling research misconduct, fraud detection, and ethical pu...scientific-integrityevaluationgovernancepolicy+1Source ↗ | Committee on Publication Ethics | Publisher guidance |
| Singapore Statement↗🔗 webSingapore Statement on Research IntegrityThis statement is a widely-cited international standard for research integrity; relevant to AI safety insofar as honest research practices, transparency, and accountability norms apply to AI research institutions and publication standards.The Singapore Statement on Research Integrity is the first international effort to establish unified principles and responsibilities for research integrity worldwide, developed ...scientific-integritygovernancepolicycoordination+1Source ↗ | World Conference on Research Integrity | Research integrity principles |
| NIH Guidelines↗🏛️ governmentNIH Guidelines on Research Misconduct (42 CFR 93)This is a foundational regulatory reference for understanding how research misconduct is formally defined in federally funded science; relevant to AI safety insofar as AI research integrity and reproducibility concerns intersect with these standards.The NIH defines research misconduct under Public Health Service policies as fabrication, falsification, or plagiarism in any stage of research, explicitly excluding honest error...governancepolicyevaluationscientific-integritySource ↗ | National Institutes of Health | US federal research standards |
| EU Code of Conduct↗🔗 web★★★★☆European UnionEuropean Code of Conduct for Research Integrity (Horizon)This EU institutional document is tangentially relevant to AI safety as a policy precedent for research governance, but primarily concerns general scientific integrity rather than AI-specific safety concerns; useful for policy and governance discussions around research accountability.This EU Horizon 2020 document establishes ethical principles and conduct standards for researchers funded under EU research programs. It outlines obligations around research int...governancepolicyscientific-integrityevaluation+1Source ↗ | European Commission | Research integrity framework |
References
This EU Horizon 2020 document establishes ethical principles and conduct standards for researchers funded under EU research programs. It outlines obligations around research integrity, data management, and responsible innovation to ensure publicly funded science meets high ethical standards. The code addresses issues like scientific misconduct, transparency, and accountability in European research.
John Ioannidis's landmark 2005 paper demonstrates mathematically that the majority of published research findings are likely false positives, due to low statistical power, publication bias, and researcher degrees of freedom. Using probability modeling, it shows that under common research conditions the post-study probability of a true finding is frequently below 50%. This work catalyzed the modern replication crisis movement across scientific disciplines.
COPE is a global membership organization promoting ethical standards in scholarly publishing, providing guidance, education, and leadership on issues like research integrity, editorial independence, and AI in publishing. With over 14,500 members across 97 countries, it serves as a central authority for publication ethics norms. Its resources include guidelines, position statements, and discussion documents relevant to research integrity challenges.
The Singapore Statement on Research Integrity is the first international effort to establish unified principles and responsibilities for research integrity worldwide, developed at the 2nd World Conference on Research Integrity in 2010. It was produced collaboratively by 340 participants from 51 countries and aims to encourage governments, institutions, and researchers to develop comprehensive standards and codes of conduct promoting honest research globally.
ImageTwin appears to be a service related to image detection or verification, likely for identifying duplicate or manipulated images in scientific publications, but the website is currently in a 'coming soon' state with no substantive content available.
This large-scale study screened over 20,000 papers across 40 scientific journals and found that 3.8% contained problematic figures with inappropriate image duplication, at least half showing signs of deliberate manipulation. The prevalence has risen markedly over the past decade, and journal-level practices like prepublication image screening appear to influence data quality.
For Better Science is an investigative blog by Leonid Schneider that exposes scientific misconduct, data fraud, paper mills, and integrity failures across biomedical and life sciences research. It critically examines problematic publications, institutional cover-ups, and the broader replication crisis. The site serves as a watchdog resource for accountability in academic publishing.
Statcheck is a free online tool that automatically checks statistical results in research papers for inconsistencies and errors, such as mismatches between reported test statistics, degrees of freedom, and p-values. It helps researchers and reviewers identify potential errors in null-hypothesis significance testing (NHST) results. The tool supports efforts to improve scientific integrity and reproducibility.
Crossref Event Data is a service that tracks and aggregates online activity and discussions around scholarly content, collecting data on how research is referenced, shared, and discussed across the web. It provides an open dataset of events linking scholarly works to online sources such as social media, Wikipedia, and news outlets. This helps researchers and institutions understand the broader impact and reach of academic publications.
12Detection tools unreliableNature (peer-reviewed)·Aditya Vempaty, Bhavya Kailkhura & Pramod K. Varshney·2018·Paper▸
A preprint study found that AI chatbots like ChatGPT can generate research paper abstracts that are convincing enough to fool scientists into believing they are human-written. The research, posted on bioRxiv in December 2022, demonstrates that current detection methods are unreliable at identifying AI-generated academic content. This finding has sparked debate within the scientific community about the implications for research integrity and the need for better detection tools or policies to address AI-generated submissions.
The Committee on Publication Ethics (COPE) provides comprehensive guidelines for editors, authors, and reviewers on handling research misconduct, fraud detection, and ethical publishing practices. It serves as a central resource for maintaining integrity in academic publishing. The guidance covers issues such as paper mills, plagiarism, data fabrication, and peer review manipulation.
This Nature article examines the problem of paper mills—organizations that produce fraudulent academic papers for sale—and their impact on scientific integrity and the replication crisis. It discusses how systematic fabrication of research undermines trust in published science and proposes strategies for detection and prevention.
Retraction Watch is a blog and database that tracks retractions of scientific papers and other issues of research integrity, providing transparency about errors, fraud, and misconduct in academic publishing. It serves as a critical resource for understanding the scale and nature of problems in scientific literature, including paper mills and reproducibility failures. The site maintains a searchable database of over 45,000 retracted papers.
PubPeer is an online platform enabling post-publication peer review, allowing researchers to comment on and critique published scientific papers anonymously. It has become a major venue for detecting research fraud, data manipulation, image duplication, and paper mill activity. The platform plays a significant role in scientific accountability and the broader replication crisis discourse.
This paper proposes Auto-DSTSGN, an automated dilated spatio-temporal synchronous graph network for traffic prediction in intelligent transportation systems. The key innovation is an automated graph structure search approach that dynamically constructs the spatio-temporal graph to adapt to different data scenarios, rather than using fixed graph construction methods. The model uses dilated layers with increasing dilation factors to capture both short-term and long-term spatio-temporal dependencies more effectively. Experiments on four real-world datasets demonstrate approximately 10% performance improvements over state-of-the-art methods.
The NIH defines research misconduct under Public Health Service policies as fabrication, falsification, or plagiarism in any stage of research, explicitly excluding honest error or differences of opinion. The guidelines establish the three core categories with specific examples, forming the regulatory backbone for research integrity enforcement in federally funded science.
A web-based tool designed to help researchers identify potentially problematic or fraudulent academic papers, supporting research integrity efforts. It screens papers for indicators associated with paper mills, fabricated data, or other forms of scientific misconduct. The tool contributes to combating the replication crisis and improving the reliability of the scientific literature.
The Retraction Watch Database is a searchable repository tracking retractions, expressions of concern, and corrections across scientific literature. It documents the reasons behind retractions—including fraud, data fabrication, and plagiarism—serving as a key resource for assessing scientific integrity. The database supports researchers, journalists, and institutions in monitoring the reliability of published science.
The Retraction Watch Database is a comprehensive, searchable repository tracking retracted scientific papers across disciplines. It provides transparency into the scientific correction process by cataloging retractions, expressions of concern, and corrections with reasons such as fraud, error, or plagiarism. It serves as a critical resource for researchers verifying the integrity of cited literature.