ArXiv, a widely used open repository for preprint research, is doing more to crack down on the careless use of large language models in scientific papers.
Although papers are posted to the site before they are peer-reviewed, arXiv (pronounced “archive”) has become one of the main ways that research circulates in fields like computer science and math, and the site itself has become a source of data on trends in scientific research.
ArXiv has already taken steps to combat a growing number of low-quality, AI-generated papers, for example by requiring first-time posters to get an endorsement from an established author. And after being hosted by Cornell for more than 20 years, the organization is becoming an independent nonprofit, which should allow it to raise more money to address issues like AI slop.
In its latest move, Thomas Dietterich — the chair of arXiv’s computer science section — posted Thursday that “if a submission contains incontrovertible evidence that the authors did not check the results of LLM generation, this means we can’t trust anything in the paper.”
That incontrovertible evidence could include things like “hallucinated references” and comments to or from the LLM, Dietterich said. If such evidence is found, a paper’s authors will face “a 1-year ban from arXiv followed by the requirement that subsequent arXiv submissions must first be accepted by a reputable peer-reviewed venue.”
Note that this isn’t an outright prohibition on using LLMs, but rather an insistence that, as Dietterich put it, authors take “full responsibility” for the content, “irrespective of how the contents are generated.” So if researchers copy-paste “inappropriate language, plagiarized content, biased content, errors, mistakes, incorrect references, or misleading content” directly from an LLM, then they’re still responsible for it.
Dietterich told 404 Media that this will be a “one-strike” rule, but moderators must flag the issue and section chairs must confirm the evidence before imposing the penalty. Authors will also be able to appeal the decision.
Recent peer-reviewed research has found that fabricated citations are on the rise in biomedical research, likely due to LLMs — though to be fair, scientists aren’t the only ones getting caught using citations that were made up by AI.
When you purchase through links in our articles, we may earn a small commission. This doesn’t affect our editorial independence.
arXiv Draws a Hard Line: New ‘One-Strike’ Rule Cracks Down on AI ‘Slop’ in Scientific Preprints
The venerable preprint server tightens its policies, issuing strict penalties for researchers who fail to properly vet AI-generated content.
Key Takeaways:
- Zero Tolerance for Unchecked AI: arXiv will impose a 1-year ban on authors whose submissions show “incontrovertible evidence” of unchecked Large Language Model (LLM) generation.
- Hallucinations & AI Artifacts are Key: Proof points for violations include fabricated references and direct LLM interaction artifacts within the paper’s text.
- Responsibility Remains Human: The policy isn’t an outright ban on LLMs, but a firm insistence that authors bear full accountability for all content, regardless of how it was generated.
In the rapidly evolving landscape of scientific research, the advent of Large Language Models (LLMs) like ChatGPT has introduced both unprecedented opportunities and significant challenges. While these AI tools promise to accelerate discovery and aid in scientific communication, their propensity for “hallucination” and generating plausible-sounding falsehoods poses a serious threat to academic integrity. At the forefront of addressing this burgeoning problem is arXiv, the world’s leading open-access repository for preprint research, which has just announced a significant tightening of its policies.
For over two decades, arXiv (pronounced “archive”) has served as a crucial nerve center for disseminating cutting-edge research in fields such as computer science, mathematics, physics, and more, long before formal peer review. Its influence is undeniable, shaping research trends and providing a real-time pulse on scientific progress. However, this open-door policy, a cornerstone of its success, now faces a formidable adversary: the proliferation of low-quality, AI-generated content, or what some are calling “AI slop.”
The Growing Threat of AI Slop in Academia
The problem isn’t theoretical. Researchers are increasingly encountering instances of AI-generated misinformation seeping into academic papers, particularly in the form of fabricated citations. A recent peer-reviewed study highlighted a worrying rise in “hallucinated references” in biomedical research, directly attributing this surge to the careless use of LLMs. This isn’t just a minor annoyance; it erodes the foundational trust upon which scientific discourse is built. When researchers can no longer rely on the veracity of cited sources, the entire edifice of cumulative knowledge is jeopardized.
While arXiv has previously implemented measures, such as requiring endorsements from established authors for first-time posters, the latest move signals a more aggressive stance against what threatens to become an epidemic. The urgency is palpable, prompting the platform to evolve not just its rules, but its very organizational structure, transitioning from its long-standing home at Cornell to become an independent nonprofit. This strategic shift aims to bolster its financial resources, enabling it to better tackle complex challenges like policing AI-generated content.
arXiv’s New Zero-Tolerance Policy: The “One-Strike” Rule
The core of arXiv’s new crackdown comes directly from Thomas Dietterich, the chair of its influential computer science section. In a recent public statement, Dietterich made it unequivocally clear: “if a submission contains incontrovertible evidence that the authors did not check the results of LLM generation, this means we can’t trust anything in the paper.” This stark declaration underscores the severity of the issue and the potential for a cascading loss of credibility once AI-generated inaccuracies are allowed to propagate.
The penalties for such transgressions are equally severe. Authors found to have submitted unchecked AI-generated content will face a “one-strike” rule: a stringent 1-year ban from arXiv. Furthermore, any future submissions post-ban will come with the prerequisite of first being accepted by a reputable peer-reviewed venue – a significant hurdle for any researcher accustomed to the rapid dissemination offered by preprint servers.
What Constitutes “Incontrovertible Evidence”?
For researchers wondering where the line is drawn, Dietterich provided concrete examples of what arXiv considers “incontrovertible evidence.” These include:
- Hallucinated References: Citations that simply do not exist, a common byproduct of LLM generation when prompted to invent sources.
- Comments to or from the LLM: Direct interactions or residual prompts within the paper’s text that reveal AI assistance was not adequately reviewed or removed.
These indicators are not just stylistic oversights; they are glaring red flags that suggest a fundamental failure on the part of the authors to exercise due diligence. The process for imposing these penalties involves moderator flagging, confirmation by section chairs, and a built-in author appeal mechanism, ensuring a degree of fairness, but the underlying message is clear: vigilance is paramount.
Human Responsibility in the Age of AI
It’s crucial to understand that arXiv’s new policy is not an outright prohibition on the use of LLMs in research. Instead, it is a powerful reaffirmation of author responsibility. As Dietterich articulated, authors must take “full responsibility” for the content of their submissions, “irrespective of how the contents are generated.” This means that if an LLM is used to draft text, generate ideas, or even assist with data analysis, the human authors are ultimately accountable for any “inappropriate language, plagiarized content, biased content, errors, mistakes, incorrect references, or misleading content” that results.
This nuanced stance recognizes AI as a powerful tool, much like a calculator or a word processor, but one that requires human oversight and critical evaluation. The convenience offered by LLMs cannot come at the cost of accuracy or integrity. Researchers are now on notice that while AI can assist, the intellectual heavy lifting of verification and validation remains firmly in their court.
Broader Implications for Scientific Publishing
arXiv’s decisive action sends a strong signal across the entire scientific publishing ecosystem. As AI capabilities continue to advance, similar challenges are bound to emerge in peer-reviewed journals and other academic platforms. The precedents set by arXiv could very well influence how other institutions adapt to the presence of LLMs in research generation and dissemination. Maintaining trust in scientific output, especially in an era rife with misinformation, is more critical than ever.
The move also highlights the proactive efforts arXiv is undertaking as it transitions to an independent nonprofit. This organizational evolution is not merely administrative; it’s a strategic repositioning designed to provide the necessary resources and agility to address contemporary challenges head-on. Securing more funding and operational independence will be vital in developing sophisticated detection mechanisms and refining policies as AI technology continues its rapid progression.
Bottom Line
arXiv’s new “one-strike” policy marks a critical moment in the ongoing battle to preserve the integrity of scientific research in the age of artificial intelligence. By drawing a clear line against unchecked AI “slop” and firmly placing the onus of responsibility on human authors, the platform is setting a vital standard for academic rigor. This assertive stance not only protects the credibility of arXiv’s vast repository but also serves as a crucial reminder to the global scientific community: while AI offers revolutionary tools, the bedrock principles of accuracy, verification, and human accountability remain non-negotiable.
Source: {feed_title}

