Cryptopolitan
2026-02-01 14:24:23

Researchers say AI “slop” is distorting science, push for mandatory disclosure

Scientists working inside the AI research world are facing a credibility problem they can no longer ignore. Major conferences focused on AI research reacted after review systems became clogged with weak submissions. Organizers saw a sharp rise in papers and peer reviews produced with little human effort. The concern is not style. The concern is accuracy. Errors are slipping into places where precision used to matter. Conferences crack down as low-quality papers overwhelm reviewers Researchers warned early that unchecked use of automated writing tools could damage the field. Inioluwa Deborah Raji, an AI researcher at the University of California, Berkeley, said the situation turned chaotic fast. “There is a little bit of irony to the fact that there’s so much enthusiasm for AI shaping other fields when, in reality, our field has gone through this chaotic experience because of the widespread use of AI,” she said. Hard data shows how widespread the problem became. A Stanford University study published in August found that up to 22 percent of computer science papers showed signs of large language model use. Pangram, a text analysis start-up, reviewed submissions and peer reviews at the International Conference on Learning Representations in 2025. It estimated that 21 percent of reviews were fully generated by AI, while more than half used it for tasks like editing. Pangram also found that 9 percent of submitted papers had more than half their content produced this way. The issue reached a tipping point in November. Reviewers at ICLR flagged a paper suspected of being generated by AI that still ranked in the top 17 percent based on reviewer scores. In January, detection firm GPTZero reported more than 100 automated errors across 50 papers presented at NeurIPS, widely seen as the top venue for advanced research in the field. As concerns grew, ICLR updated its usage rules before the conference. Papers that fail to disclose extensive use of language models now face rejection. Reviewers who submit low-quality evaluations created with automation risk penalties, including having their own papers declined. Hany Farid, a computer science professor at the University of California, Berkeley, said “If you’re publishing really low-quality papers that are just wrong, why should society trust us as scientists?” Paper volumes surge while detection struggles to keep up Per the report, NeurIPS received 21,575 papers in 2025, up from 17,491 in 2024 and 9,467 in 2020. One author submitted more than 100 papers in a single year, far beyond what is typical for one researcher. Thomas G. Dietterich, emeritus professor at Oregon State University and chair of the computer science section of arXiv, said uploads to the open repository also rose sharply. Still, researchers say the cause is not simple. Some argue the increase comes from more people entering the field. Others say heavy use of AI tools plays a major role. Detection remains difficult because there is no shared standard for identifying automated text. Dietterich said common warning signs include made-up references and incorrect figures. Authors caught doing this can be temporarily banned from arXiv. Commercial pressure also sits in the background. High-profile demos, soaring salaries, and aggressive competition have pushed parts of the field to focus on quantity. Raji said moments of hype attract outsiders looking for fast results. At the same time, researchers say some uses are legitimate. Dietterich noted that writing quality in papers from China has improved, likely because language tools help rewrite English more clearly. The issue now stretches beyond publishing. Companies like Google, Anthropic , and OpenAI promote their models as research partners that can speed up discovery in areas like life sciences. These systems are trained on academic text. Farid warned that if training data includes too much synthetic material, model performance can degrade. Past studies show large language models can collapse into nonsense when fed uncurated automated data. Farid said companies scraping research have strong incentives to know which papers are human-written. Kevin Weil, head of science at OpenAI, said tools still require human checks. “It can be a massive accelerator,” he said. “But you have to check it. It doesn’t absolve you from rigour.” Join a premium crypto trading community free for 30 days - normally $100/mo.

获取加密通讯
阅读免责声明 : 此处提供的所有内容我们的网站,超链接网站,相关应用程序,论坛,博客,社交媒体帐户和其他平台(“网站”)仅供您提供一般信息,从第三方采购。 我们不对与我们的内容有任何形式的保证,包括但不限于准确性和更新性。 我们提供的内容中没有任何内容构成财务建议,法律建议或任何其他形式的建议,以满足您对任何目的的特定依赖。 任何使用或依赖我们的内容完全由您自行承担风险和自由裁量权。 在依赖它们之前,您应该进行自己的研究,审查,分析和验证我们的内容。 交易是一项高风险的活动,可能导致重大损失,因此请在做出任何决定之前咨询您的财务顾问。 我们网站上的任何内容均不构成招揽或要约