An AI-authored paper just passed peer review. The scientific community isn’t ready
March 27, 2026
4 min read
Add Us On GoogleAdd SciAm
An AI-authored paper just passed peer review. The scientific community isn’t ready
The arrival of AI-generated research papers marks a turning point that could radically accelerate discovery—or drown it in automated mediocrity
By Jacek Krywko edited by Eric Sullivan
AI can generate research infinitely faster than humans can read it, threatening to bury an already strained peer-review system under a mountain of automated submissions.
Vince Talotta/Toronto Star via Getty Images
Science has always relied on a curious human’s mind forming a hypothesis, designing an experiment, analyzing the results and presenting the case to that person’s peers. Over centuries, we’ve built better tools such as electron microscopes, particle accelerators and supercomputers, but the core loop of scientific discovery has remained stubbornly human. Now, for the first time, that loop has started with a new kind of mind.
So far, scientists have often had artificial intelligence help them with solving a predefined, narrow task such as folding proteins, says Jeff Clune, a professor of computer science at the University of British Columbia. “We’re saying the AI gets to be the scientist,” he says.
In a recent Nature study, Clune and his colleagues unveiled the AI Scientist, an AI system that wrote a paper without human involvement that passed peer review for a workshop at the 2025 International Conference on Learning Representations (ICLR), a top-tier venue in the field of machine learning. The paper was mediocre, according to Clune and other experts. But its existence marks a turning point that the scientific community is only beginning to grapple with: AI has quickly moved from assisting scientists to attempting to be one.
On supporting science journalism
If you're enjoying this article, consider supporting our award-winning journalism by subscribing. By purchasing a subscription you are helping to ensure the future of impactful stories about the discoveries and ideas shaping our world today.
The AI Scientist comprises multiple modules. After it is given a general topic prompt by researchers, it surveys available literature and generates hypotheses. “We’re just giving it a general direction like ‘Come up with something interesting to study on how the AI learns,’” Clune explains. The system then evaluates and refines those ideas, filtering out any that are not novel. From there, further modules plan and execute experiments, analyze and plot the data and, finally, write the paper. It even does its own internal peer review process to find flaws in its papers, Clune says. (The system relies on existing foundation models such as Anthropic’s Claude Sonnet or OpenAI’s GPT-4o; the team’s contribution is the pipeline orchestrating these models).
To see if The AI Scientist’s output could meet human standards, the team submitted three papers generated by it to the I Can’t Believe It’s Not Better (ICBINB) workshop at the 2025 ICLR. One was accepted. (The conference organizers gave their permission for the AI-generated papers to be submitted, and all of the AI Scientist’s papers were withdrawn from the conference after the review process.)
The team behind the AI Scientist admits the bar for this workshop was lower than that of a main conference publication. “Would a mediocre graduate student get one paper in three accepted at a place that accepts 70 percent of papers? Sure!” says Jodi Schneider, an associate professor of information sciences at the University of Wisconsin–Madison, who was not involved in Clune’s study.
The AI’s papers “are okay but not great,” Clune says. To him, some of the AI’s ideas seemed truly creative, yet the system struggled with execution. “The logic and the writing and the thinking throughout the whole paper didn’t all fit together beautifully,” he notes. Further issues included hallucinated references, duplicated figures and a lack of methodological rigor.
Overall, Clune and his colleagues’ new study has received a lukewarm reception. “The approach is agentic and without any real novelty,” says Maria Liakata, a professor of natural language processing at Queen Mary University of London, who was not involved in the work.
There was one metric, though, where the AI Scientist did outperform human researchers by a huge margin: it produced a formally passable paper on machine learning within 15 hours at a cost Clune estimated to be around $140. Compare that with the capability of a graduate student, who might take a full semester to write their first accepted workshop paper, according to Schneider.
As costs drop and output speeds increase, AI-authored papers present the scientific community with an immediate challenge. “The AI-written papers are