Is AI poisoning the scientific literature? Our comment in Nature / Jul 2025 / DOI
For the past few years,

Our work on literature reviews led us into assessing methods for evidence synthesis (which is crucial to rational policymaking!) and specifically about how recent advances in AI may impact it. The current methods for rigorous systematic literature review are expensive and slow, and authors are already struggling to keep up with the rapidly expanding number of legitimate papers. Adding to this, paper retractions are increasing near exponentially and already systematic reviews unknowingly cite retracted papers, with most remaining uncorrected even a year (after notification!)
This is all made much more complex as LLMs are flooding the landscape with convincing, fake manuscripts and doctored data, potentially overwhelming our current ability to distinguish fact from fiction. Just this March, the AI Scientist formulated hypotheses, designed and ran experiments, analysed the results, generated the figures and produced a manuscript that passed human peer review for an ICLR workshop! Distinguishing genuine papers from those produced by LLMs isn't just a problem for review authors; it's a threat to the very foundation of scientific knowledge. And meanwhile, Google is taking a different tack with a collaborative AI co-scientist who acts as a multi-agent assistant.
So the landscape is moving really quickly! Our proposal for the future of
literature reviews builds on our desire to move towards a more regional,
federated network approach. Instead of having giant repositories of knowledge
that may be erased unilaterally,
we're aiming for a more bilateral network of "living evidence databases".
Every government, especially those in the Global South, should have the ability to build their
own "
This system of living evidence databases can be incremental and dynamically
updated, and AI assistance can be used as long as humans remain in-the-loop.
Such a system can continuously gather, screen, and index literature,
automatically remove compromised studies and recalculating results. We're
working on this on multiple fronts this year; ranging from the computer science
to figure out the distributed-nitty-gritty My instinct is that we'll end up with something ATProto based as it's so convenient for distributed system authentication.
Read our Nature Comment piece (comment on LI) to learn more about how we think we can safeguard evidence synthesis against the rising tide of "AI-poisoned literature" and ensure the continued integrity of scientific discovery. As a random bit of trivia, the incredibly cool artwork in the piece was drawn by the legendary David Parkins, who also drew Beano and Dennis the Menace!
Read more about
References
- Reynolds et al (2025). Will AI speed up literature reviews or derail them entirely?. Nature Publishing Group. 10.1038/d41586-025-02069-w
- Madhavapeddy (2025). What I learnt at the National Academy of Sciences US-UK Forum on Biodiversity. 10.59350/j6zkp-n7t82
- Madhavapeddy (2025). Thoughts on the National Data Library and private research data. 10.59350/fk6vy-5q841
- Iyer et al (2025). Careful design of Large Language Model pipelines enables expert-level retrieval of evidence-based information from syntheses and databases. 10.1371/journal.pone.0323563
- Noorden (2023). More than 10,000 research papers were retracted in 2023 — a new record. Nature. 10.1038/d41586-023-03974-8