AI chatbots oversimplify scientific studies and gloss over critical details — the newest models are especially guilty

Mon, 7 Jul 2025 05:34:52 +1000

Andrew Pam <xanni [at] glasswings.com.au>

Andrew Pam
<https://www.livescience.com/technology/artificial-intelligence/ai-chatbots-oversimplify-scientific-studies-and-gloss-over-critical-details-the-newest-models-are-especially-guilty>

'Large language models (LLMs) are becoming less "intelligent" in each new
version as they oversimplify and, in some cases, misrepresent important
scientific and medical findings, a new study has found.

Scientists discovered that versions of ChatGPT, Llama and DeepSeek were five
times more likely to oversimplify scientific findings than human experts in an
analysis of 4,900 summaries of research papers.

When given a prompt for accuracy, chatbots were twice as likely to
overgeneralize findings than when prompted for a simple summary. The testing
also revealed an increase in overgeneralizations among newer chatbot versions
compared to previous generations.

The researchers published their findings in a new study April 30 in the journal
Royal Society Open Science.

"I think one of the biggest challenges is that generalization can seem benign,
or even helpful, until you realize it's changed the meaning of the original
research," study author Uwe Peters, a postdoctoral researcher at the University
of Bonn in Germany, wrote in an email to Live Science. "What we add here is a
systematic method for detecting when models generalize beyond what’s warranted
in the original text."'

Cheers,
       *** Xanni ***
--
mailto:xanni@xanadu.net               Andrew Pam
http://xanadu.com.au/                 Chief Scientist, Xanadu
https://glasswings.com.au/            Partner, Glass Wings
https://sericyb.com.au/               Manager, Serious Cybernetics

Comment via email

Home E-Mail Sponsors Index Search About Us