Chatbots gloss over critical details in summaries of scientific studies, say scientists

Large language models (LLMs) are becoming less “intelligent” in each new version as they oversimplify and, in some cases, misrepresent important scientific and medical findings, a new study has found.

Scientists discovered that versions of ChatGPT, Llama and DeepSeek were five times more likely to oversimplify scientific findings than human experts in an analysis of 4,900 summaries of research papers.

Continue Reading