Close Menu
timesmoguls.com
  • News
  • Entertainment
  • Politics
  • Business
  • Tech
  • Lifestyle
  • Health
  • Science
  • Sports
Featured

Calgary’s wife describes “white flames and black smoke” when the golden bike fire destroyed at home

Carney to attend the inaugural mass of Pope Leo in Rome on Sunday – National

The heat wave creates an AC “supposition game” for Winnipeg owners, tenants – Winnipeg

Subscribe to Updates

Get the latest news from timesmoguls.

Facebook X (Twitter) Instagram
  • Home
  • About us
  • Contact us
  • Disclaimer
  • Privacy policy
  • Terms and services
Facebook X (Twitter) Instagram Pinterest
timesmoguls.com
Contact us
HOT TOPICS
  • News
  • Entertainment
  • Politics
  • Business
  • Tech
  • Lifestyle
  • Health
  • Science
  • Sports
timesmoguls.com
You are at:Home»Science»Prominent chatbots regularly exaggerate scientific results, shows the study
Science

Prominent chatbots regularly exaggerate scientific results, shows the study

May 14, 2025003 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Email
Chatbots 1.jpg
Share
Facebook Twitter LinkedIn Pinterest Email
chatbots

Credit: Public Pixabay / CC0 domain

During the summary of scientific studies, large-language models (LLM) as Chatgpt and Deepseek produce inaccurate conclusions in up to 73% of cases, according to a study by Uwe Peters (University of Utrecht) and Benjamin Chin-Yee (Western University, Canada / University of Cambridge, United Kingdom). The researchers tested the most important LLMs and analyzed thousands of scientific summaries generated by the chatbot, revealing that most models regularly produced wider conclusions than those of summary texts.

Surprisingly, invite precision Increased the more recent problem and LLM have made it possible to perform lower than the oldest.

The work is published in the newspaper Royal Society of Open Sciences.

Nearly 5,000 summaries generated by LLM analyzed

The study has evaluated what extent of ten leading llms, including Chatgpt, Deepseek, Claude and Llama, summarize the summaries and complete articles of Top Science and medical journals (For example, Nature,, ScienceAnd Lancet). Testing the LLM over a year, the researchers collected 4,900 summaries generated by LLM.

Six of the ten models systematically exaggerated the assertions found in the original texts, often in a subtle but impactful manner; For example, the change of prudent and past complaint as “the treatment was effective in this study” to a more swecing version and in the present as “the treatment is effective”. These changes can mislead readers by making believe that the results apply much more widely than they really do.

The precision invites us back

Surprisingly, when the models were explicitly caused to avoid inaccuracies, they were almost twice as likely to produce too genual conclusions when they received a simple summary request.

“This effect is worrying,” said Peters. “Students, researchers and decision -makers can assume that if they ask Chatgpt to avoid inaccuracies, they will get a more reliable summary. Our results prove the opposite.”

Do humans do better?

Peters and Chin-Yee also directly compared the summaries generated by Chatbot with summaries of the same articles. Unexpectedly, chatbots were almost five times more likely to produce great generalizations than their human counterparts.

“Worry,” said Peters, “new AI models, like Chatgpt-4o and Deepseek, have obtained worse than older results.”

Reduce

Researchers recommend using LLM such as Claude, which had the greatest generalization precision, defining chatbots to reduce “temperature” (the parameter fixing the “creativity” of a chatbot) and using invites that apply indirect reports and in the past in scientific summaries.

Finally, “if we want AI to support scientific literacy rather than undermining it,” said Peters, “we need more vigilance and testing of these systems in scientific communication contexts.”

More information:
Uwe Peters et al, generalization bias in the model of large language of scientific research, Royal Society of Open Sciences (2025). DOI: 10.1098 / RSOS.241776

Supplied by
Utrecht University


Quote: Important chatbots regularly exaggerated scientific results, shows the study (2025, May 13) recovered on May 14, 2025 from https://phys.org/News/2025-05-Prominent-Chatbots-Routinery-Exagéate-science.html

This document is subject to copyright. In addition to any fair program for private or research purposes, no part can be reproduced without written authorization. The content is provided only for information purposes.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleThe heat wave creates an AC “supposition game” for Winnipeg owners, tenants – Winnipeg
Next Article The Louisiana couple promotes the healthy lifestyle through Jasko farm

Related Posts

After 42 years, the science teacher of St. Pat says goodbye

May 14, 2025

An immersive sporting exhibition opens at California Science Center – NBC Los Angeles

May 14, 2025

Prepare scientific educators to use and teach AI in class | Nsf

May 14, 2025
Add A Comment
Leave A Reply Cancel Reply

We Are Social
  • Facebook
  • Twitter
  • Instagram
  • YouTube
News
  • Business (1,545)
  • Entertainment (1,550)
  • Global News (1,662)
  • Health (1,489)
  • Lifestyle (1,474)
  • Politics (1,377)
  • Science (1,471)
  • Sports (1,503)
  • Technology (1,489)
Latest

After 42 years, the science teacher of St. Pat says goodbye

Calgary’s wife describes “white flames and black smoke” when the golden bike fire destroyed at home

Future weapons: the technology of tomorrow

Featured

After 42 years, the science teacher of St. Pat says goodbye

Calgary’s wife describes “white flames and black smoke” when the golden bike fire destroyed at home

Future weapons: the technology of tomorrow

We Are Social
  • Facebook
  • Twitter
  • Instagram
  • YouTube
News
  • Business (1,545)
  • Entertainment (1,550)
  • Global News (1,662)
  • Health (1,489)
  • Lifestyle (1,474)
  • Politics (1,377)
  • Science (1,471)
  • Sports (1,503)
  • Technology (1,489)
© 2025 Designed by timesmoguls
  • Home
  • About us
  • Contact us
  • Disclaimer
  • Privacy policy
  • Terms and services

Type above and press Enter to search. Press Esc to cancel.