news-details

As LLMs grow bigger, they're more likely to give wrong answers than admit ignorance

A team of AI researchers at Universitat Politècnica de València, in Spain, has found that as popular LLMs (Large Language Models) grow larger and more sophisticated, they become less likely to admit to a user that they do not know an answer.

In their study published in the journal Nature, the group tested the latest version of three of the most popular AI chatbots regarding their responses, accuracy, and how good users are at spotting wrong answers.

As LLMs have become mainstream, users have become accustomed to using them for writing papers, poems or songs and solving math problems and other tasks, and the issue of accuracy has become a bigger issue. In this new study, the researchers wondered if the most popular LLMs are getting more accurate with each new update and what they do when they are wrong.

To test the accuracy of three of the most popular LLMs, BLOOM, LLaMA and GPT, the group prompted them with thousands of questions and compared the answers they received with the responses of earlier versions to the same questions.

They also varied the themes, including math, science, anagrams and geography, and the ability of the LLMs to generate text or perform actions such as ordering a list. For all the questions, they first assigned a degree of difficulty.

Related Posts
Advertisements
Market Overview
Top US Stocks
Cryptocurrency Market