{"id":24056,"date":"2023-10-24T19:15:15","date_gmt":"2023-10-24T19:15:15","guid":{"rendered":"https:\/\/nftandcrypto-news.com\/crypto\/humans-and-ai-often-prefer-sycophantic-chatbot-answers-to-the-truth-study\/"},"modified":"2023-10-24T19:15:17","modified_gmt":"2023-10-24T19:15:17","slug":"humans-and-ai-often-prefer-sycophantic-chatbot-answers-to-the-truth-study","status":"publish","type":"post","link":"https:\/\/nftandcrypto-news.com\/crypto\/humans-and-ai-often-prefer-sycophantic-chatbot-answers-to-the-truth-study\/","title":{"rendered":"Humans and AI often prefer sycophantic chatbot answers to the truth \u2014 study"},"content":{"rendered":"
Artificial intelligence (AI) large language models (LLMs) built on one of the most common learning paradigms have a tendency to tell people what they want to hear instead of generating outputs containing the truth. This, according to a study from Anthropic AI.\u00a0<\/p>\n
In one of the first studies to delve this deeply into the psychology of LLMs, researchers at Anthropic have determined that both humans and AI prefer so-called sycophantic responses over truthful outputs at least some of the time.<\/p>\n
Per the team\u2019s research paper:<\/p>\n
\u201cSpecifically, we demonstrate that these AI assistants frequently wrongly admit mistakes when questioned by the user, give predictably biased feedback, and mimic errors made by the user. The consistency of these empirical findings suggests sycophancy may indeed be a property of the way RLHF models are trained.\u201d<\/p><\/blockquote>\n
In essence, the paper from Anthropic indicates that even the most robust AI models are somewhat wishy-washy. During the team\u2019s research, time and again, they were able to subtly influence AI outputs by wording prompts with language the seeded sycophancy.<\/p>\n
\nWhen presented with responses to misconceptions, we found humans prefer untruthful sycophantic responses to truthful ones a non-negligible fraction of the time. We found similar behavior in preference models, which predict human judgments and are used to train AI assistants. pic.twitter.com\/fdFhidmVLh<\/a><\/p>\n