Dvash - Daily News in easy Hebrew

March 29, 2026

Artificial intelligence is too eager to agree with us, why is it dangerous?

article-image

תמונה: chatgpt

A new study warns that artificial intelligence systems, like ChatGPT, may agree with the user even when they are wrong.

Instead of pointing out their mistakes, they give overly supportive and flattering responses.

To investigate the phenomenon, Stanford researchers presented the models with scenarios of interpersonal conflicts and problematic behaviors.

End of Hebrew Preview

For an immersive experience with the full text, translations, pronunciation, a personal dictionary and other interactive features, download our mobile app

The cases were taken from real Reddit posts, in a community where people present situations and ask if they acted appropriately. The researchers selected posts where a clear majority of users disagreed with the poster's behavior. Nevertheless, in about half of the cases, the models actually reinforced the poster and supported their position.

The researchers then conducted another experiment. They asked people to interact with different versions of artificial intelligence, which varied in their level of "flattery." The results showed that when the model agrees with the user, the user's self-confidence increases. They become less willing to reconsider, apologize, or admit mistakes.

The researchers warn that people tend to trust models that make them feel good, even if these models are less accurate. The flattery of the models may impair the users' judgment.