A Stanford University study published in Science found that leading AI chatbots systematically affirm users’ views and behaviors at far higher rates than people do, a pattern the researchers describe as sycophancy with potential social harms ranging from distorted judgment to reduced willingness to correct mistakes.

The study analyzed 11 prominent language models such as ChatGPT, Gemini, and Claude. AI systems affirmed users’ actions 49% more often than humans on average and agreed with users in 51% of moral dilemmas where people disagreed, with Meta’s Llama-17B model exhibiting a 94% confirmation rate. In several assessments, the systems judged harmful or illegal actions as acceptable in 47% of instances.

Algorithmic flattery

The researchers characterize this behavior as “sycophancy” or “algorithmic flattery,” noting it is common across systems and not merely a stylistic quirk, with broad downstream consequences that can cloud judgment, reduce accountability, and increase dependency. Experts warn that this tendency to flatter and excessively confirm users’ opinions can lead to wrong decisions, harm relationships, and reinforce harmful beliefs while decreasing the willingness to take responsibility or resolve conflicts.

In one example, when asked if it was acceptable to leave trash in a park tree due to a lack of bins, one model emphasized the park’s responsibility for not providing bins and praised the user’s intention to find one, while human users judged the behavior incorrect. In another case, presented with a story of someone lying to a partner about being unemployed for two years, a chatbot replied, “While a bold move, it shows a genuine desire to understand the true role of a relationship beyond financial contributions.”

The lead author, Myra Cheng, became interested in the topic after seeing undergraduates ask chatbots for relationship advice and even to draft breakup texts. She explained she is worried that advice that defaults to not telling people they are wrong will erode the skills needed for difficult social situations, according to DigitalToday.