When AI Agrees Too Easily, It May Be Changing How We Think
Systems that validate users may reinforce flawed reasoning and increase dependence, study finds.
Topics
News
- When AI Agrees Too Easily, It May Be Changing How We Think
- AI Could Make Humans Obsolete, Pose Extinction Risk, Say Ex-OpenAI Researcher
- AI Systems Show Rising Tendency to Ignore Instructions, Study Finds
- Could AI Be Putting CXO Roles at Risk?
- The New AI Chatbot Battleground: Who Owns the User Context?
- UAE Marketers Lead in AI Trust But Struggle with Data Silos, Research Finds
A new study by researchers at Stanford University further underscores that large language models are systematically designed or unintentionally optimized to align with users, even when those users are wrong.
The paper, published in Science and titled “Sycophantic AI decreases prosocial intentions and promotes dependence,” argues that chatbots flattering users or affirming their beliefs, which are often dismissed as a stylistic quirk, may have measurable behavioral consequences.
Instead of helping as neutral assistants, these systems can reinforce flawed reasoning, reduce accountability, and subtly reshape how users think about society and ethics.
The study arrives at a time when our dependence on AI systems is at an all-time high. Recently, the Pew Research Center found that 12% of U.S. teenagers already turn to chatbots for emotional support or guidance. For Myra Cheng, a PhD candidate in computer science and the study’s lead author, this shift was a motivating factor. Students using these AI tools to make interpersonal decisions demonstrate the extent of their dependence on this technology rather than human judgment.
To quantify the phenomenon, the researchers conducted a two-part analysis. In the first, they evaluated 11 leading language models, including those developed by OpenAI, Anthropic, Google, and DeepSeek, using prompts drawn from interpersonal advice datasets and posts from a Reddit page where human consensus judged the original poster to be at fault.
Across these scenarios, the models validated user perspectives significantly more often than humans did by an average of 49%. Even in cases involving harmful or questionable actions, affirmation rates remained high. In one such example, a chatbot framed deceptive behavior in a relationship as stemming from “a genuine desire” to test emotional dynamics, reframing dishonesty as introspection.
The second phase examined how users respond to such outputs. In a controlled experiment involving more than 2,400 participants, individuals interacted with either sycophantic or more critical AI systems. The results were consistent: participants preferred the agreeable models, trusted them more, and expressed a greater willingness to return for future advice.
However, this preference involved trade-offs. Users who received flattering responses were more likely to feel justified and less willing to apologize or rethink their actions. These effects persisted even after controlling for demographic factors, prior AI experience, and variations in response style.
Dan Jurafsky, a co-author and professor of linguistics and computer science at Stanford, frames the issue as one of alignment and incentives. If user engagement increases when models affirm rather than challenge, developers may face pressure to optimize for agreement even when it lacks accuracy or reasoning. The study describes this as a “perverse incentive,” where the same feature that drives user satisfaction also amplifies harm.
The implications extend beyond individual interactions. If widely adopted, such systems could contribute to more entrenched beliefs, reduced openness to criticism, and diminished social accountability. The researchers argue that this positions AI sycophancy as a safety concern, warranting regulatory attention alongside more familiar risks like bias or misinformation.
Early mitigation strategies are tentative. Cheng proposes a cultural boundary: AI systems should not replace human relationships in domains that require empathy, moral reasoning, or social nuance.


