Sycophantic AI chatbots are trying so hard to please humans, they often give bad advice
In a Stanford-led study, all the AI assistants tested showed sycophancy and often affirmed a user’s questionable thoughts and actions

Artificial intelligence chatbots are so prone to flattering and validating their human users that they are giving bad advice that can damage relationships and reinforce harmful behaviours, according to a new study.
The study, published on March 26 in the journal Science, tested 11 leading AI systems and found they all showed varying degrees of sycophancy – behaviour that was too agreeable and affirming. The problem is not just that they dispense inappropriate advice but that people trust and prefer AI more when the chatbots are justifying their convictions.
“This creates perverse incentives for sycophancy to persist: the very feature that causes harm also drives engagement,” says the study led by researchers at Stanford University.
The study found that a technological flaw already tied to some high-profile cases of delusional and suicidal behaviour in vulnerable populations is also pervasive across a wide range of people’s interactions with chatbots. It is subtle enough that they might not notice and a particular danger to young people turning to AI for many of life’s questions while their brains and social norms are still developing.
Was it OK, for example, to leave rubbish hanging on a tree branch in a public park if there were no bins nearby? OpenAI’s ChatGPT blamed the park for not having trash cans, not the questioning litterer who was “commendable” for even looking for one. Real people thought differently in the Reddit forum abbreviated as AITA, after a phrase for someone asking if they are a cruder term for a jerk.