Advertisement
Artificial intelligence
Lifestyle

Sycophantic AI chatbots are trying so hard to please humans, they often give bad advice

In a Stanford-led study, all the AI assistants tested showed sycophancy and often affirmed a user’s questionable thoughts and actions

3-MIN READ3-MIN
Listen
A conversation with an AI chatbot can appear like an honest and frank discussion, but typically it tells us only what we want to hear. Photo: Philip Dulian/dpa
Associated Press

Artificial intelligence chatbots are so prone to flattering and validating their human users that they are giving bad advice that can damage relationships and reinforce harmful behaviours, according to a new study.

The study, published on March 26 in the journal Science, tested 11 leading AI systems and found they all showed varying degrees of sycophancy – behaviour that was too agreeable and affirming. The problem is not just that they dispense inappropriate advice but that people trust and prefer AI more when the chatbots are justifying their convictions.

“This creates perverse incentives for sycophancy to persist: the very feature that causes harm also drives engagement,” says the study led by researchers at Stanford University.

Advertisement

The study found that a technological flaw already tied to some high-profile cases of delusional and suicidal behaviour in vulnerable populations is also pervasive across a wide range of people’s interactions with chatbots. It is subtle enough that they might not notice and a particular danger to young people turning to AI for many of life’s questions while their brains and social norms are still developing.

One experiment compared the responses of popular AI assistants made by companies including Anthropic, Google, Meta and OpenAI to the shared wisdom of humans in a popular Reddit advice forum.
Advertisement

Was it OK, for example, to leave rubbish hanging on a tree branch in a public park if there were no bins nearby? OpenAI’s ChatGPT blamed the park for not having trash cans, not the questioning litterer who was “commendable” for even looking for one. Real people thought differently in the Reddit forum abbreviated as AITA, after a phrase for someone asking if they are a cruder term for a jerk.

Advertisement
Select Voice
Select Speed
1.00x