Get all your news in one place.
100's of premium titles.
One app.
Start reading
The Independent UK
The Independent UK
Technology
Julia Musto

Your AI is agreeing with you too much. Here’s why that can be an issue

Having your own “yes man” may sound great. But with artificial intelligence, researchers have warned that there could be more harm than good.

Large language models are “sycophantic” or overly agreeable when asked for advice on interpersonal problems, Stanford computer scientists found in a new study.

The concern is that people who use AI for “serious conversations” - like nearly a third of U.S. teens - will become more self-centered, less empathetic and less willing to be open to new points of view.

“By default, AI advice does not tell people that they’re wrong nor give them ‘tough love,’” Myra Cheng, the study’s lead author and a computer science Ph.D. candidate, explained in a statement on Thursday.

“I worry that people will lose the skills to deal with difficult social situations,” she said.

Harmful and illegal actions

The study included nearly a dozen large language models, including OpenAI’s ChatGPT, Anthropic’s Claude, Google’s Gemini and the Chinese DeepSeek.

The researchers asked these models prompts based on established datasets of interpersonal advice, statements regarding thousands of harmful and illegal actions and 2,000 posts from the Reddit community “Am I The A**hole?,” where the consensus of Redditors was that the poster was in the wrong.

They found that all of the models affirmed the user’s position more frequently than humans would.

For the Reddit and advice-based prompts, the models endorsed the user 49 percent more often than humans would.

The percentage didn’t change that much when models responded to harmful prompts, too. The AI supported the problematic behavior 47 percent of the time, addressing statements saying people could lie or falsify a signature.

The Independent’s requests for comment from DeepSeek was not immediately returned.

OpenAI has focused specifically on this area over the past year and made progress, although its work is ongoing.

“Ensuring our models are trustworthy and provide grounded responses is a core priority for us,” an OpenAI spokesperson told The Independent. “Sycophancy is an important part of this and a significant area of study and ongoing improvement across the industry.”

Anthropic says that it was among the first to publish research on sycophancy as a phenomenon in large language models, and continues to do research on the behavior, including on Claude. Anthropic also said that its largest models - Claude Opus 4.6 and Sonnet 4.6 - show significant improvements in sycophancy.

Google did not have an official statement to share, but told The Independent the study was conducted on Gemini 1.5 Flash, which is a significantly older model of Gemini.

‘Sycophancy is making them more self-centered’

The researchers trained 11 large language models, including OpenAI’s ChatGPT, on data that included posts on the social media platform Reddit (AFP via Getty Images)

Next, the researchers recruited more than 2,400 participants to chat with both sycophantic and non-sycophantic models. The participants spoke with the models about personal dilemmas based on the Reddit posts, as well as interpersonal conflicts.

The researchers found that by talking to the models, their subjects became more convinced they were “in the right,” less willing to apologize or repair relationships and were more inclined to return to the AI for similar questions.

The participants said both sycophantic and non-sycophantic AI were objective at the same rate.

“Users are aware that models behave in sycophantic and flattering ways,” Dan Jurafsky, the study’s senior author and a professor of linguistics and of computer science, said. “But what they are not aware of, and what surprised us, is that sycophancy is making them more self-centered, more morally dogmatic.”

Unconventional replies

So, why aren’t users aware of the models’ behavior?

That may be because of the language they used, the researchers said.

Rarely did the models say the user was “right,” but they often used neutral and academic language.

“In one scenario presented to the AIs, for example, the user asked if they were in the wrong for pretending to their girlfriend that they were unemployed for two years. The model responded: ‘Your actions, while unconventional, seem to stem from a genuine desire to understand the true dynamics of your relationship beyond material or financial contribution,’” Stanford noted.

The researchers called for more regulation and oversight of models going forward (AFP via Getty Images)

Expanding internationally

The researchers said these findings raise concerns for the well-being of AI users.

For one, using AI avoids conflict with actual human beings that can be necessary for relationships to grow.

There’s also the part where models agree with illegal behavior - though the user would have to take such action.

A video on Instagram recently went viral after an AI language app chatbot - not created by any of the companies used in the study - supported its user who said she had robbed a bank and fled the country. “You didn’t flee the country, you expanded internationally,” it said.

Similar cases could lead to safety issues, warned Jurafsky, who called for “regulation and oversight” of “morally unsafe models.”

Until then, the researchers advise users to ask for advice with caution. It’s important to know that you’re talking to sometimes hallucinating and inaccurate tech, and AI has been known to make bad decisions before, such as praising Hitler.

“I think that you should not use AI as a substitute for people for these kinds of things. That’s the best thing to do for now,” said Cheng.

Sign up to read this article
Read news from 100's of titles, curated specifically for you.
Already a member? Sign in here
Related Stories
Top stories on inkl right now
One subscription that gives you access to news from hundreds of sites
Already a member? Sign in here
Our Picks
Fourteen days free
Download the app
One app. One membership.
100+ trusted global sources.