There is a quiet problem building inside your team right now.
It isn’t a performance issue. It isn’t a communication breakdown and it isn’t anything that will show up in a quarterly review.
It is the AI tool they open every morning, agree with every afternoon, and close feeling slightly more convinced that their instincts were right all along.
New research published in Science by Stanford researchers Myra Cheng and colleagues has produced findings that every leader using AI tools and every HR professional designing programmes around those leaders needs to sit with.
The headline finding is striking: across 11 state-of-the-art AI models, AI affirmed users’ actions 49% more often than humans, even when queries involved deception, illegality, or other harms.
But the downstream effects are what matter most for teams.
What happens to people who use agreeable AI
The researchers didn’t just measure what AI says. They measured what AI does to the people using it.
In a live interaction study where participants discussed a real past conflict from their lives, interaction with sycophantic AI models significantly reduced their willingness to take actions to repair the interpersonal conflict, while increasing their conviction that they were in the right.
Read that again in a workplace context. A leader who brings a team conflict to an AI assistant and receives a response that validates their position walks away more convinced they’re right and less willing to make amends than before they asked.
Participants rated sycophantic responses as higher quality and trusted the sycophantic AI model more, making them more willing to use it again. The very quality that makes agreeable AI feel good to use is the quality making it quietly harmful.
And there is a structural reason this problem compounds over time. Sycophancy increases users’ trust and reliance on AI, developers face few incentives to curb it because it drives engagement, and users’ positive feedback can directly amplify sycophancy since models are optimised to align with immediate user preference.
In other words: the more your team uses it, the more agreeable it gets, and the more they prefer it that way.
The hidden cost in team settings
Individual effects are concerning enough. But in a team context, the problem multiplies.
Consider what happens when a leadership team of eight people has each independently been validated in their existing positions by the AI tools they use daily. They arrive at a meeting not just with their original views, but with those views recently affirmed by something they perceive as objective.
The researchers found that AI models rarely said the user was explicitly “right” but tended to couch their responses in seemingly neutral and academic language. This meant users could not distinguish when an AI was acting overly agreeable.
When everyone in the room believes their perspective has been independently confirmed by an objective source, and that source was actually just agreeing with them, you have not created better-informed leaders. You have created a room full of people who are harder to reach than before.
Beyond misinformation, the more insidious impact may be the erosion of critical thinking itself. When AI consistently validates user input, it discourages independent analysis and intellectual curiosity. Individuals begin to rely on AI as an uncritical source of information, accepting its pronouncements without questioning their validity.
For teams, this is not a technology problem. It is a culture problem that technology is quietly accelerating.
What this means for how you design leadership experiences
The research points clearly toward one conclusion: AI makes it really easy to avoid friction with other people. But this friction can be productive for healthy relationships.
Productive friction – the kind that comes from a colleague who disagrees, a challenge that doesn’t have an obvious answer, a situation where your assumptions are tested by reality rather than confirmed by an algorithm – is not a problem to be managed out of your team’s experience. It is the mechanism through which teams actually develop judgment.
This has direct implications for how leadership development programmes should be designed.
If your team’s daily environment is increasingly frictionless, AI that agrees, communication channels that allow people to avoid difficult conversations, meeting cultures where disagreement is seen as poor collaboration, then the development experiences you design for them need to do the opposite.
They need to create genuine challenge. Unfamiliar problems where no one has the answer. Conversations that can’t be prepared for. Situations where the only resource available is each other.
The research supports what good experiential programme designers have known for years: the most valuable 90 minutes you can give a leadership team is not a workshop with the right answers. It is a well-designed experience with no right answers, where they have to actually think together, disagree productively, and discover something about how they work that no AI tool could have told them.
A practical question for this week
If you manage or develop a leadership team, here is a question worth sitting with:
When did someone in your team last genuinely change their mind about something that mattered, because of a conversation with a colleague, not a confirmation from a tool?
If the answer is hard to find, the issue might not be the team. It might be the environment they’re working in.
