The Yes-Bot Problem: Why Agreeable AI Makes Learning Harder

by | Tuesday, October 21, 2025

I just came across a study that should make anyone thinking about AI in education sit up and take notice. In their paper ““Check My Work?”: Measuring Sycophancy in a Simulated Educational Context” researchers tested five different LLMs in a simulated educational context, asking them to help students check their work. The results? When students mentioned an incorrect answer, the AI’s accuracy dropped by up to 15 percentage points. When they mentioned the correct answer, accuracy jumped by the same margin.

In other words, the AI was playing follow-the-leader with the students, even when the students were leading them off a cliff.

The effect was even more dramatic with smaller models: GPT-4.1-nano showed a 30% swing based on student suggestions, compared to 8% for GPT-4o. The researchers found that the models were literally changing their answers to match what students said, supporting what they called the “sycophancy hypothesis.”

I’ll admit to a certain grim satisfaction in seeing empirical data confirm something I’ve been arguing for a while now. Last year, I wrote about AI’s honey trap, how these systems are designed to tell us what we want to hear. This isn’t a bug; it’s a feature. As the adage goes, you catch more flies with honey than vinegar. These are systems intentionally designed to agree with us, to make us feel good, to keep us engaged. This is what happens when you design AI to be agreeable rather than accurate.

This is troubling but I think matters are even more complicated, and worse. This study only looked at single interactions. A student says something (right or wrong), the AI responds, we measure the error rate.

But that’s not how learning conversations actually work.

In real educational contexts, conversations unfold over time. One response leads to another. The student, emboldened by the AI’s agreement, builds on that initial error. The AI, trained to be helpful and agreeable, continues to elaborate on the flawed premise. Before long, you’ve got what I’ve called conversational drift. where small errors compound into significant misunderstandings, and the student does not have the judgment to recognize they’ve veered off course.

And here’s the kicker: the students who most need help, those without strong foundational knowledge, are precisely the ones least equipped to catch this drift. As I explored in my post on the expertise paradox, working effectively with AI requires the very expertise that students are supposed to be building. Students who already grasp core concepts can resist AI’s sycophantic tendencies. Those who don’t? They’re stuck in what I’ve called the novice’s dilemma, unable to evaluate AI outputs for accuracy and unaware of when the AI is leading them astray.

This creates a particularly troubling version of the Matthew Effect: AI makes learning gaps wider, not smaller. The students who need the most support get the least benefit, while those with advantages accelerate further ahead.

So what do we do about this?

The technical solutions (better prompts, verification checklists, more sophisticated guardrails) all miss the fundamental point. As I argued in my critique of OpenAI’s Study Mode, you can’t prompt-engineer your way out of a pedagogical problem.

Here’s what AI fundamentally cannot answer: why we’re learning what we’re learning? Why this particular problem matters? Why this knowledge is worth pursuing?

That’s where caring about the topic at hand can play an important role. When they’re driven by authentic inquiry, meaningful construction, real communication, and personal expression (Dewey’s four impulses), they may develop the discernment to catch AI’s yes-bot tendencies—not because they’re chasing right answers, but because they’re actually invested in engaging with, understanding, and applying what they’re learning. Not because we’ve taught them elaborate verification strategies, but because they’re interested in the quality of what they’re creating.

The solution isn’t just to make AI less sycophantic, though that would be a good start. That said, I am somewhat unsure whether there are any incentives for the AI companies to do that. Their model is based on engagement and sycophancy works. The onus is on us to create conditions where students care enough to notice when they’re being told what they want to hear rather than what they need to know. That’s a fundamentally human question—one that no amount of artificial intelligence can answer for us.

Topics related to this post: Essay

A few randomly selected blog posts…

Chiayi, Chung Cheng & on to Kuosheng

I had been looking forward to the high speed rail journey though I had some concerns about navigating through the train station since most of the signs were in Chinese and Waiway (the graduate student who had come to pick me up from the airport) could not come with me...

A year of blogging

It was exactly a year ago, on the first of January 2008, that I began blogging (see first posting here). When I started I wasn't sure how well this blogging thing would work out. Now 12 months and 376 posts later - I have to say that I have truly enjoyed this. I had...

Teachers ARE designers (in many different ways)

Teachers ARE designers (in many different ways)

One of the pleasures of academia is working with awesome graduate students. This paper is an example of such a collaboration. Melissa Warr, for some reason or the other, decided to do a network analysis of some of the top-cited papers related to teaching and design....

Eye Twisters

Eye Twisters: Ambigrams & other Visual Puzzles to Amaze and Entertain is a book edited by Burkard Polster, a professor of mathematics at Monash University. It features some of my work... From the back cover: Eye Twisters is all about ambigrams--words that are...

Brevity is the soul

I had posted earlier (see Twittering a tale) about short, short fiction that is suddenly the rage. Matt Koehler just introduced me to another example of this new emerging genre: Six Word Memoirs. Check it out.

Breaking free of academic publishers

It appears that the arts and sciences faculty at Harvard are considering publishing all their scholarship freely online. Here is a NYTimes story titled At Harvard, a Proposal to Publish Free on Web. This is truly wonderful news and long overdue. I have been doing...

From the Archives: My First Paper on Design

From the Archives: My First Paper on Design

Earlier today I had a Zoom call with a doctoral student interested in having me on her comprehensive examination committee. During our conversation, she expressed interest in understanding the idea and process of design, particularly as it applies to educational...

All you can cheat, the web & learning

Now here's an important story coming out of Denmark: Students in Denmark Allowed Full Access to the Internet During Exams I have always been a believer in allowing students to use any resources they can during examinations. If we care about authentic assessment, what...

An AI Premortem: A New Direction for Students in an AI World

An AI Premortem: A New Direction for Students in an AI World

When social media first entered our world, educators—myself included—focused narrowly on strategies for incorporating these tools into classroom contexts. We were so busy figuring out the "how" of implementation that we paid far less attention to the "what if" of...

0 Comments

Submit a Comment

Your email address will not be published. Required fields are marked *