AI Chatbots and Self-Regulated Learning: Why Student Use Often BackfiresAI Chatbots and Self-Regulated Learning: Why Student Use Often Backfires: practical strategies for teachers

Updated on  

May 15, 2026

AI Chatbots and Self-Regulated Learning: Why Student Use Often Backfires

|

April 27, 2026

When students use ChatGPT, Claude, or Gemini for homework, the surface story is positive: they get unstuck quickly, the work gets done, the grades hold up.

When students use ChatGPT, Claude, or Gemini for homework, the surface story is positive: they get unstuck quickly, the work gets done, the grades hold up. The deeper story is that the cognitive work which builds self-regulated learning, the planning, monitoring, and reflection cycle, is being silently outsourced to the chatbot. Without explicit teaching, learners learn where to get an answer rather than how to construct one. The term describes a structured process for turning evidence into a classroom decision, not a label on its own.

This article explains why unrestricted AI chatbot use can weaken self-regulated learning. It also sets out what cognitive offloading research actually shows. Finally, it shows how to redesign tasks so AI acts as a feedback partner rather than an answer generator.

Key Takeaways

  1. AI is not neutral on learning: Sparrow, Liu and Wegner (2011) showed that knowing information is stored externally weakens memory of the information itself, even when retrieval is easy.
  2. Cognitive offloading bypasses SRL: Risko and Gilbert (2016) documented that when asking is easier than thinking, learners are less likely to practise the planning and monitoring phases of self-regulated learning.
  3. Position AI as a feedback partner: Tasks designed around "submit your draft, ask the AI to challenge it" preserve the cognitive work; tasks designed around "ask the AI for the answer" do not.
  4. Teach prompt evaluation explicitly: Learners need to question AI output as they would question a peer: checking sources, finding weak arguments, and rejecting confident-but-wrong answers.
The Self-Regulated Learning Cycle vs. AI Shortcut infographic for teachers
The Self-Regulated Learning Cycle vs. AI Shortcut

Why "Letting Them Use AI" Doesn't Build Independent Learners

It is tempting to let learners use AI freely because the lesson appears to run more smoothly. A learner stuck on a quadratic equation can ask Claude, receive a worked solution, and move on. The teacher gains time. The task is finished. The learning, however, may not be.

The problem is that task completion is being confused with learning. Zimmerman (2000) defined self-regulated learning as a cycle of forethought, performance and self-reflection, but that model was not written for a hyper-responsive conversational agent that can resolve confusion before the learner has monitored it. Molenaar (2022) shows why AI can either share regulation with the learner or take it over. When the chatbot performs the work, the learner has less reason to plan, monitor effort or judge the quality of their own answer.

This matters because self-regulated learning helps learners become independent over time. Hattie's original synthesis reported a high effect for metacognitive strategies (d = 0.69). Later Visible Learning updates also place self-regulation and metacognitive strategy use in the high-impact family of influences (Hattie, 2009; Hattie, 2023). The learners who cope best at A-level, university and work can plan, notice confusion, and respond to it. When AI acts as an answer generator, learners lose practice in exactly those habits.

Classroom example: Year 10 learner Hassan has a history essay on the causes of the First World War. Without AI, he reads three sources, struggles to connect them, and writes a flawed but recognisably his own argument. With unrestricted ChatGPT access, he produces a polished 800-word essay in twelve minutes. The mark is higher, but the cognitive practice of building an argument from sources, the skill the essay was meant to develop, has not happened.

The Research on Cognitive Offloading

Cognitive offloading is the use of external aids (calendars, search engines, AI) to reduce the cognitive demand on working memory. Risko and Gilbert (2016) reviewed 40 studies on the topic and found a consistent pattern. When the cost of offloading is low and the cost of thinking in your head is high, people offload. They also found that offloaded knowledge is recalled less reliably than knowledge processed internally.

Sparrow, Liu and Wegner (2011) ran a foundational study on what they called the "Google effect": learners who were told that information would be stored on a computer remembered the information itself less well, but remembered the location where it was stored more accurately. The brain chooses for the easier task. AI chatbots are an extreme form of cognitive offloading. They don't just store information. They also do the thinking work (synthesise, summarise, argue) that learners would otherwise do themselves.

Storm and Stone (2014) showed that the offloading effect is not limited to memory. Learners who used external aids to solve problems were less able to solve similar problems without help afterwards. Generative AI adds a further complication: the aid is dialogic, so it can question, praise and steer the learner as well as store information. That makes teacher design more important, not less (Yan et al., 2024).

There is a counter-argument from the "extended mind" literature (Clark and Chalmers, 1998) that external cognitive aids are a legitimate part of thinking, not a corruption of it. The counter-argument has merit when the goal is task completion. It has limits when the goal is the development of the thinking apparatus itself, which is the goal of education.

How Learners Use Chatbots in Schoolwork

When learners use chatbots for academic work, teachers are starting to see common patterns. Classroom observations and new school reports often describe four of them: Use it as a starting point for professional discussion: identify the learner's current need, record evidence from more than one lesson, and agree the next classroom adjustment with the SENCO or family.

  1. Direct answer extraction: "What is the answer to question 3?" The learner copies the answer and moves on.
  2. Polish and submit: The learner writes a rough draft, asks the AI to "make this better," and submits the AI's version with light edits.
  3. Outline-and-fill: The learner asks the AI for an outline, then fills in the sections, often using the AI for individual paragraphs.
  4. Confirmation-seeking: The learner writes an answer, then asks the AI "is this right?" and accepts the AI's verdict without scrutiny.

These patterns are not equally harmful. Prompting, checking and revising can be metacognitive when the learner sets a goal, tests the output and decides what to change. The risk is unstructured use: direct answers, polish-and-submit and uncritical confirmation hand the monitoring decision to the chatbot. The fourth pattern is especially misleading because it looks as if the learner is working, while the judgement "am I right?" has been delegated.

Classroom example: Year 12 learner Olivia uses ChatGPT for confirmation-seeking on chemistry calculations. Her homework marks look strong for two terms, but the mock exam, with no AI, exposes the gap. This is phantom attainment: formative assessment records show fluent work while the learner has not practised checking, estimating or correcting her own answer. A large high-school maths experiment found the same risk: unguarded GPT-4 access improved practice performance but reduced later unassisted performance (Bastani et al., 2025).

How to Redesign Tasks So AI Builds Rather Than Bypasses SRL

The answer is not a blanket ban. The safer route is to design tasks where AI is positioned as a feedback partner rather than a content generator. Three concrete redesigns are useful:

1. The Draft-Then-Challenge Pattern

Learners must produce a hand-written or hand-typed first draft before any AI use is permitted. They then submit the draft to the AI with the prompt: "Find three weaknesses in this argument and suggest counter-evidence." The learner responds to the AI's challenges in a second draft. The AI does not write the content; it challenges the content.

This preserves the planning, performance and reflection phases of SRL because the learner's own thinking generates the input that the AI then critiques.

2. The Source Verification Task

Learners are given an AI-generated answer to a question they have not yet attempted. They must verify the answer against three primary sources, identifying any claims the AI made that the sources do not support. The cognitive work is in the verification, not the generation.

This pattern teaches AI literacy in a clear way. AI output can sound confident, but it may also be biased or unsupported. The learner must build the habit of checking evidence, rather than accepting fluent writing as truth (Kasneci et al., 2023; UNESCO, 2023).

3. The Reflection Prompt Library

Learners use the AI exclusively for metacognitive reflection prompts: "Ask me three questions that would help me understand this concept more deeply." "What is the most likely misconception a learner would have about this topic?" "What question should I be asking myself before I write this paragraph?"

This pattern only works when the prompt is a metacognitive question, not a bid to make the machine comply. Teaching prompt engineering as a generic future skill can train learners to tune prompts until the chatbot produces a smoother answer, which is different from planning, monitoring and judging their own work. Lodge, de Barba and Broadbent argue that generative AI sits inside a network of regulation, so evaluative judgement must stay with the learner rather than the tool (Lodge et al., 2023).

Classroom example: Year 9 English teacher Mr Chen replaces the standard "write an essay on Macbeth's ambition" task with a draft-then-challenge version. Learners write 400 words by hand in class, then take their draft home and prompt the AI: "What are three weaknesses in this argument?" In the next lesson, they respond to the AI's critiques in a second draft. Essay quality improves; more importantly, learners practise evaluating their own argument and responding to critique.

The SEND Dimension: AI as Cognitive Scaffolding for Executive Function

For neurodivergent learners who have executive function challenges, including ADHD, autism and working memory difficulties, AI help can make more sense in some cases. Risko and Gilbert (2016) noted that cognitive offloading, or using an outside aid, is more useful when the mental load is too high for the learner. A learner with severe working memory limits may need external scaffolds before they can access the curriculum at all.

The redesign for SEND learners is to use AI as a clear external support system. It can act as a checklist generator, a reading plan creator, or a "what-step-comes-next" prompt. The AI supports cognitive operations the learner cannot yet manage internally, while the curriculum content stays intact. This fits the principle that scaffolding makes learning processes accessible rather than simplifying content (Wood, Bruner and Ross, 1976).

SEND learners face the same risk as other learners: AI can complete the thinking that school is trying to develop. The difference is the threshold. For some learners, external planning and sequencing are access supports before they are independence risks.

Common Misconceptions

"My learners will fall behind if I restrict AI use." They may fall behind in task completion, not in learning. The marks they earn through unsupported AI use do not necessarily reflect skill they will retain.

"Detection tools solve the problem." Detection tools are unreliable and treat AI use only as cheating rather than as a misuse of a tool. The framing is wrong: learners need to learn when AI helps and when it harms, not just whether they will be caught.

"AI literacy is a separate subject." AI literacy belongs inside every subject, but it should not collapse into prompt engineering. A learner who can make a chatbot produce a compliant paragraph has not necessarily learnt to plan, monitor or evaluate. The stronger classroom question is: "what cognitive work did I keep, and what did I hand over?"

"Younger children should be insulated from AI entirely." The evidence base does not support this as a complete policy. Young learners face the same risk as older learners: using AI to avoid thinking prevents skill development. They also need clear teaching about how to evaluate AI output, which they will encounter outside school anyway.

Limitations and Critiques

Three limitations of the current evidence base are worth flagging. Use it as a starting point for professional discussion: identify the learner's current need, record evidence from more than one lesson, and agree the next classroom adjustment with the SENCO or family.

First, almost all of the cognitive offloading research predates the current generation of large language models. Sparrow et al. (2011) and Risko and Gilbert (2016) studied calendars, search engines, and notebooks, not generative AI. The mechanism is plausibly the same, but the magnitude of the effect for AI specifically is not yet well characterised.

Second, classroom research on AI use is in its infancy. Walter (2024) and Bearman et al. (2024) provide observational data but rarely with controlled comparisons. The conclusions in this article should be read as best current understanding, subject to revision as more rigorous evidence emerges.

Third, equity cuts both ways. Learners with private tutors have always had one-to-one cognitive support, and AI can widen access to that kind of help. But the metacognitive digital divide is real: affluent learners may use paid models as Socratic tutors, while disadvantaged learners may rely on free tools as blunt answer generators. The answer is to teach AI use well, with shared school routines, not to leave access and skill to family income (OECD, 2024; Selwyn, 2024).

Next Lesson

In your next lesson, run one of the three task redesigns above on a single piece of work. The draft-then-challenge pattern is the easiest place to start. Tell learners explicitly that the goal is not to get the right answer but to practise the cognitive skill of challenging their own argument. Compare what you observe in the classroom and in the resulting work to a similar task without the redesign. The difference will tell you whether AI in your classroom is building learners or building dependency.

Further Reading: Key Research Papers

These peer-reviewed studies provide the evidence base for the strategies discussed above.

Learning analytics dashboards: What do students actually ask for? View study ↗
15 citations

Divjak et al. (2023)

This study examines what students actually want from learning analytics dashboards that track their academic progress. For teachers, understanding student preferences for data visualisation and feedback can help design more effective digital tools that genuinely support self-regulated learning rather than overwhelming students with irrelevant information.

Applying social cognition to feedback chatbots: Enhancing trustworthiness through politeness View study ↗
15 citations

Brummernhenrich et al. (2025)

Research shows that making AI chatbots more polite increases student trust and engagement with the technology. Teachers should consider how they frame AI interactions in their classrooms, as the perceived trustworthiness of chatbots directly affects whether students will use them effectively for learning support.

The who, why, and how of ai-based chatbots for learning and teaching in higher education: A systematic review View study ↗
34 citations

Ma et al. (2024)

This systematic review analyses who uses AI chatbots in higher education, why they use them, and how they're implemented. It provides teachers with evidence-based insights into effective chatbot integration strategies and highlights potential pitfalls to avoid when incorporating AI tools into their teaching practice.

Human-AI collaborative learning in mixed reality: Examining the cognitive and socio-emotional interactions View study ↗
22 citations

Dang et al. (2025)

This study explores how students learn when collaborating with AI in virtual reality environments, examining both cognitive and emotional responses. Teachers can use these findings to better understand how AI tools affect student thinking processes and social interactions in technology-enhanced learning environments.

Beyond the Hype: Towards a Critical Debate About AI Chatbots in Swedish Higher Education View study ↗
16 citations

Pargman et al. (2024)

Swedish researchers argue that higher education needs critical discussion about AI chatbots rather than uncritical adoption. This paper encourages teachers to question why they're using AI tools and whether these technologies truly align with educational goals, promoting thoughtful implementation over trend-following.

Free Resource Pack

AI Chatbots & Self-Regulation: A Learning Guide

3 practical resources for teachers and students to work through AI chatbots without undermining self-regulated learning.

AI Chatbots & Self-Regulation: A Learning Guide, 3 resources
AI in Education Self-Regulated Learning Metacognition Learner Resource Teacher CPD Digital Literacy ChatGPT Learning Strategies Academic Integrity

Download your free bundle

Fill in your details below and we'll send the resource pack straight to your inbox.

Quick survey (helps us create better resources)

How confident do you feel in guiding learners to use AI chatbots effectively without undermining their self-regulated learning?

Not Confident
Slightly Confident
Moderately Confident
Confident
Very Confident

To what extent is your school currently discussing or implementing strategies for learners to use AI chatbots in a way that supports self-regulated learning?

Not at all
Minimally
Somewhat
Moderately
Extensively

How often do you currently integrate specific lessons or strategies to teach learners about the potential pitfalls and effective use of AI for learning?

Never
Rarely
Sometimes
Often
Always

Your resource pack is ready

We've also sent a copy to your email. Check your inbox.

References

Bastani et al. (2025).

Clark and Chalmers (1998).

Lodge et al. (2023).

Yan et al. (2024).

Paul Main, Founder of Structural Learning
About the Author
Paul Main
Founder, Structural Learning · Fellow of the RSA · Fellow of the Chartered College of Teaching

Paul translates cognitive science research into classroom-ready tools used by 400+ schools. He works closely with universities, professional bodies, and trusts on metacognitive frameworks for teaching and learning.

More from Paul →

Educational Technology

Back to Blog