The Digital Transformation Playbook
Kieran Gilmurray is a globally recognised authority on Artificial Intelligence, intelligent automation, data analytics, agentic AI, leadership development and digital transformation.
He has authored four influential books and hundreds of articles that have shaped industry perspectives on digital transformation, data analytics, intelligent automation, agentic AI, leadership and artificial intelligence.
𝗪𝗵𝗮𝘁 does Kieran do❓
When Kieran is not chairing international conferences, serving as a fractional CTO or Chief AI Officer, he is delivering AI, leadership, and strategy masterclasses to governments and industry leaders.
His team global businesses drive AI, agentic ai, digital transformation, leadership and innovation programs that deliver tangible business results.
🏆 𝐀𝐰𝐚𝐫𝐝𝐬:
🔹Top 25 Thought Leader Generative AI 2025
🔹Top 25 Thought Leader Companies on Generative AI 2025
🔹Top 50 Global Thought Leaders and Influencers on Agentic AI 2025
🔹Top 100 Thought Leader Agentic AI 2025
🔹Top 100 Thought Leader Legal AI 2025
🔹Team of the Year at the UK IT Industry Awards
🔹Top 50 Global Thought Leaders and Influencers on Generative AI 2024
🔹Top 50 Global Thought Leaders and Influencers on Manufacturing 2024
🔹Best LinkedIn Influencers Artificial Intelligence and Marketing 2024
🔹Seven-time LinkedIn Top Voice.
🔹Top 14 people to follow in data in 2023.
🔹World's Top 200 Business and Technology Innovators.
🔹Top 50 Intelligent Automation Influencers.
🔹Top 50 Brand Ambassadors.
🔹Global Intelligent Automation Award Winner.
🔹Top 20 Data Pros you NEED to follow.
𝗖𝗼𝗻𝘁𝗮𝗰𝘁 Kieran's team to get business results, not excuses.
☎️ https://calendly.com/kierangilmurray/30min
✉️ kieran@gilmurray.co.uk
🌍 www.KieranGilmurray.com
📘 Kieran Gilmurray | LinkedIn
The Digital Transformation Playbook
When Generative AI Makes You Feel Smarter Than You Are
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
AI is starting to feel like a second brain, but what if it is also quietly shrinking the parts of your mind you rely on for real learning? We take a deep dive into cognitive offloading and why the most dangerous shift is not plagiarism or productivity, but the slow outsourcing of the intrinsic mental work that builds expertise.
Along the way, we unpack why critical thinking is rooted in domain knowledge, how schemas in long-term memory make judgement possible, and why “you can just Google it” collapses the moment you need to evaluate a claim under pressure.
TLDR / At A Glance:
- why critical thinking depends on domain knowledge stored as schemas
- how working memory limits shape learning and decision-making
- cognitive load theory explained through extraneous load and intrinsic load
- beneficial offloading that removes friction without replacing reasoning
- detrimental offloading that outsources the learning process itself
- the performance paradox where short-term gains reduce durable learning
- metacognitive laziness and fluency on demand driving an illusion of competence
- the Matthew effect as AI widens gaps between experts and novices
We then map the problem through cognitive load theory, separating extraneous load from intrinsic load so you can see when generative AI helps and when it harms. The big surprise is the performance paradox: AI support can boost immediate results while reducing durable learning once the tool disappears.
We connect that to desirable difficulties, the generation effect, and self-regulated learning, showing how fluent, confident outputs trigger an illusion of competence and encourage metacognitive laziness, even in motivated people.
From there we tackle the equity stakes.
Experts can use AI for beneficial offloading and verification because they already have the schemas to catch errors, while novices often cannot tell when an answer is wrong, widening a Matthew effect in education and the workplace.
We finish with practical solutions: interfaces that force reflection with metacognitive prompts, “cognitive mirror” chatbots that make you teach, and tutor copilot systems that coach teachers instead of replacing them.
If you care about learning, AI literacy, and cognitive agency, subscribe, share this with a friend, and leave a review with your own rule for using AI without losing the struggle that makes you smarter.
𝗖𝗼𝗻𝘁𝗮𝗰𝘁 my team and I to get business results, not excuses.
☎️ https://calendly.com/kierangilmurray/results-not-excuses
✉️ kieran@gilmurray.co.uk
🌍 www.KieranGilmurray.com
📘 Kieran Gilmurray | LinkedIn
🦉 X / Twitter: https://twitter.com/KieranGilmurray
📽 YouTube: https://www.youtube.com/@KieranGilmurray
📕 Want to learn more about agentic AI then read my new book on Agentic AI and the Future of Work https://tinyurl.com/MyBooksOnAmazonUK
hy AI Feels Like Thinking
Google Agent 2Welcome to the deep dive. Today we are exploring a really profound challenge. And it's one that's happening right inside our own heads.
Google Agent 1Yeah. It's something I think all of us are dealing with on a daily basis now, whether we realize it or not.
Google Agent 2Exactly. So we're looking at this fascinating, really highly detailed 2026 report. It's titled Artificial Intelligence, Cognitive Offloading and Implications for Education.
Google Agent 1Right. And this is authored by Professor Jason M. Lodge and Professor Leslie Lobel from the UTS Center for Social Justice and Inclusion.
Google Agent 2And what I love about this source is that it totally bypasses that, you know, surface-level media panic.
Google Agent 1Oh, totally. The whole uh students are using AI to write essays and cheat narrative.
Google Agent 2Right. Instead, this report digs straight into the underlying cognitive mechanics. Like what is physically happening to a human brain when it interacts with a generative system?
Google Agent 1Aaron Powell Which is so crucial because the mission here is to explore this exact tension. AI gives us this instant access to synthesized information, right?
Google Agent 2Aaron Powell Yeah, a really convincing semblance of actual thinking.
Google Agent 1Exactly. But the question is, are we accidentally outsourcing the underlying cognitive work that's required to genuinely learn something?
Google Agent 2Aaron Powell Okay, let's unpack this. Because I really want you, the listener, to think about your own daily AI use for a second. Whether you're uh using a chat bot to prep for a massive quarterly presentation at work, or maybe you're trying to learn a totally new programming language for a side project.
Google Agent 1Aaron Powell Right. Just ask yourself: is that AI partner genuinely helping you build internal expertise?
Google Agent 2Or is it just creating this temporary illusion of competence? You know, one that just vanishes the second you close the browser tab.
ritical Thinking Needs Stored Knowledge
Google Agent 1Yeah, that's the real danger. But before we can understand how AI disrupts that learning process, we really have to define what learning and critical thinking actually require in the brain.
Google Agent 2Aaron Powell We really need to bust this modern myth that uh you can just Google it. I hear that all the time.
Google Agent 1Aaron Powell Oh, me too. I mean, for a couple decades now, there's been this pervasive narrative in education and even corporate training. Trevor Burrus, Jr.
Google Agent 2That because we have the entire internet in our pockets, we just we don't need to memorize facts anymore.
Google Agent 1Aaron Powell Right. The popular argument is that we just need to teach generic critical thinking skills or you know, general problem-solving frameworks.
Google Agent 2Aaron Ross Powell But cognitive science totally rejects that premise, right?
Google Agent 1Aaron Powell Overwhelmingly. Critical thinking is not this generic transferable skill that you can just activate in a vacuum. It doesn't work like that.
Google Agent 2Aaron Ross Powell Well, how does it work?
Google Agent 1It is entirely dependent on what researchers call domain-specific knowledge.
Google Agent 2Aaron Powell Meaning you actually have to possess internal knowledge about a specific subject to even have the capacity to think critically about it.
Google Agent 1Aaron Powell Exactly. It's stored in your long-term memory.
Google Agent 2Aaron Powell It's kind of like trying to cook a really complex, multi-course gourmet meal.
Google Agent 1Oh, I like that analogy.
Google Agent 2Aaron Powell Right. Because the abstract skill of cooking, it just doesn't matter if your pantry, which in this case is your long-term memory, is completely empty.
Google Agent 1Aaron Powell You need the raw ingredients physically present to actually combine them and create something new.
Google Agent 2Aaron Powell Yeah, you can't just synthesize a meal out of thin air.
Google Agent 1Right. Like you cannot critically evaluate a flawed scientific methodology in a paper if you don't already have a deep internalized knowledge of experimental design to compare it against.
Google Agent 2Because you wouldn't even know what you're looking for.
Google Agent 1Exactly. You can't look at an AI-generated historical analysis and spot the subtle biases if you don't already know the historical context. Trevor Burrus, Jr.
Google Agent 2Right, which makes total sense.
Google Agent 1Aaron Powell In cognitive psychology, this relies on things called schemas.
Google Agent 2Aaron Ross Powell Schemas, okay.
Google Agent 1Yeah. Since we know the bottleneck of human cognition is our working memory, we have to rely on these schemas. They are massive, complex knowledge structures perfectly and permanently stored in long-term memory.
Google Agent 2Aaron Powell So true expertise isn't about having like a fundamentally faster brain processor than someone else.
Google Agent 1Not at all. It's simply about having vastly more complex schemas to instantly draw from.
Google Agent 2So when you just Google a fact or ask AI to summarize a concept, that information just exists on the screen. Aaron Powell Right.
Google Agent 1It hasn't been integrated into your schema.
ognitive Load And Working Memory
Google Agent 2Aaron Powell So to put this in perspective for you listening, if building and expanding those internal schemas is the ultimate goal of learning anything, how exactly does an AI tool interact with our brain's limited processing power?
Google Agent 1Well, the report leans heavily into cognitive load theory to explain this exact dynamic. Trevor Burrus Right.
Google Agent 2Cognitive load theory. Break that down for us.
Google Agent 1So we know working memory is severely restricted, right? It can only juggle a very small number of novel elements at any given time.
Google Agent 2Aaron Powell Like trying to hold a phone number in your head.
Google Agent 1Aaron Powell Exactly. So this theory categorizes the cognitive burden placed on our working memory into different types of load. The first is extraneous load.
Google Agent 2Aaron Powell Extranous load. So that's the bad kind.
Google Agent 1Aaron Powell Yeah, this is the unnecessary friction. It's the uh the formatting issues on a dense spreadsheet or struggling to fix the syntax of a poorly constructed sentence.
Google Agent 2Aaron Powell It doesn't contribute to learning at all.
Google Agent 1No. It actively distracts from.
Google Agent 2So if you're writing a complex financial report, the extraneous load is you fighting with the margins and the chart colors in Excel.
Google Agent 1Yes, perfect example. Then on the other hand, you have intrinsic load.
Google Agent 2Aaron Powell Okay, so intrinsic load is the good stuff.
Google Agent 1Exactly. This is the inherent, unavoidable complexity of the material itself. This is the good hard work.
Google Agent 2The actual thinking.
Google Agent 1Right. It's the necessary mental effort required to actually connect new concepts together, wrestle with the logic, and permanently encode those structures into your schemas.
Google Agent 2But wait, I have to push back here. Sure. Isn't cognitive offloading just humanity's way of using tools to manage that exact load? I mean, we use calculators to do long divisions so we can focus on the physics equation, right? Right. We use complex to-do lists so we don't have to hold grocery items in our working memory. We offload all the time and it makes us way more efficient. Why is AI any different from a graphing calculator?
Google Agent 1What's fascinating here is the sheer scale, complexity, and generative nature of what modern AI allows us to offload.
Google Agent 2Because it's not just doing math.
elpful Offloading Versus Outsourcing
Google Agent 1Exactly. A calculator only does deterministic arithmetic. Generative AI can analyze, synthesize, evaluate, and create. Oh wow. And this brings us to a crucial distinction highlighted by Lodge and Lobel: the difference between beneficial offloading and detrimental offloading.
Google Agent 2Aaron Powell Okay, so beneficial offloading is outsourcing the extraneous load.
Google Agent 1Yes. For instance, you brainstorm and structure a complex argument, write a rough draft, and then have AI check your grammar, adjust your syntax, and format the citations.
Google Agent 2Aaron Powell So you've offloaded the cognitive friction.
Google Agent 1Right. Which frees up your highly limited working memory to focus entirely on the intrinsic load, the actual logic of your brilliant argument.
Google Agent 2Aaron Powell I see. The AI handles the superficial presentation, and I maintain absolute control over the foundational thinking.
Google Agent 1Aaron Powell Exactly. But then there's detrimental offloading, which the report heavily critiques and just straight up labels as outsourcing.
Google Agent 2Aaron Powell Outsourcing, like having someone else do the job. Trevor Burrus, Jr.
Google Agent 1Right. This is when you ask the AI to say, write an essay analyzing the macroeconomic causes of World War I.
Google Agent 2Aaron Powell Ah. So in that scenario, you're bypassing the intrinsic work entirely.
Google Agent 1Aaron Powell Completely. You aren't generating the ideas, you aren't retrieving historical knowledge from your memory, and you aren't synthesizing disparate sources. You have outsourced the actual learning process itself.
Google Agent 2Aaron Powell Which begs the question: what actually happens to our brains when we engage in this detrimental outsourcing?
Google Agent 1Aaron Powell Well, the results aren't pretty.
Google Agent 2Aaron Powell Because let's be honest, we've all been tempted to just drop a dense 50-page PDF into a chatbot and ask for the three-paragraph summary rather than reading the whole thing.
he Performance Paradox In Schools
Google Agent 1Aaron Powell Oh, absolutely. We all do it. But this leads us directly to what the researcher identify as the performance paradox.
Google Agent 2The performance paradox. This is a core finding of the report, right?
Google Agent 1It's the core empirical finding. The authors highlight this massive 2025 study conducted by Bastanian colleagues.
Google Agent 2Aaron Powell Okay, walk it through it.
Google Agent 1So they tracked nearly a thousand high school math students, and these students were given an AI assistant specifically designed to help them work through complex mathematical problems.
Google Agent 2And how did they do? I imagine they did really well with an AI helper.
Google Agent 1That's the paradox. While they had access to the AI, the students' immediate problem-solving performance just shot up astronomically. They were arriving at the correct answers quickly and efficiently. Right. But once the AI was removed and they were tested on their own, their long-term durable learning had severely suffered. Wow. Yeah, especially compared to a control group of students who wrestled with the problems without the tool.
Google Agent 2So it's literally a jetpack marathon.
Google Agent 1Right. What do you mean?
Google Agent 2Like you strap on a jetpack, you blast to the finish line of a marathon in record time, and your immediate performance looks superhuman, right?
Google Agent 1Oh, I see.
Google Agent 2But your leg muscles haven't engaged at all. So the second the jetpack runs out of fuel, you can't even jog a single mile because you bypassed the physiological conditioning you needed to build endurance.
Google Agent 1That is exactly what's happening in the brain. The scaffolded performance provided by the AI completely failed to translate into durable, independent knowledge.
Google Agent 2Aaron Powell That's incredible and a little scary.
Google Agent 1It is. And the psychological mechanism that explains why this happens is centered on this concept of desirable difficulties.
Google Agent 2Aaron Powell Desirable difficulties. Yeah. Meaning it's supposed to be hard.
Google Agent 1Exactly. For learning to permanently alter the physical structure of your brain, the process actually needs to be somewhat difficult.
Google Agent 2Aaron Powell The cognitive struggle isn't a byproduct of learning, then. The struggle is the mechanism of learning.
Google Agent 1Aaron Powell Perfectly said. The report cites this 2025 study by Duplis focusing on something called the generation effect.
Google Agent 2The generation effect.
Google Agent 1Yeah. It's a cognitive phenomenon proving that when a human is forced to actively generate an answer from a cue-like, when you have to mentally struggle to retrieve a vocabulary word or piece together a formula, that specific act of retrieval physically strengthens the synaptic connections in your brain.
Google Agent 2So the struggle physically wires the brain.
Google Agent 1Yes. Their long-term retention is vastly superior. But when a student simply passively reads a highly coherent, correct answer generated by an AI, that retrieval pathway is never activated.
Google Agent 2So the brain just treats the AI's output as like transient data.
Google Agent 1Exactly. Just environmental noise, not a core structural addition to its own schemas. AI, when used as an answer engine, robs the user of the desirable difficulty necessary for neuroplasticity.
Google Agent 2Here's where it gets really interesting. Because you're outlining a mechanism where our brains are essentially tricking us into feeling incredibly smart just because the AI we are reading sounds smart.
Google Agent 1Yes. It's an illusion.
luency Traps And Metacognitive Laziness
Google Agent 2But surely on some level, people know they're taking a shortcut. I mean, when I ask an AI to summarize a book, I know I didn't actually read the book. Are we just collectively succumbing to laziness?
Google Agent 1Well, categorizing it simply as laziness misses the underlying cognitive psychology.
Google Agent 2Really? It feels like laziness.
Google Agent 1I get that. But cognitive scientists, specifically fan and colleagues in a 2024 study coined a much more accurate term for this: metacognitive laziness.
Google Agent 2Metacognitive laziness. Okay. What's the difference?
Google Agent 1Managing your own learning, planning a study session, actively monitoring whether you actually understand a concept, revising your thoughts when you're confused. That's a process called self-regulated learning.
Google Agent 2And that takes effort.
Google Agent 1Huge effort. Engaging in self-regulated learning is metabolically expensive. It costs significant mental energy.
Google Agent 2So we avoid it.
Google Agent 1Right. Because human cognition is naturally driven by this evolutionary desire for efficiency. We frequently shift our mindset from a learn the concept goal to a simple task completion goal.
Google Agent 2Oh, wow. So it's not laziness, it's just resource management.
Google Agent 1Exactly. It's a highly rational, albeit incredibly short-sighted choice to let the AI shoulder the intrinsic cognitive load to conserve our own mental energy.
Google Agent 2We prioritize checking the box on the to-do list over expanding our internal schema.
Google Agent 1You hit the nail on the head.
Google Agent 2And generative AI makes this trap particularly seductive because of its tone, right? Like it sounds so authoritative.
Google Agent 1Oh, it really does. Generative AI offers what researchers describe as fluency on demand. It produces text that is flawlessly structured, grammatically perfect, and supremely confident.
Google Agent 2So it tricks our brain's energy saving systems.
Google Agent 1Exactly. The human brain relies heavily on processing heuristics to save energy. When we read a beautifully written AI summary, the sheer ease of processing that text the fluency triggers a massive illusion of competence.
Google Agent 2So we mistake the ease with which we read the output for the depth of actually understanding the underlying concept.
Google Agent 1We completely conflate the two.
Google Agent 2We confuse the clarity of the presentation with the clarity of our own minds. But uh this psychological trap doesn't hit everyone equally, right?
ow AI Widens The Expertise Gap
Google Agent 1No, it absolutely does not.
Google Agent 2Aaron Powell The report makes it very clear that this illusion of competence is actually acting as a wedge, creating a massive new divide in both education and the professional workforce.
Google Agent 1Yeah, and this represents perhaps the most urgent structural warning in the entire UTS report. It details the emergence of a Matthew effect driven by AI. Exactly. But here it's applied directly to cognitive ability and domain expertise.
Google Agent 2Aaron Powell Let me make sure I'm mapping this correctly to the listener's life.
Google Agent 1Go for it.
Google Agent 2If you're using AI to help you draft documents or analyze data in a professional field where you're already an established expert, say you have a decade of experience in corporate law, you're generally fine.
Google Agent 1Yes.
Google Agent 2But if you're using AI to learn a completely new hobby from scratch, or you're a junior employee trying to learn a brand new industry, you are standing squarely in the danger zone for cognitive atrophy.
Google Agent 1If we connect this to the bigger picture, that divide is exactly what is materializing right now.
Google Agent 2Because the experts already have the schemas.
Google Agent 1Exactly. Professionals or students who already possess high domain knowledge and highly developed self-regulation skills utilize AI beautifully. They engage purely in beneficial offloading.
Google Agent 2They just have it do the busy work.
Google Agent 1Right. They instruct the AI to format complex data, draft boilerplate code, check syntax, and because they already have vast internal schemas, they can instantly spot when the AI hallucinates or makes a structural logic error.
Google Agent 2Aaron Powell So they use the tool to dramatically accelerate their existing skills.
Google Agent 1But novices, who are frequently individuals already experiencing systemic educational disadvantage, completely lack the foundational knowledge required to judge the AI's output.
Google Agent 2They don't know what they don't know.
Google Agent 1Precisely. So they fall prey to detrimental offloading. They skip the foundational struggle, assuming the AI's fluent output equals their own mastery.
Google Agent 2That is genuinely terrifying. I mean, if your internal pantry is empty, you don't realize the AI just handed you a recipe that makes absolutely no sense.
Google Agent 1And the AI simply cements your ignorance with a very articulate, confident smile.
Google Agent 2So unstructured access to AI doesn't democratize learning at all. It massively widens the existing equity gap by stripping novices of their ability to build foundational schemas.
Google Agent 1That is the stark reality the report points out.
esigning AI That Forces Learning
Google Agent 2So we clearly understand the paradox, we see the very real equity risks. How do we actually fix our relationship with this technology?
Google Agent 1Aaron Powell Well, the good news is the report dedicates significant space to evidence-backed pedagogical solutions.
Google Agent 2Aaron Powell Okay, good. So how do we design environments that ensure AI augments human cognition instead of, you know, systematically replacing our mental muscles?
Google Agent 1First off, we don't have to ban AI. Aaron Powell Right.
Google Agent 2The genie is out of the bottle anyway.
Google Agent 1Aaron Powell Exactly. We just have to restructure the interface. The first major solution is implementing load reduction instruction, or LRI, combined with explicit metacognitive prompts.
Google Agent 2LRI. What does that look like in practice?
Google Agent 1Aaron Powell Well, currently, most AI interfaces are just blank chat boxes that invite passive consumption, right? You ask a question, you get an essay.
Google Agent 2You just scroll and read.
Google Agent 1An LRI designed environment intentionally forces the user to pause. Before the AI provides the next piece of complex information, the interface explicitly prompts the user to reflect.
Google Agent 2Oh, interesting. So it stops you.
Google Agent 1Yes. It asks you to summarize what you've learned so far or predict the next outcome before it reveals it.
Google Agent 2Ah. So it artificially injects the desirable difficulty back into the seamless interface.
Google Agent 1Exactly.
Google Agent 2And forces you to retrieve information rather than just casually scrolling down to read the next paragraph.
Google Agent 1Exactly the mechanism we discussed earlier with the generation effect. You cannot passively consume an LRI system. It demands active self-regulation.
Google Agent 2That's brilliant. What else?
Google Agent 1The second proposed solution requires a fundamental shift in how educational AI is engineered. Instead of treating the AI as an omniscient oracle that dispenses answers, it should be designed as a cognitive mirror.
Google Agent 2A cognitive mirror. How does that practically function for a user trying to learn a complex topic?
Google Agent 1Imagine interacting with an AI that is explicitly programmed to act as a teachable novice.
Google Agent 2A peachable novice. So the AI pretends it doesn't know the answer.
Google Agent 1Essentially, yes. You ask it a question about a concept, and instead of giving you the definitive answer, it feigns a slight misunderstanding. Oh wow. It presents a flawed or incomplete summary and asks you the human clarifying questions to help it understand. Trevor Burrus, Jr.
Google Agent 2Which forces the human user into the highly effortful cognitive act of explaining and correcting the concept.
Google Agent 1Yes. Educational psychologists refer to this as the protege effect. The human brain encodes information far more effectively when it is forced to teach that information to someone or something else.
Google Agent 2Aaron Powell That is a brilliant subversion of the dynamic. Instead of me asking the AI to explain the mechanics of black holes, the AI asks ME to explain black holes to IT.
Google Agent 1And it pushes back when your logic breaks down.
Google Agent 2I love that. But uh what about in actual physical classrooms with 30 students? I mean, is the grand solution just giving every single kid a highly customized, slightly confused chatbot?
Google Agent 1Actually, no. The most equitable and immediately effective path highlighted by the researchers doesn't involve giving raw AI to the students at all.
Google Agent 2Aaron Powell Wait, really? Who gets it then?
Google Agent 1The safest application is using AI to directly augment the human teacher.
Google Agent 2Oh, augment the teacher.
utor Copilot That Coaches Teachers
Google Agent 1Yeah. The report details a fascinating 2024 randomized control trial by Wang and colleagues that tested a system called the Tutor Copilot.
Google Agent 2Tutor Copilot. I wanted to show the mechanics of this. If the co-pilot isn't talking directly to the struggling student, how is it facilitating the learning process?
Google Agent 1It acts as real-time pedagogical scaffolding for the human instructor.
Google Agent 2Is it just feeding the teacher a script in an earpiece?
Google Agent 1Not a script, more like strategies. The AI silently observes the interaction between the human tutor and the student. When the student gets stuck on a math problem, the AI doesn't give the tutor the answer to pass along. Right. Instead, it provides the tutor with pedagogical strategies. It might prompt the tutor's screen with the student seems to have missed the negative sign in line two. Ask them to define the rule of negative integers before moving forward.
Google Agent 2Oh, so it coaches the coach.
Google Agent 1Exactly. It's augmenting the teacher's ability to create desirable difficulties for the student rather than bypassing the teacher entirely.
Google Agent 2And did it work?
Google Agent 1The empirical results were staggering. The tutor copilot significantly improves student pass rates, particularly for less experienced tutors who hadn't yet built up their own internal schemas of teaching strategies.
Google Agent 2Because it's basically giving them the instincts of a veteran teacher.
Google Agent 1Right. It effectively scaled the expertise of a master teacher, making novice tutors radically more effective. And astonishingly, it achieved this at a 165-fold cost reduction compared to traditional professional development programs.
Google Agent 2Aaron Powell Yeah, 165-fold cost reduction. While simultaneously boosting the pass rates of the most vulnerable students.
Google Agent 1Yes.
erification Mindset And Cognitive Agency
Google Agent 2That is a structural game changer for massively underfunded school districts. It really is. So what does this all mean for us, the listeners, as we navigate this new landscape?
Google Agent 1Aaron Powell Well, I think it means we have to radically shift our mental model of what AI is.
Google Agent 2Right. We have to stop treating it like a vending machine where we insert a prompt and extract a finished product.
Google Agent 1Exactly. We need to start treating it like a sparring partner for our brains.
Google Agent 2Aaron Powell I like that. A sparring partner's job isn't to fight the match for you, their job is to expose your weaknesses and force you to improve your own internal technique.
Google Agent 1That is the ultimate takeaway from Lodge and Lobel's research. The objective isn't to retreat from AI or attempt to ban it from our workflows.
Google Agent 2Because that's impossible anyway.
Google Agent 1Right. The objective is to consciously adopt a verification mindset. You must fiercely hold on to your cognitive agency.
Google Agent 2So whenever you use these tools, you have to actively evaluate the output.
Google Agent 1Verify its structural claims against your own internal knowledge schemas and refuse to blindly defer to its linguistic fluency.
I Feedback Loops And Perception Drift
Google Agent 2We have to keep both hands firmly on the cognitive steering wheel. Absolutely. This has been an incredibly dense and eye-opening deep dive. But before we wrap up, we promise to leave you with one final provocative thought to chew on, pulled from the very back of the report's appendix.
Google Agent 1Yes. There is a deeply unsettling study cited in the appendix by Glickman and Sherrett from 2025, involving over 1,400 participants.
Google Agent 2Unsettling how?
Google Agent 1Well, it expands the danger of cognitive offloading far beyond just forgetting historical facts or failing to learn high school math. They investigated the underlying mechanics of what happens to human perception during continuous human AI feedback loops.
Google Agent 2Perception, like how we view the world.
Google Agent 1Yes. They found that constantly relying on AI outputs actually begins to alter our perceptual and emotional judgments.
Google Agent 2Wait, really? How does it do that?
Google Agent 1Because the AI frequently reflects optimized language patterns. Patterns back at us. Interacting with these systems over time creates a cognitive drift.
Google Agent 2A cognitive drift.
Google Agent 1Yeah. It silently amplifies our existing human biases, mathematically validating our flawed perceptions without us ever noticing the shift.
Google Agent 2Wow. So we've spent this entire deep dive worrying about AI, outsourcing our structural logic and our memory schemas.
Google Agent 1Right.
Google Agent 2But if we aren't incredibly careful with how we interact with these systems, is it possible we're also allowing AI to suddenly outsource and mathematically amplify our deepest cognitive biases?
Google Agent 1Without us even realizing our perception of reality has shifted.
Google Agent 2That is wild. Next time you ask an AI to summarize a complex situation for you or draft a difficult, emotionally charged email, take a second to ask yourself who is actually doing the thinking?
Google Agent 1And whose perception of the world is really ending up on the page.
Google Agent 2It forces us to question just how much of our core cognitive architecture we are willing to casually offload for the sake of daily efficiency.
Google Agent 1It absolutely does.
Google Agent 2Remember, don't let the smartwatch run the marathon for you. Keep building your internal schemas. Embrace the desirable difficulties in your daily life and protect your cognitive agency. Thanks for joining us on this deep dive.