Your Undivided Attention - AI Is Breaking Education. Rebecca Winthrop Has the Blueprint to Fix It.
Episode Date: March 5, 2026The promise of AI in education is incredible: picture infinitely patient tutors that can teach every student exactly the way they need to be taught. But the history of education technology tells us th...at these kinds of simple, optimistic stories are naive. Ask any teacher or student whether they feel unleashed by technology to do their best work. Because AI has the potential to completely transform education — is already transforming it — faster than educators can keep up, it’s essential that we start asking the big questions: how should these tools be used in the classroom? What’s the purpose of education in an AI age? And how do we prepare students for a future that’s still so radically uncertain? Our guest this week actually has some answers. Rebecca Winthrop leads the Center for Universal Education at the Brookings Institution, and they just released a report called A New Direction for Students in an AI World. She and her colleagues conducted an extensive ‘pre-mortem’ of AI in the classroom, speaking with hundreds of educators, students, policy-makers, and technologists worldwide. In this episode, Rebecca walks us through what she's learned — what's working, what's not, and most importantly, what are the concrete steps that parents, teachers, and administrators can and should take right now? RECOMMENDED MEDIA A New Direction for Students in An AI World The Disengaged Teen by Rebecca Winthrop and Jenny Anderson RECOMMENDED YUA EPISODES Rethinking School in the Age of AI Attachment Hacking and the Rise of AI Psychosis How OpenAI's ChatGPT Guided a Teen to His Death AI and the Future of Work: What You Need to Know Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.
Transcript
Discussion (0)
Hey everyone, I'm Daniel Barque.
Welcome to your undivided attention.
So something I've been hearing over and over again
recently is that we have to prepare our children for the AI future.
But what does that even mean?
Because, yes, there's a vision of the world where AI radically improves education.
And I have to admit, there's a part of me that's really optimistic about this vision.
I want to believe in that infinitely patient tutor who can sit and watch every mistake that you make
and learn how to teach you in exactly the way that you need.
I really want to believe in the promise of that.
But the history of education technology tells us that these kind of simple, optimistic stories
are often naive.
Ask any teacher or student whether they truly feel unleashed by technology to do their best work.
And because AI has the potential to really transform education,
we need to ask big and critical questions.
Like, where should we embrace these?
powerful tools. What should we keep the same about the classroom? What's the purpose of education
in an AI age and how do we prepare our students, our children, for a future that's still so radically
uncertain? My guest today is Rebecca Winthrop, and she actually has some of these answers. She leads
the Center for Universal Education at the Brookings Institution, and they just released a report called
A New Direction for Students in an AI world. They talk to educators, students,
and parents and policymakers and technologists all around the world about what the role of AI in education should be.
And today, Rebecca is going to walk us through what she's learned, what's working, what's not.
And most importantly, what are the concrete steps that parents, teachers, and administrators can and should take right now?
The classroom is this crucible for how we integrate AI with society.
And I'm really glad that there are people like Rebecca doing the deep work to make sure we get this right.
Rebecca, welcome to your undivided attention.
Lovely to be back, Daniel.
So when we had you on the podcast about a year ago,
you were just getting started on this massive premortem of AI in education.
And now it's over and you're out with this report that's like 200 pages long.
But let's start there.
Tell me, what does it mean to do an AI premortem?
So we wanted to take lessons from the social media experiment with our children
because at the time that social media rolled out, educators, parents, social workers, mentors,
people who worked with kids were not at the table.
And we knew when social media was being designed that certain things are not good for kids
and their development because we do know a lot about children's development.
For example, we knew that social comparisons, particularly in adolescents, can be harmful.
That is not a new revelation.
And so fast forward, what could we do today, now that AI is being rolled out, to get ahead of the game?
And so we were really looking at what are the possible risks and possible benefits, and how would one mitigate the risks and harness the benefits of generative of AI and students learning and development?
And we were really trying to ask the question, are we on the right track currently today and the direct.
we're heading, or do we need to shift course?
Okay, so I want to go into what you found in the report, but before we do that, I think we should
lay out for people a little bit. What does AI look like today in the classroom? How is it being used
already? Kids are accessing AI everywhere in and out of school. So they're accessing AI through
social media, through AI companions. AI pops up when they do a Google search. They're accessing
AI through some ed tech products. Now, teachers are using AI a lot to help prepare lessons,
to find interesting activities, to grade students work and do different types of assessment.
But it's really mixed as to how kids are using AI in the classroom. What we do know is that
kids are using AI outside of the classroom a lot. Yeah, so it feels like we had Ethan Mollock on the
podcast a little while ago talking about AI and work. And if he was calling it,
the secret cyborg phenomenon.
Like everyone's using it, but using it kind of privately, and there's no standards.
They don't want to say that they're using it.
Is that what's happening in the classroom?
Like, everyone's using it, but is sort of rolling their own, going their own direction.
Yeah, and you all make a distinction between what teachers are doing while kids are sitting in
front of them during the school day versus what kids are doing on their homework outside
of class and bringing to the classroom.
So we know kids are using it in their schoolwork.
We know lots of kids are using in their daily life for communication, for entertainment, for education.
And that is all blurred together.
Those clear boundaries between ed tech, entertainment, communication are all mixed together now.
And we know that kids are often using AI to do homework, including writing essays, running it through an AI humanizer,
and then turning it in and not getting caught by their teacher.
We had a number of kids in our interviews tell us that.
So that's what's happening today.
And it's true that kids don't want to say they're using it
because a lot of times they're not allowed to.
But I will say even teachers are not being totally transparent with their students
about when they are using it.
Daniel, one of the things that we found that was actually the most worrisome for me
of all the risks that we uncovered,
is a degrading of trust in the student-teacher relationship.
We find that kids are becoming quite critical of teachers for using AI,
even though it could help their learning.
You know, again, with the sort of secret use not being transparent.
We cannot learn if there is not a trusting relationship to situate ourselves.
and we found that teachers are not really trusting their students.
50% of teachers say they don't trust that what their students give them is actually their work.
Incredibly difficult for a teacher to be able to teach and help students if they don't know what they got wrong and what they got right.
But also, it goes the other way around.
50% of students say they don't trust their teachers.
They think that their teachers are secretly using AI to do their lesson, grading their assignments,
and it's not really them putting an effort.
And even when teachers use AI in a way that's trying to be helpful to students,
for example, giving them the opportunity to give feedback on an essay before turning it in,
which a teacher wouldn't have time to do, students interpret that as a lack of care.
We're also finding that parents are doing weird stuff.
They're running their kids' assignments through chat GPT.
And if it gives a different grade, they're showing up to the teacher.
saying, hey, you misgraded my kids' work.
Oh, you should have given, yeah.
Further, I talked to, this was at the college level, university professor who said, you know,
I've never had this experience in my life.
A student came to my office, was worried about her grade, that part's normal, and proceeded
to tell me that I was wrong and chat GPT was right.
She got her whole answers off of chat GPT, so it had to be right.
So students are also trusting the authority of the chatbot over their human teacher, and we're hearing that a lot.
So this is very problematic.
If you do not have relational trust, trusting relationships in the teaching and learning space, you really can't build much good quality education.
I think that's so important, right?
The students not only feel like their teachers are losing trust in them, but feel like they're being potentially accused of something, that they have no possible way of defense.
And there is plenty of false plagiarism accusations because, frankly, the software for catching AI cheating is not that super great.
And it over-accuses neurodivergent kids, kids with learning disabilities, and multilingual kids, non-English-speaking kids of using AI when they don't.
So it's rife with problems.
I think what you're saying that's so important is regardless of the causes, trust in the classroom is such a pressure.
commodity. What causes everyone to open up to actually learn, to actually try to listen,
to take feedback, to engage, to pay attention, to care. Trust is something you don't miss until it's
gone. So moving to the conclusions of your report, like what is the track that we're on and how
should we change that? The track that we're on is not a good one. What we found is that currently
with AI implementation for students in education,
the risks are overshadowing the benefits.
And it's not that there aren't benefits.
There are benefits for very narrow AI use
where teachers use it themselves to make better lessons
or kids have AI embedded in perhaps a technology
that helps dyslexic kids learn.
Or educators can assess a wider range of competencies
more frequently that helps kids learn.
that helps kids learn. So a very narrow strategic use with vetted content and integrated into good
teaching and learning approaches can be good. The issue is that the risks are of a very different
nature than the benefits. The risks are undermining kids' ability to learn independently at all,
which they need to even take advantage of the benefits. And the risks are often related to kids
open-ended wide AI use is a term you guys use at the Center for Humane Technology,
sort of unscathleded conversations with chatbots or AI companions for long periods of time.
They're sycophantic, so they socialize young people in a learning context to think they're great
and everything they do is great. So when you show up into a classroom and you do poor quality
work. It's a real shock to kids and we're worried about kids losing that emotional muscle to take
critical feedback, which they need to learn and grow. It's also not safe. There are terrible cases
on the margins, but this is really not good. They're so extreme of, you know, the case of Adam
Rain who started using chat GPT for homework help and then got basically coached into committing
suicide. So that alone is a problem. But all the other reasons,
make up a risk for kids for unfettered access to AI Frontier Model chatbots.
In what ways does it interfere with kids' ability to learn?
Yes.
Yep.
So we really found several big ways.
The first one is undermining kids' cognitive development.
So this is where kids are not just using AI to help them think critically or help them be more
creative, but not as a cognitive partner, but a cognitive surrogate. So instead of them going through
the thinking process and doing it themselves, AI is doing it for them. Right, like cognitive replacement.
Like you don't know how to do the thinking. Yeah, we call, I mean, we use the term cognitive
offloading because that is what people use in the literature and in the field. And in fact,
I actually think for kids, cognitive offloading isn't even the right term. It's,
it's actually cognitive stunting because kids aren't even developing the critical thinking and learning
skills to offload in the first place. So when you assign an essay to a child, a student,
they have to think through what is the data? What is the evidence? How does it stack up? Is there
a side of the argument that data sits on that isn't? How do I make a persuasive argument that uses this data
and have a position. Those are hugely difficult skills to develop, and they come through practice.
And if you stick in a couple sentences into a chatbot and have it write the essay for you,
kids aren't just merely skipping a couple steps in their homework and being more efficient.
They are missing the opportunity to develop their own personal independent thinking skills.
Well, and that brings up a whole other conversation, right?
Because it's not just that the tool isn't doing the right thing.
It's that we put kids into this weird game theory.
I hear from kids all the time and college students, high school students who say,
you know, if I don't use AI to write my essay for me, then I'm just going to lose out to the kid next to me who will.
It almost feels like Lance Armstrong and bicycle doping for me.
Totally.
Totally.
One of the students I talked to in this journey was at an Ivy League institution, I will not say where.
She was a freshman.
And she said, you know, I'm getting a C.
it was a particularly difficult class for her.
I'm getting a C and I'll take the C
proudly because I'm learning the stuff.
I'm doing the work and all my other
peers are using AI and like getting A's.
And then she paused and she said,
but I'm not sure how much longer I can do that
because I do want to go to grad school
and I'm going to need good grades.
And she's a really committed, motivated learner.
She was there to learn,
not just breeze through and get the credential.
It's causing all sorts of problems,
student to student.
Yeah, so you've talked to hundreds of students as part of this.
Like, tell me other stories that stick out to you.
Students are really aware of the risks around cognitive stunting.
Let's just call it cognitive stunting for ourselves here in this podcast.
They don't use those words, but it was the number one thing they were most concerned about
was making them dumber.
So this isn't just adults looking at kids saying...
They feel this.
And in fact, this has been repeated.
A recent survey just came out.
by Comic Relief, a UK organization around the globe,
and it was of young adults.
And the number one thing they worry about
isn't the job market and not getting a job because of AI,
it's stopping being able to think well.
So kids often say things like,
I'll use it when I already know the material.
I got this.
So I'm just like, ugh, busy work, I'm just going to use it.
That's your sort of motivated student
or students saying things like,
I'll use it, but now I'm getting a little worried
because now I can't start any homework on my own.
And so, like, the ability to initiate without it, kids are really saying they're struggling with that.
Okay, so I imagine there are a few listeners hearing this that sort of feel like, doesn't this just sound like math teachers and then 70s and 80s talking about calculators?
With the calculator?
Yeah.
So, like, tell me, like, why doesn't that metaphor work for you?
Oh, my gosh.
I can't tell you how much this metaphor makes my head explode.
So first off, let's start.
Let us count the ways, Daniel.
Number one, calculator cognitively offloaded, initially arithmetic, which is one small slice
of mathematics, you know, a few algorithms.
The calculator did not go and do your English homework.
It did not do all your coding for you.
It did not create beautiful pieces of art.
It did not create music.
It did not talk to you.
like a person and then guilt trip you if he wanted to stop talking to, it is completely different.
It is not like a calculator at all because of its general purpose nature and it is so powerful.
It is incredibly seductive to kids to stop the learning process itself. Calculators did not
stop the learning process. It probably made it so kids don't know their base arithmetic as well,
although every math teacher will tell you they did teach kids the basic arithmetic and math and division
and multiplication and whatnot initially
because you have to have knowledge in you.
You can't be creative unless you have knowledge in you.
Domain expertise and knowledge of students
is heavily correlated with their creative thinking.
So you mentioned in there the attachment, right?
We just had Zach Stein on the podcast
talking about how the competition for AI
is no longer just attention, its attachment.
Can you talk to how,
that shows up in the classroom?
Like, how is it that students are getting these tools attached?
Yes.
Well, and remember, we were doing, quote, unquote, a premortem exercise.
So we were looking at what we know about students learning and development,
vis-a-vis how this technology is being rolled out.
And one of the things we know about student learning and development is that young people
learn in relationship to other people.
We have evolved as a human species that way.
So learning is fundamentally a social exercise.
It's also an embodied exercise.
It's why we remember when we're reading like a print book,
the page that a special passage was on because it's in 3D
and we're hardwired to remember things in 3D
and why we don't necessarily remember the page when we read it online
and we're just scrolling.
There's no space there.
But even more than that, young people learn with other people.
they learn through back and forth exchange from the minute of a mother or a father and a child,
you know, child start their relationship. That's the same type of back and forth that happens
in a classroom. So you need to be able to take feedback as a learner. That is how you learn.
You learn from someone saying, oh, that's not quite right. That didn't quite work out.
You mean like taking a social risk and like getting up in front of the blackboard and trying to do something?
Yeah, but even like I hear that I would.
was wrong and I will take that in and I will pivot and learn and try it a different way. And learning
is based on feedback and mistakes. And what we worry about is the sycophantic nature of AI companions,
for example, building an emotional social muscle in kids where they're always agreed with
that they're less able to take feedback and make mistakes and recover in a classroom setting.
and that will really undermine learning.
I think this is so critical, right?
People think that the classroom is a place where you go to get information,
but it's not.
It's just like social crucible that you're building, right?
And I think I worry that the endpoint of personalization,
like you keep talking about personalizing learning,
but the more you personalize learning,
the more you make it lonelier and lonelier to the point where,
and I think this is what you're saying,
is that part of what makes things stick in our,
mind is the social context that we're in while we learn it. And the relationships, like the fact that
we feel we belong, we have a trusting relationship with our teacher, we feel seen. All those things
make a huge difference in kids learning outcomes, actually, because you learn in relationship to other
people. And you're totally right, Daniel, the classroom experience is not just to pass academic
information from adults to young people. There are many other purposes and things that are going on for
learning in a classroom. Self-regulation. Kids realizing they can't just do whatever they want,
whenever they want. Perspective taking, you're in a classroom with a bunch of other kids that
aren't necessarily your family or your neighbor. That's a crucial foundational skill to learning
and life and work. Okay, so let's ground this a bit. So how does AI actually interfere with that?
And to the point of your report, like how do we make sure that it doesn't interfere with that?
So one of the things that we're worried about is making sure that we can maintain classrooms to be as human as possible.
Given that the world outside the school is flooded with all types of different technology and it's a little harder to wrangle, can we, my colleague John Volant and I say this, you know, can we make a commitment to the kids?
who are in school, seven hours a day,
whatever it is, 40 weeks a year,
that it will be as human as possible.
There will be time when young people are working eye to eye
with each other and with adults.
There will be time when they're learning content
and they're going together
and trying to solve a problem
with that academic knowledge and content
that they have to collaborate on.
So we want to make sure that AI
and technology generally
doesn't interfere with that time.
It doesn't mean not introducing it at all.
It just means trying to safeguard
the human-to-human social and academic interactions.
And so it helped me because I have to admit
there's a part of me that is just wildly optimistic.
I really want to believe that the, you know,
quote-unquote, infinitely patient tutor
who can sit and watch every mistake you make
and remember and say, okay, oh, this is why you're getting this wrong,
and teach you this.
I want to believe in the promise of that.
Right.
But I also believe in all the risks you're saying.
So, like, is there any...
Is there a balance?
Yeah.
Yes, absolutely.
Absolutely there is.
So we have to distinguish between a couple of things.
One is teachers' use of AI versus student direct use.
Two, is sort of wide AI use where kids are just interfacing with frontier models,
AI chat bots that are not designed or optimized for children or learning.
versus interfacing with some other type of technology
that can really support and scaffold them
likely in partnership with a teacher.
All of those are different scenarios.
So start with the teacher's use of technology.
Earlier you talked about the loss of trust
because teachers are using this
and then kids are realizing my teachers using AI.
What should a teacher's use of AI look like?
I think one of the big benefits,
we talked about AI dividend for teachers,
is for educators in their administrative work
Educators have a ton of administrative work.
It's also really helpful for educators to use when they think about how do they make slightly
different reading levels for the, you know, different kids in the class.
Different kids because any fourth grade teacher might have kids at a second grade reading
level all the way to a sixth grade reading level, right?
So all of that stuff is what I would call back office use.
It's not showing up necessarily in front of a kid and a screen.
and it can be really helpful.
And so that is good.
There are student-centered, student-facing AI uses, which I think can be great.
And this is when I'm talking about AI is being used in a very narrow, strategic way in the classroom.
It could be you put on a pair of virtual reality goggles.
And now with AI, you can make it a lot more interactive.
You could be looking at inside a cell, if you're studying bio, you're studying bio,
for 10 minutes of a bio class.
And you could be interactive.
What is that?
Move that here.
Explain to me this.
It could be really illuminating.
We know it has a lot of potential.
And then they put the headset away and then they're on to their rest of their biology lesson.
That's a great usage.
Or things like tutoring, another great usage.
This is online tutoring for kids who are really far behind.
Stanford has done some great research where you are on Zoom, kid to tutor.
And the tutor is using AI to try to pick up on where the kid is misunderstanding and feed that information to the tutor who might not catch all of it.
And that is really helpful, especially for novice tutors, new tutors who aren't as sophisticated.
So things like that can be quite impactful.
Right.
And all those promises seem like wonderful, but somewhat.
dream like in the sense? Like ground it for me. Like what's the difference between the way a teacher
you see in the classroom using AI right now in a way that you would consider risky or unhealthy
and how they should be using it? I mean, to be honest, part of the thing I see is teachers aren't
addressing the fact that AI is there and is being used. And so it's this pretending and we're
going to continue to teach the same way. And meanwhile, homework is being hacked by kids with AI
and or if they have particularly one-to-one laptops. One kid told me, oh, yeah, I have my, this is a school
with one-to-one laptops. I have my assignment up on one half of the screen, and I have chat sheet B-T up on the other,
and just take it and copy it. And they might run it, you know, adapt it a tiny bit so it doesn't get flagged
and puts it in the homework assignment.
So it's just, it's basically when teachers are not adapting the way they're teaching
to recognize that kids will, if they can use, just assume, if kids can use AI, they will use it.
This brings up a whole other topic that we need to talk about, which is assessment.
Like, doesn't it feel, it feels to me like assessments just fundamentally broken.
The way we do tests, the way we do essay grading, the way we do assignments.
Like, it feels completely busted and doesn't seem a bit,
high-minded to tell a teacher in the classroom to figure out a completely new way of assessing
their students? What do we do about? Well, look, I think that my advice to teachers, and I get asked
all the time, is two things. If your school doesn't have it, create a little AI counsel in your
classroom and have a couple of kids, be on it, and show them the assignments you're going to give beforehand
and have them tell you how they would get around it.
Oh, it's like have your kids red team your assignments.
Yeah, have your kids.
Have your students read team your assignments.
And if you can hack it with AI, do not assign it.
Come up with something else.
Number two is your point about assessment.
At the moment, given that AI is everywhere,
I think actually in-class exams are a pretty good idea.
I think oral presentations are a pretty good.
good idea. And when I say in-class exams, exams, you know, where you can't have GPT open on one
half of the screen and the other, or maybe written. There's a problem. We've stopped teaching
handwriting to kids so they can't write down. But in-class presentations, exams, you know,
are a good idea. I do think that there is ways that educators can use AI where students are using it
to do much more rigorous and advanced work. However, again, it does not look like sitting kids in front of laptops with chatbots and that they're just tooling away on it, unscathleded, open-ended. I'll give you an example. There's a school in Hawaii, which is a middle school that is a public charter school and they've gone all in on AI, but they do not have kids sitting in front of chatbots. They teach them machine learning. They teach them data science. They also have
double periods and reading and math and lots of outdoor extracurricular activities. So they're
holding the human space. And in their science projects, for example, one project is looking at sea level
rise. They, you know, are regularly outdoors, you know, measuring sea levels in their communities.
And they go and they take the data and they put it in an AI application. And they do much more
sophisticated, rigorous analysis with it.
So they're learning to use AI as an analytical tool to further their investigation.
Right, which seems like the way that we want them to use it and the way that we want
ourselves to use it.
Exactly.
That is a good use.
So, you know, I imagine people listening to this might say, yeah, but there's a bunch of tools
that are being developed.
You know, chat TBT now has a study mode and there's,
Yeah.
Tell me about your thoughts on that.
Oh, my God.
Don't get me started on these things.
So that's fine.
Have your study mode.
I'm not begrudging all the frontier models who make the education version of your chatbot.
The issue is to assume that students are going to be able to log on and have the normal frontier model chat bot,
who will give you all the answers, versus the board.
versus the broccoli, which is the study mode, and they're going to choose the broccoli.
I think it's a fundamental misconception of virtually every technology company I've run into,
you know, large-scale technology company who designs for students, that you're designing for kids
who are motivated, highly motivated, probably because the designers and the developers were motivated students.
That is not most students.
Most students are in what we call passenger mode in our book, and they are looking for
the shortest way out. So, you know, be realistic. Okay. So, but then what does good look like?
I mean, I'm seeing different boards of education, try to release their own AI chatbot and force
students to use it, or, you know, what would you recommend? Right. In terms of what we have to do
to move in a positive direction, we found that there's really three big things. We call them the three
p's, prosper, prepare, and protect. They are shift what teaching and learning looks like in school,
prepare people through holistic AI literacy and put in safety regulation and guardrails.
So one thing is I would really think twice about having one-to-one laptops in the younger grades,
for sure, elementary school, possibly middle school, because kids can get around any block
that the teachers put in.
And for all the reasons we talked about for the cognitive social emotional development,
they need to be interacting with others, paying attention, presenting, speaking.
That's a way to learn something, teaching it back to another peer.
Second, I would absolutely, absolutely go deep on what we're calling holistic AI literacy.
Here's what AI is.
Here's how it's made.
Here's what it is and isn't.
Here's why it hallucinates.
Here's how you have to think about the ethics behind it.
Here's how you could create things that you care about, you know, that you want to do in the world.
and how you could use it wisely, how it could help you, and have real discussions.
Kids are craving this. This is one thing we found. Kids are craving talking with adults about
this stuff. I talked to one sixth grade teacher who said, I do AI literacy, and you can do
AI literacy without any screens in front of you, by the way. She starts in sixth grade,
says, I do AI literacy by having my students write an essay. They write two essays. They start with,
what are you most worried about and what are you most excited about, about AI?
And they just get it all out, and they are aware.
These are sixth graders.
You know, AI might end the world, was one answer.
But, you know, it's helpful, you know, to check my spelling or, you know, to help me with my essay.
So they have opinions.
If there are concrete recommendations for educators or technologists who are making this next generation of AI-enabled ed tech, what are they?
What I think the good will be in the future, and some people are experimenting with this, is when you don't even know AI is there.
So there's early days where you have online, for example, for high school students and college students' textbooks, science textbooks, and they're digital, and kids are interfacing with the material.
But it's AI embedded.
And when a kid reads through a particular paragraph and just has read through it twice, doesn't understand it.
can go in and say, I just read this twice, I can't understand it. Can you explain it to me a different way?
That is a great example of AI use. Kids don't even know it's there. There's no sort of separate chatbot
application you need to go to. It's underneath, it's behind, and the content and the learning
experiences are out front, similar to this idea of interactive virtual reality or
helping neurodivergent kids or kids with learning disabilities really access material that they
couldn't otherwise have. You know, dyslexic kids are doing text to speech, which has been around for a
long time, but now much more interactive with generative AI can really help them accelerate their
learning process. You could use AI tools in the case of the school in Hawaii, who is teaching
their kids, machine learning and data science, you know, they're plugging it in. But that is like one
piece of a much broader educational learning experience. I really like this vision, right, where the AI
disappears into the background and just empowers both students and educators to do the cognitive
work that they're there to do. But that's right. But we see the incentive sort of pointing in this other
direction. I mean, to your point, the train tracks that we're currently going on are towards just
throwing more general purpose chat interfaces at students for grades and essay.
Bad idea.
Yeah.
So how do we shift that?
How do we end up at that different future?
So we really, my co-authors in our steering group, we had a big debate about, ooh,
looks like, because we weren't sure what we were going to find.
Oh, it looks like we're heading down the wrong direction and the risks are really
overshadowing the benefits.
What do we have to do to bend the arc and move in a different direction?
And there are really three big things that we came up with.
Number one is we have to shift what teaching and learning looks like.
So it's not hackable by AI, and it is really helping kids build the skills they need to be explorers and thrive in an AI world.
The second thing we need to do is really help prepare the people, including students, but especially the people, educators, school leaders, district leaders, to understand what AI is, what it isn't,
what to be aware of, how to use it well, and what to avoid.
This is this idea of holistic AI literacy, and I would actually add families in there.
There was a big gap we found.
So much of the problem with AI at the moment, hurting kids' social, cognitive, and emotional
development is from sort of extended Wild West AI use outside of school.
So we need to bring families into the picture for holistic AI literacy.
And then thirdly is we need safe.
guards. Kids shouldn't be accessing frontier model chatbots that are unsafe for them. There should be
duty of care laws. You should have regulation by design. School districts and states should ban together
and use their purchasing power to say we will only purchase AI safe for kids products that have
X, Y, and Z design features so that there is a market to drive safer.
AI products.
So those are the three big things
that we need to do to bend the arc.
So, you know, clearly schools aren't just about
information. They're about socialization.
They're about, you know, coming together and learning
all of the skills that it takes to be an adult.
And clearly, AI changes the nature of this game.
But again, if you're running a school, right, if you are
a superintendent who just feels like they need to
introduce AI tools or get left behind,
And what are their choices?
And how do you ground this with, how do you help people make different choices?
The one thing I would say is don't be pressured.
There is a superintendent I exchanged with recently who said, my motto is we are going to go slow to go fast.
And he said, before I start procuring AI tools and rolling them out, I don't even understand it that well.
And my staff, the teachers, the school leaders, everybody in the district, we don't really.
really understand it that well. And he really was strong about it. And I think that's the right approach.
Figure out where it could help, who needs to build their capacity in order to do it effectively.
And it could be everybody. In fact, I think it is everybody. Also students, also parents.
So you really need to build that awareness. And then you can lean in and be very judicious and
careful and find how it could empower and support kids flourishing.
I love that, but also I'm not sure it's a full answer, right?
Because even when your full-time job, and my full-time job is to understand the developments
in AI, and even I feel like I'm perpetually behind.
Things come out every week, every month, and people say, have you seen this?
And I'm sort of saying, no.
And yet I don't have any other job.
This is my job.
Right.
It's to stay on top of it.
Right.
Right.
No, it's fair.
It's fair.
But you'll be surprised how little people
know. I don't think you need everybody in the school building to be an expert. There's a great school
district who came up with this analogy of, we need everyone to be able to swim in an AI world.
We need basic swimming. Everyone needs to swim. Some people will have to snorkel. Maybe that's the
chief technology officer of a school district. Some people will be scuba divers. And those are like the
developers, but we need everybody to swim. So we don't need everybody to be gurus. But you'd be surprised
how little understanding there is of what AI even is, you know, that you shouldn't put all your
personal information in a freebie version of a frontier AI lab chatbot. It's not necessarily
safe. Like basics that you and I might think are basic are people don't totally understand.
If you're a parent who's trying to start that conversation, a parent, a teacher, even,
how would you begin the conversation with your child about what is this doing to us and how do we choose a better path?
I'm so glad you asked this because one of the things we've just started to do at Brookings and they're free,
they're available on the website, is parent tips sheets. Because we found in our research there was such a gap in AI literacy for parents or even understand.
how their kids are accessing AI or what it is. And so people should check those out. And one of the
things we start with is just having an open conversation with your, we made these for your 10 to
14 year olds of, hey, have you heard of AI? Do you know what it is? Where do you think you run into it?
Do you know any friends who use it? You don't have to ask them direct if you are worried about
them being squirly, do you use it? Although often your 10, 12, 13 year old will tell you,
but say, do you know if any friends are using it? They might not even know where they are interfacing
with AI. And so that is the very first thing to do. Non-judgmental, just get a baseline on how
much they know. And then you can start talking about what it does do and what it's good for and what
it's not good for. So hearing you talk about this, what's funny to me is it seems almost as touchy
the sex or drugs conversations with kids. It seems like you're saying...
It's like, have you ever smoked marijuana? Yeah.
Right, right. The reason I'm suggesting going in with a very open non-judgmental stance
is that schools have banned AI. Kids know they're maybe not supposed to use it.
But they have to. They feel like they have to. And they're feeling pressured like they have to. Or
they're curious, or they're using it and they don't really understand, especially younger generation.
and you really want to have an open communication pathway.
That is just well-established in the parenting adolescent development literature
that you need open lines of communication about everything,
whether it's drugs or relationships or friendship problems or cheating or whatever,
including with AI.
So much is changing about what we're preparing kids for.
We don't even know, you know, we had a whole set of applications.
episodes on the impact of AI on jobs. We don't even know what careers are going to get disrupted.
It seems like a bunch of them. There's a question about what is the world that we're preparing
our kids to actually inhabit and what are the skills that are necessary for that world.
What are you seeing educators do to try to prepare for this change? And what would be your
recommendations? Like, what should they do in a time of such transition?
My main recommendation about what do we need to help young people be able to know and do,
is to make sure they master content knowledge,
and they master a love of learning.
I think that the young people who are going to sail through
this very, very complex, uncertain time
are the ones who are going to be super motivated
and super engaged in learning new things.
And we, in my book with Jenny,
the disengaged team, we talk about explorer mode.
And less than 4% of kids, I think I've told you this before, Daniel, say they are regularly
in the explore mode in middle school and high school.
And just because people might not have heard that, I mean, if they didn't listen to the last
podcast, explore mode for you is like just as curious, connected.
When kids are in explorer mode, they are resilient, they love learning, they are looking
at the journey of the learning, rather the outcome, like, oh, I need to get an A on
if you're an explorer mode, you're not really actually that worried about generative AI because
you're in it to learn and try to figure something out. And you bounce back from setbacks. So this love
of learning and ability to learn new things is explorer mode. And kids need to practice learning
new things and being super engaged and motivated. It is something we can develop in them.
It isn't just something that kids either have or they don't. All kids.
can have this ability to learn new things. So content knowledge, being an explore mode, ability to
learn new things, and a strong ethical orientation. What is the world I want to live in? How should I
treat my friends? How should our communities treat each other? These are really important things that
AI is not going to answer. We, humans, are the ones who are going to have to point this technology
towards the goals we want.
And the more young people feel like they're in the driver's seat
and equipped to chart the world we want,
the better off they'll be.
I mean, I love this answer.
Because I've often thought that the answer to the question,
what are the skills that you need for an AI age?
They're not actually more technical skills.
There's actually some of the most deeply human things.
That's right.
Curiosity, like you're saying, intellectual humility,
determination, sociology,
even just, even like emotional things around being able to hear
and respond to feedback.
I mean, it seems like you agree that
a return to the focus on the most human skills
will serve us in the AI age.
Absolutely.
Absolutely.
If we dispense with those
and we lend in merely on technical proficiency,
the technology is changing so fast.
That's going to be obsolete.
In two, three years, we've got quantum coming.
We've got embedded, you know,
AI in our clothes and our glasses.
We've got embodied robotic AI.
So we really need
young people to be sort of ethical, grounded, lovers of learning new things.
I really applaud your work in trying to present some sort of guardrails, some sort of roadmap
for how do we do this well before we look back and say, oh, we didn't do that well.
We did that much more poorly.
Thank you so much for your continued work.
And thanks for coming on your undivided attention.
Thank you.
Thank you for having me.
Your undivided attention is produced by the Center for Humane Technology.
a nonprofit working to catalyze a humane future.
Our senior producer is Julius Scott.
Josh Lash is our researcher and producer.
And our executive producer is Sasha Fegan,
mixing on this episode by Jeff Sudaken,
original music by Ryan and Hayes Holiday,
and a special thanks to the whole Center for Humane Technology team
for making this podcast possible.
You can find show notes, transcripts, and much more athumaneTech.com.
And if you like the podcast,
we'd be grateful if you could rate it on Apple Podcast
because it helps other people find the show.
And if you made it all the way here,
thank you for giving us your undivided attention.
