The Joy of Why - Will AI Ever Understand Language Like Humans?

Episode Date: May 1, 2025

Large language models (LLMs) are becoming increasingly more impressive at creating human-like text and answering questions, but whether they can understand the meaning of the words they gener...ate is a hotly debated issue. A big challenge is that LLMs are black boxes; they can make predictions and decisions on the order of words, but they cannot communicate the reasons for doing so.Ellie Pavlick at Brown University is building models that could help understand how LLMs process language compared with humans. In this episode of The Joy of Why, Pavlick discusses what we know and don’t know about LLM language processing, how their processes differ from humans, and how understanding LLMs better could also help us better appreciate our own capacity for knowledge and creativity.

Transcript
Discussion (0)
Starting point is 00:00:00 I'm Jana Levin. And I'm Steve Strohgatz. And this is The Joy of Why, a podcast from Quantum Magazine exploring some of the biggest unanswered questions in math and science today. Steve, hi. Hey, Jana. How's it going? Good. I wanted to tell you about this conversation I had about AI and large language models.
Starting point is 00:00:29 OK. Have you been thinking about AI a lot right now? Is it on your mind? Sure, I can't resist. It's fun playing with it. And now my interest has peaked. Well, it's interesting because Quanta actually just published a whole series of articles about AI
Starting point is 00:00:41 to kind of fill in some of the blanks that are out there in the conversation, right? Because we're kind of going over the same material a lot. Will they replace our jobs? And what does it mean for creative fields? But there's this almost neuroscience of AI. How do you understand what your AI is doing? And that really surprised me. You would think, well, you built the thing. How come you don't know what it's doing? But that's kind of like saying I had a child. That doesn't mean you have transparency into their mind. Right. This feels like a real frontier question
Starting point is 00:01:15 because we keep hearing AIs referred to as black boxes. It's as hard as us opening the black box of our minds. I mean, it's not as though I can explain to you the neuroscience of my mind as I'm talking to you, right? I don't know how this black box is working. There's an old essay by Louis Thomas at one point says something like, if I had to do consciously what my liver does, I would just be vibrating, you know? Right. A lot of what we consider consciousness, I sometimes think is because we can't process that much data. So we need the consciousness as a very quick approximation so we can do lots of tasks. We have to be able to breathe automatically. We have to be able to recognize
Starting point is 00:01:59 a chair versus a person instantly and loosely. And these have all been difficult things to teach an AI. Oh, huh. Because of its nature to want to be exact? I mean, I guess the AI will have to learn. The fact that it makes mistakes to me is almost reassuring. Oh, that's interesting. What a cool thought, because we so often make fun of them for hallucinating.
Starting point is 00:02:23 And that never occurred to me, that that might be a sign of being on the road to real intelligence. I think these advances in AI and language, specifically large language models have been really intriguing. So I had the chance to talk with Ellie Pavlik. She's a computer scientist in linguistics at Brown University and she heads this language understanding and representation lab, which is trying to understand not just language and language models, but how they actually work. We had a chance to talk about all of this.
Starting point is 00:02:51 So let's hear from Ellie. So Ellie, welcome to the Joy of Why. We're thrilled to have you today. Thank you. Yeah. This topic is really all over the news right now and it's in our lives actually, this issue of AI. Before we get too deep into it, I'm curious about your own trajectory.
Starting point is 00:03:09 You started in economics and you started playing saxophone. How did you go from that to studying computers and how they encode semantics? I always wish I had a really like literary answer where like it all comes full circle or it's like only because I began where I did, could I have ended up where I am. Some profound life lesson. Exactly. It turns out it wasn't like scripted and perfect. So I think the path into CS was very much through econ,
Starting point is 00:03:35 because I had a research gig with a microeconomics professor and the grunt work I was given was to like make plots in Matlab. And that was so overwhelming for someone with no CS background. And I was like, okay, well, maybe I need to learn how to code. So I took an intro class just so I didn't feel so out of my element. And there's this very pleasant nature to like writing a little thing and running it. It works and it does what you said. And then I've always thought I liked the idea of research.
Starting point is 00:04:03 So I started working with the one professor who was doing language stuff, but then really kept working with him because he was working more and more on semantics and that resonated like that, like tapped into something I think I was always interested in. Slightly the overachievers response, I have to make a plot,
Starting point is 00:04:17 therefore I must get a degree in computer science. I wish it was that, but I think it was like absolute confusion about what, like I didn't know what skill I was missing. What was required? It's just like I don't even understand what's going on. I don't even know what question to ask. So I can imagine years ago if you had said to somebody,
Starting point is 00:04:34 oh, I work on how computers encode semantics at a dinner party, you might have ended the conversation. But these days, has reaction changed? Yeah. When you tell people you're working on things like large language models. Absolutely. I've said this is like a blessing and a curse. So I used to say I do natural language processing, which is getting computers to understand languages like English or Chinese or Spanish as opposed to
Starting point is 00:04:55 computer languages like Python or Java. And yeah, most people were zoned out. But now it's like an open invitation to talk about all of the kind of science philosophical questions that's on everyone's mind. And we're going to ask you all of those too. Yes, exactly. Before we get into the philosophical aspects, which I do believe you integrate into your work, give us a little synopsis of what it is that you do. You said natural language processing, you said large language models, LLMs. Yes, natural language processing is like the broader field that kind of gave rise to LLMs. Yeah, so natural language processing is like the broader field that kind of gave rise to LLMs
Starting point is 00:05:26 that could encompass anything that involves getting computers to work with human language. NLP isn't really about the approach you're using, it's about the kinds of problems you're trying to solve. So before large language models, maybe you would have something like a sentiment classifier or a spam filter or information retrieval like Google search or machine translation, right? All of these tasks would be NLP and they might use machine learning or they might not. And if they use machine learning, they might use neural networks and deep learning or they might not. And so then large language models are like one type of model that are neural networks predicting the next word. And it's turned out that as a consequence of building these things, they can be used to solve lots of different tasks.
Starting point is 00:06:06 And so there's this feeling that they're subsuming a lot of the things that traditionally other models in NLP were being created to solve. But definitely I would say NLP is a broad field that cares about solving language problems using computational tools. Excellent. And then what exactly is it that you're looking into
Starting point is 00:06:22 around things like large language models and chat GPT? Yeah. So right now when I talk about what my lab does, we're basically working on large language models. The kinds of questions we're really interested in is the same questions we would have asked about humans and still do ask about humans, which is just like, how do they represent language such that they do the things they do? What does it mean to represent language and how
Starting point is 00:06:45 does that representation of language support the various kinds of interesting linguistic behavior that we get and other behavior? Now that you have language models that produce often human-like behavior and then sometimes a little bit alien, weird behavior, but obviously are so linguistic in a way that non-human things have never been before, it's just interesting to ask how they do it and then ask in what ways is that the same or different from humans and is that a difference that really matters for something we might care about like comprehension or meaning. Hmm, so let's think about this relationship between how these large language models are processing language versus how humans are. I think that's very intriguing.
Starting point is 00:07:26 I understand why we don't have immediate transparency in how humans are processing language. We didn't make humans. Evolution made humans, and we are these black boxes. We can interrogate ourselves. We can self-reflect. We can analyze other humans. Why is a computer a black box if it's human made? That is something I think people struggle with.
Starting point is 00:07:48 What do you mean you don't know how it's doing what it's doing? You made it. Yeah. It's somewhat unique where we are right now that it's a computational system that we're treating as though it's an organic system, like as though it was created by something that wasn't us. It's a hard one to answer because you really
Starting point is 00:08:05 have to answer with some kind of an analogy. And it's like, what's the right analogy? So the direct answer is, well, we understand the actual code we wrote. You can go through line by line and say, this is what this line of code is doing. But what that code is doing is it's calling a machine learning program, which means it's setting up
Starting point is 00:08:22 a set of principles and rules. But then the model is going to follow these to gradually fit patterns of data. We understand the basic constraints on how that learning happens, but you can't then explain exactly the system that comes out the other side. In particular, you can't explain why the system that comes out has the properties and the behaviors it does. There's not a direct reduction of the behavior you see from an LLM to the lines of code and the principles that that gave us.
Starting point is 00:08:50 So there's different analogies you can play with. One I really like is we have a recipe for how to make large language models and you can understand the recipe. Like, you know what the steps are that you're doing and you understand some levels. Like if I don't put baking soda in the cake, it will turn out – I actually don't know if I'm a very good baker – it'll turn out too flat, too chewy, something. And you can even do some kind of substitutes like, oh, if I don't have eggs, I can use smashed banana or whatever, and it'll have these different consequences. But that doesn't mean you understand the chemistry. You can't precisely say exactly
Starting point is 00:09:21 why the cake is this exact way that it turned out. And so I think that's an important distinction to me from being able to build something or create something and understanding how it works. And as we've moved towards machine learning and deep learning, that just pulls those two things apart. So the large language model, do I call it a computer? It must be a network of computers. How do I refer to this entity? I don't want to anthropomorphize.
Starting point is 00:09:43 I actually think this is an interesting issue even in how to talk about them. Because they're producing behaviors that until recently only humans produced, we just don't have the language for talking about that thing without using anthropomorphized language. So you call them LLMs? I call them large language models. And they sometimes are on one computer. They're sometimes on many computers. It's like a virtual entity. it's not a physical entity.
Starting point is 00:10:06 It's a meta something. So here's this meta black box. That's still a mystery. Why can't we ask it? Hey, what are you doing? How'd you do that? Yeah, so we have a complicated mathematical model, the whole goal of which is to say, given a sequence of words, predict the next word. So if I just say, I just saw a school bus drive past my house, car, yard, whatever, like you can predict what the next word might be. And that's primarily what they're optimized to do. That's what they're designed to do. And then they're doing all kinds of crazy maths to support that. But then if you say something like, why did you just say what you said? The objective is not to faithfully explain
Starting point is 00:10:47 why it just said what it said, if it even knows what you refers to here, which it doesn't, but instead to say what kinds of words are likely to come next after that question, right? And it's gonna be sourcing its understanding of what's likely to come next from having seen lots and lots of data
Starting point is 00:11:04 of questions similar to that, followed by answers. And so that in and of itself is completely untethered to any reference to the language model's internal state, for example. The way the systems are designed and trained, right, there's absolutely nothing that constrains its answer to this question to be useful or correct or accurate. There's nothing that guarantees that its explanation of its behavior not only is not right, but has anything to do with its behavior. And we have some studies that look at these explanations where we're trying to see how much what it explains its behavior actually aligns with what it does. And I've just been surprised by the degree to which they are inconsistent
Starting point is 00:11:42 with each other. And we're trying to figure out why that is because there's nothing that would objectively require it. It's the same kind of argument of like, why can't I just ask you like how your nervous system works, how your brain works? Like you're using it to not know. Like it's your brain that's telling me you don't know how your brain works. Right. And you're like, what do you mean?
Starting point is 00:12:02 Of course the mechanism by which the language model doesn't know how it works is very different than the mechanism by which humans don't know how they work. But it's still this kind of point that those two things don't really operate that way. Yeah, it does make me wonder if trying to correct the neuroscience of how a human mind works will be equally challenging problems in parallel. Are you working on sort of neuroscience aspects and how to think about this? Yeah, that's a direction I've been super excited about. Every time you work with a new discipline, it just brings in a whole new set of types of ways of thinking about things, terminology, insights, right? So it brings new stuff. There are ways in which I think neuroscience
Starting point is 00:12:35 is going to be very informative here on certain aspects. We often talk in AI and in cognitive science about levels of analysis, which is just saying there's many different ways to understand a system. But it's like this idea that like, what level should we be trying to understand it before trying to analogize them to humans? Is it more like the brain? Is it more like the mind? Is it more like society? Is it like a chaotic system that's more like multiple people and we're looking at emergent behavior because it is trained on the whole internet? And there's nothing that's like the one true analogy. And so neuroscience brings this really low-level way of thinking about how might a lot of small numerical operations allow certain more complex behaviors to emerge, and cognitive science
Starting point is 00:13:15 can provide other kinds of insights. So, but we do know some things that they're doing, which, for instance, they're looking at these semantic relationships, as you described. They're guessing what word comes next, and they're doing, which for instance, they're looking at these semantic relationships as you described. They're guessing what word comes next and they're doing this mathematically. How has that process achieved for them? So there's different types of math that are relevant here. The go-to is like the probabilistic model estimating what the probability of the next words. And so you're just saying I've seen a set of words so far and I need to encode this into some state. And then you're saying, what is the probability of the next word given this state? But then something that becomes quite complex and one of the reasons they are harder to explore is
Starting point is 00:13:51 that the way of representing that state, it's not like the coin flipping example where you say it's either heads or it's tails, right? Because there's an infinite number of these things. And so the way that gets encoded is more of a linear algebraic notion or even more calculus. It's like this high dimensional space where there's a ton of different states here. And it's really hard to know exactly what the shape of this thing is and how you move around it. And so this is where a lot of the complexity comes in. Like on the one hand, we can fairly easily think about the probability of next word given a state. And we can think about kind of there are similar states in this space and similar states will give rise to similar probabilities
Starting point is 00:14:30 And there's stuff we understand about that But it's not at a complete enough level that we can for example place guarantees or even predict the behavior of a system Without just running it. I know that you've been really careful not to Invest too much emotion in this idea that they're thinking. But how can we tell what they're understanding or if they know the information that's being provided? Yeah.
Starting point is 00:14:55 I wouldn't say I don't invest emotion in this. I feel like I've spent a lot of time thinking about this and worrying about it and caring about it, but I have it picked aside because like the thing that I'm most excited about in terms of what we can get from language models is being forced to be precise about what we mean by these things. So the thing I'm quite sure like, no, they're not human. Like in these intangibles that we're thinking about when we ask these questions about like meaning and understanding and stuff, I don't think they have it.
Starting point is 00:15:23 But I think the thing that's so hard is how intangible that thing is. The truth is we don't know what those words mean. We don't really know what we mean when we say those things, like understanding, meaning, thinking, knowing. Like any of these very anthropomorphized, very loaded words, we kind of know how little we understand what those things mean
Starting point is 00:15:42 because when we talk, we have to say stuff like, yeah, they know, but they don't really know. And bank on the fact that the person we're talking with kind of gets it. These are very intuitive concepts. And what LM's are forcing us to do is make them precise and scientific.
Starting point is 00:15:56 And I think my feeling is, as we try to do that, these words will very much fall apart into many smaller concepts that can be made precise. So the thing that we refer to as knowing or understanding is not one thing that you have or you don't have. It's like a shorthand for a collection of things, one of which might just be being human, right? Like it might be that part of what we mean when we say really know or really understand
Starting point is 00:16:23 is being a human and having all these other properties, like making a correct prediction given a certain thing and making these inferences and behaving consistently across so many states or whatever. But I think that none of these words are actually, they're just not scientific words. And we are like feeling obligated as scientists to confront them.
Starting point is 00:16:42 So the thing I stubbornly push back on is saying whether or not they're thinking, because on some aspects of what it means to be thinking, they are, right? And it's actually more productive to say, what are we actually going for? What does it mean? And very importantly, why does it matter? If we're asking it for some technical, practical reason,
Starting point is 00:16:57 they might be good enough for many cases. If we're asking it for some much deeper, much more existential reason, then they're probably not. But like actually teasing those apart is really important. It's interesting to me that you're not dismissing it outright. You're not saying, no, it's just MATLAB, you know, which is a kind of computer code that you can write. But you're not doing that right now, which is very intriguing. I'm not and definitely not everyone in my field, but a lot of people in my field really
Starting point is 00:17:20 don't reserve anything in the human mind that's not computational, right? So saying something like it's just math is like a weird dismissal. It's not clear to me that that same thing couldn't be used to dismiss what we would call natural intelligence because almost by definition, somebody who's working on trying to understand the human mind scientifically thinks that there's ultimately some model there. So it's like the dismissal on the grounds that the thing isn't human and therefore not thinking invalidates the whole field that we're in and like what was the point? You look back to when Turing began to think about mechanizing thought, which led him to algorithms and the idea of a universal machine that is a computer that used to be, human
Starting point is 00:18:02 beings were called computers. He also reflected back and said, well, you know, we're machines, too. Our thought is mechanized. I mean, we're born out of laws of physics. And do you feel that it's feeding back into your understanding of human intelligence? You're talking about it in a way where you've already said things that are very provocative along those lines. But is it making you think, well, we're kind of computational in the way the structure of our minds work, too? I wouldn't say feeding back because I think I thought that originally, hence my attraction
Starting point is 00:18:34 to the field. Again, I think there's plenty of people who work in both cognitive science and AI who think you can make a ton of technological progress and never need to go as far as saying it's possible to build actual intelligence. But many do. Many, whether they admit it or not, are drawn for a more romantic notion of what it is possible to do in AI, which is that you think humans ultimately are computational things and that there's nothing outside something metaphysical to humans that couldn't be replicated in a computer. There's actually a lot of interesting debates on this about what kinds of properties might be inherent to a digital computer versus something
Starting point is 00:19:10 else. There's a lot of room for talking about whether the digital computer itself is the right medium for replicating human intelligence. I'm open to the possibility that that's the difference, but I don't have any particular data to point to that convinces me that's the case. So yeah, I would say that I do have a fundamental belief that things are computational, right? Again, it's based on nothing, right? This is a personality trait. But if you do believe it ultimately is,
Starting point is 00:19:33 then I think you actually have a pretty hard argument to make for why being a computer precludes you from thinking, right, for why you can say it's not thinking because it's just compiling or something. I think that's actually a pretty hard philosophical argument that I haven't heard made particularly well. People are kind of holding out something special, which is the human part of what we mean when we say something like understanding.
Starting point is 00:19:54 I love it. Deep question there. It's almost like the soul free will questions, right? What is it that's intrinsic about us? And is it the mind now? Now it's the mind? Yeah, right? It used to be that living things had some vital essence that made them different from non-living things.
Starting point is 00:20:13 But when we came to believe in atoms and that we're all atoms in various states of organization, it was hard to see where the soul or the vital essence fits in there. So now what? We've retreated to saying, well, at that level, yes, we're all atoms, but intelligence, that's something else. Only we get to be intelligent. The machines are just doing math. Yeah.
Starting point is 00:20:35 It sounds like you don't buy it. I don't, but I was interested in the comment that Ellie makes that maybe there's a way out by talking about digital versus I don't know what analog that somehow that's where we get to keep the special ownership of intelligence because we're analog the way our neurons work is not exactly digital is that I mean she doesn't seem to believe that but if I heard her right makes it sound like some people think that might be the escape hatch. Yeah, I get the impression she is quite open to these digital machines thinking and that we're starting to understand how to even formulate the question.
Starting point is 00:21:16 Now, we're being pressed by these advances to formulate the question better. And what does it mean to be computational? I don't think we're doing something magical. We're doing it gooey and maybe sloppier, magically, right? This idea that consciousness is this magic. Cluj, for the fact that we're not infinitely computational, is really interesting to me. But I do think the mind is computational. And so why couldn't a digital machine
Starting point is 00:21:42 achieve something like a mind? I just wonder if we'll be able to recognize it, if it will need consciousness the way that you and I do Aha, that's another question, right? Recognize it far before we do Will it know it's aware will it be having conversations and also even it even that I'm saying it we're gonna have to start thinking differently It's not even a single entity, right? There's multiple computers that can go into a single large language model by being in the thick of it. I think we're starting to get more precise and also realizing, wow, we haven't ever really tackled
Starting point is 00:22:15 this. Beautiful. Well, there's a lot more to contemplate. So think about it during the break and we'll be right back. Welcome back to The Joy of Why. We've been speaking with computer scientist Ellie Pavlik about AI, language, and the human mind. Now, when these large language models are first trained on these enormous datasets, do they continue to learn and develop in their relationship, let's say, with the user? Or as new ideas are fed into the internet? Or are they kind of frozen until there's a big new training initiative? Everything comes down to definitions, right? It kind of depends on what you mean by learn and develop.
Starting point is 00:23:18 There's what we call the weights, which is basically it solved some really complicated set of equations to be really good at predicting next words. And those equations are stored somewhere in a file. If you want to talk to this particular instance of chat GPT or this particular instance of cloud, you basically load those equations from that file and that's who you're talking to. And so those are called the weights. And often what we think of as updating the weights as being this kind of initial learning. And there's lots of different ways to update those weights. There's update the weights themselves. There's basically add a little side file that tells you how to pretend you updated those
Starting point is 00:23:52 weights so that can allow you to spawn different models that feel like different models. But you could argue about whether they're clones of the same model or they're different models and that's a consumption question. But also a lot of the things that are being sold as learning and adapting have to do with storing a side knowledge base that could be specific to you. You have a chat with the model and say, I'm planning my daughter's birthday and I have a whole discussion about budget and her name and her friend's names and who I want to invite and where I live.
Starting point is 00:24:22 And then I come back the next day and it remembers this stuff. It's not like everyone who's using Cloud or chat GPT now has access to my daughter's name and my address. That didn't get pushed into the main model, but it still feels like it learned or developed because it has information now that didn't have yesterday and it's retained that information. So there's different mechanisms for models to learn and adapt. And depending on the particular tool
Starting point is 00:24:45 and the endpoint you're using, it might be any combination of these different things. Yeah, I'm wondering if my chat GPT is going to behave differently after lots of interaction with me than yours will with you, for instance. And as though, you know, I have my dog, and my dog is trained to behave a certain way
Starting point is 00:25:04 and react to me in a certain way. It's sort of wondering if it keeps learning and keeps feeding back in that way. Yes, there's lots of ways to customize a model to you. And maybe a useful differentiating factor is like how easy it is to reset the model so that we have the same model. In some of these versions, if there's like this add-on file that contains some information about you that this model is reading from, maybe some small things that adapt weights,
Starting point is 00:25:32 you could basically delete that file and get straight back to the exact same base model that I have. There's another version in which like, if I take ChatGBT yesterday and I train it on today's news and it updates the weights, it would actually be really hard for me to like get back to yesterday's version. I don't know which weights to go and reset. I would have to like go retrain the whole
Starting point is 00:25:52 thing exactly as it was up until I retrained it today in order to get back. And even then it might be hard. And both types of things are learning. Both things have made a change and allowed the model to develop and adapt and stuff but like some of them we can easily undo and others you can't. So they're qualitatively very different types of learning that probably have different consequences, different interpretation. It is fascinating in the human analogy where I can teach a group of students a subject, even a very mathematical subject that we consider concrete and objective, and we don't
Starting point is 00:26:22 really understand how they learn it, why some understand it more deeply and can take it further than what you taught them. And it's just fascinating that this is happening in parallel in a machine. Absolutely. I think an area that I haven't really collaborated with yet but would like to is the cognitive science of education because there's so much interesting about how do humans
Starting point is 00:26:41 learn and how do we teach them and what's going on there and how do people misunderstand things and I think there's like a lot to be shared in like what we're thinking about the black box of a LLM and the black box of a human from like education sciences. Fascinating. So you use large language models as well as study them. What's your relationship like with these large language models? I mostly use them when I study them.
Starting point is 00:27:05 I've tried to use them for a few things. I would be embarrassed to be on the record, but I've already admitted I recently got tenure and as a consequence became involved in administration. Oh yes, no good deed goes unpunished, yes. Exactly, and so as soon as I got involved in administration instead of research, I was like, oh, I start to see the use for large language models.
Starting point is 00:27:23 So I tried to do it to do things like generate the minutes of a faculty meeting, help me sort through some data I was trying to process. And actually they weren't good enough, like for even these very basic tasks. But beyond that, I haven't actually used them for many things in my day-to-day life. And I don't know if it's because a few experiences weren't quite good enough or because I'm like jaded and cynical about them despite everything I just said. Let's say there was never another update. This is it.
Starting point is 00:27:53 These are the models that we're all going to be using. So we trained them on all of our examples, for instance, translating English to French to Zwaily and back again. And now it's chaining us. Where does that put us in this chain? And will we cease to expand? Language modernizes all the time. We speak differently than we did 100 years ago.
Starting point is 00:28:15 Are we going to kind of freeze in time because we're in a loop with something? Now all our students are learning to write and speak from the chat GPTs or the clods as opposed to the other way around. The classic academic answer is like nothing is that new. I actually remember a talk I saw like early in grad school about how basically Google had trained people to use keyword searches. And this was an example of humans adapting the language technology early information retrieval but just to lead out all of your words.
Starting point is 00:28:43 If you said who was Thomas Jefferson's wife, it would just say Thomas Jefferson's wife, right? And you'd just scramble it, alphabetize it, right? That's what got you the best results out of the system at the time. Now they actually wanted the full language back and they were really struggling to get people to write full questions. And so there's already this example of people talking to a computer and adapting their language to get the best results out of the computer. And so I think you will see this. People are getting good at prompting language models and talking to language models in this way. I haven't yet seen it carry over into how people talk to each other, but technology definitely does influence how people talk to each other.
Starting point is 00:29:17 Like my Gen Z students say punctuation when they're talking. They'll say something like, do you think this is a good idea? Question mark. Like, they'll say that. And I'm like, I think this is like a spillover from like texting. It almost makes me optimistic. Language has always been very dynamic and very responsive to the technology and the context. And still, I think as long as we continue talking to humans as humans,
Starting point is 00:29:42 I think it's really cool and like cute when you see things like people saying the word question mark and dot dot dot out loud. It's like a sign of how plastic and dynamic and interesting language is. I would worry about the kind of collapse of linguistic diversity and innovation if people start talking to language models
Starting point is 00:30:00 almost exclusively. I don't know, I guess I'm an optimist. I imagine that people do like to talk to people. Even speaking as an introvert who doesn't particularly love talking to people, like I think that people will continue to have human interactions and that will save language. I appreciated when you pushed back at this idea that when computers are just doing math, that was different than when computers create poems or novels or artwork or songs. What do you think this means for human creativity? This is of course a question that people are
Starting point is 00:30:32 semi-panicked about. Yeah. So I've been teaching this class this semester with a professor at Browname, John Cayley, who's a literary artist, does poetry and other language arts projects, and has always used technology in the course of doing that. And I think it's exactly this question about our humans' mathematical objects. Like, even if you agree or grant that some neurons firing in your brain in a particular way caused you to write this poem, it doesn't devalue the poem in a particular way. Like, I don't think you have to assert divine
Starting point is 00:31:05 intervention was involved in the creation of the poem to believe that the poem itself has aesthetic and artistic value. I don't think we have to reduce it to the thing that created it in a human. And even if I understood the brand activations, it doesn't mean there's not value in analyzing this poetry. And I think the same argument can apply to language models. There is a way of thinking about what they create on its face without caring about what math and whether it was math that caused it. And there's probably room for criticism depending on what you're going for, depending on why you care, depending on who you're talking to in the context.
Starting point is 00:31:43 There's a sense in which you can say this came from a language model and therefore it's not interesting, it's meaningless, and everything in between. But I don't think like humans being mathematical devalues our creativity in any particular way. It reminds me of the sort of infinite loops of the free will and soul arguments that were unresolvable and are still debated and might be forever, but here we are and we care if people intentionally do harmful things or not or intentionally make beautiful things. That's just how we are. That's the human condition. Exactly. Again, everyone kind of relates to these situations
Starting point is 00:32:16 differently, but like if I'm thinking about the time I was like particularly connected to a piece of literature, piece of art, I don't think I spent a ton of time thinking about how causal the person was in it. Sometimes you care about the person's story, but I'm rarely hung up on whether this was preordained by the universe. That's not interfering with my ability to appreciate it. You can be a physical determinist and still... And still appreciate art.
Starting point is 00:32:40 Enjoy the Tate Modern. So I wonder if, even though you were thinking about these things and deep in this subject, if the revelation of the functional LLMs that came out practically as tools, if you were surprised by them and also do you feel in a position to predict what the future is going to be like? How rapid is this change going to be? I don't think I've been super surprised by the technology, but I think I've been a little surprised by the pace of the rollout. I wouldn't even say surprised because I think it's economically driven, not technologically driven. It's not like the technology is moving faster than I realized, or at least not now maybe. My early surprise moments were back in 2018,
Starting point is 00:33:23 2019 with what I would say were the precursors to the large language models. There's one My early surprise moments were back in like 2018, 2019, with what I would say were the precursors to the large-anguage models. There's one called Elmo, one called Bert. There was a little cute period where we had a Sesame Street theme going. Unfortunately, died after a stretch of a few models. It was a very exciting time where it felt like research was turning a corner. I think a lot of people in academia would point back to that time as being like, we're at a pivoting moment in NLP. And then there was like the
Starting point is 00:33:46 chat GBT moment, which is where it was like, suddenly pulling back the curtain and like now everyone's involved. And so that was a really important time that I think surprised me in that pace at which then the world was paying attention and reaction and then the deployment. It does surprise me how quickly people are pushing things out and how willing people are. I'm generally an optimist, but it does scare me a little bit. I think we're going to have a few oh crap moments that could have been avoided, right? What would you imagine would be a moment like that? I could imagine some kind of big security things, some kind of either intentional or
Starting point is 00:34:19 unintentional glitch or attack where a lot of systems are implicated. AI, it seems like it's lots of different technologies, but they're actually all the same technology, which makes you think they're deeply correlated errors or vulnerabilities. There's like a small amount of open source software that many things are based on. I mean, it could be overblown because a lot of things
Starting point is 00:34:40 are based on the Linux kernel, and that's quite safe. The Linux kernel being pre-Unix, which a lot of our apples run on this kind of operating system. Exactly. It's like kind of core operating system code that is then repurposed and reused. But Linux was free, right? And it was open source, and it was part of that utopian idealistic movement. And obviously, it could still have bugs in it and things, but was like understood in a level that is different
Starting point is 00:35:08 from large language models. I think there's also the obvious one that people talk about, which is just the proliferation of scams and this lack of trust, because if you don't know what that language is coming from a human anymore, you can just fundamentally start doubting everything. Like I've already felt myself do this.
Starting point is 00:35:21 Every time I see a news story or an image, if I didn't see it on kind of mainstream media, then I just preface everything with, I haven't fact-checked it myself. So I think there are a lot of these things that it surprised me how willing people are to try things out so far. We go right back to it, human beings, man. We try to be suspicious and we just kind of can't help ourselves. Yeah.
Starting point is 00:35:44 Right, Exactly. There's a question I always like to ask of our guests. What about your work brings you joy? I'm glad we turned that, because now we just talked about the pessimistic thing. But I think I ultimately am extremely optimistic. I think the potential value of the system is far outweighs the cost.
Starting point is 00:36:03 A lot of people come into AI more as dreamers than anything else. It is just very exciting. It's fascinating. There's nothing more fascinating than the human mind and brain. Of course we're obsessed with this thing. We're like a narcissistic species. It's like, we're so great. We're so incredible.
Starting point is 00:36:18 How do we work? Then the concept that we would stumble upon something computational that replicates parts of that. Being able to study these things and ask questions that seem like they don't have answers but then take them seriously as though they do have answers, I feel like it feels like a big privilege. Treating these philosophical questions as rigorous, scientific, concrete questions that you can actually make progress on. Yeah, like a lot of people get a few late nights in college to like think about these
Starting point is 00:36:42 things and then you go and have a real job where you don't get to think about it again. Yeah, that's my whole real job and that's wonderful. Ellie, thanks so much for joining us. It's been a real pleasure. It's a pleasure. What a charming take on this, that she gets to think about what she wanted to think about as a college student. I think a lot of scientists feel this way, that it's a privilege to be able to really spend our time
Starting point is 00:37:07 doing what we want to do. Our hobby is our job. Yeah, and hers seems to me particularly elusive in the science space. It's getting so philosophical, right, that how do you make progress in the same way that you do in science? I mean, philosophy can really spin your wheels
Starting point is 00:37:24 for a very long time. Yeah, that makes me wonder, does philosophy always turn into science, just a matter of time? It used to be a question, how is life different from non-life? But after Watson and Crick, it started to really look like it's going to boil down to molecules and atoms. And Bertrand Russell, of course, famous British philosopher, also turned to science in many ways. I mean, he was trying to write a kind of mathematical principia, right? Logic, science, we're involved with things that we're setting up.
Starting point is 00:37:54 What Turing did, what Cantor did, what Gödel did. I don't know, it's an interesting question. You can send all your mail to Steve. But seriously, let's just ask, what are going to be the longest holdouts? For instance, most people would say values are not something that can be quantified. But I'm not even sure about that because with morality being studied nowadays through evolution of cooperation from a biological perspective, I'm not even sure that values are outside of science. I guess I'm espousing what the critics call scientism, that it's all just science at the
Starting point is 00:38:30 bottom. And that's a big naughty thing to do, isn't it? Okay, just thinking out loud here. I feel like you're lost in thought and I need to give you some space to ponder the process. Always great talking to you. Can't wait to see you again. This is fun. Yeah.
Starting point is 00:38:47 See you next time. Still have questions about AI's impact? Wondering how researchers devise experiments or how mathematicians think about proofs? Head to quantummagazine.org slash AI for a special series that looks beyond prosaic AI-based research tools to explore how AI is changing, what it means to do science, and what it means to be a scientist. Thanks for listening.
Starting point is 00:39:20 If you're enjoying the joy of why and you're not already subscribed, hit the subscribe or follow button where you're listening. You can also leave a review for the show. It helps people find this podcast. Find articles, newsletters, videos, and more at quantamagazine.org. The Joy of Why is a podcast from Quanta Magazine, an editorially independent publication supported by the Simons Foundation. Funding supported by the Simons Foundation. Funding decisions by the Simons Foundation have no influence on the selection of topics, guests, or other editorial decisions in this podcast or in Quanta Magazine. The Joy of Why is produced by PRX Productions.
Starting point is 00:39:58 The production team is Caitlin Folds, Livia Brock, Genevieve Sponsor, and Merritt Jacob. The executive producer of PRX Productions is Jocelyn Gonzalez. Edwin Ochoa is our project manager. From Quanta Magazine, Simon France and Samir Patel provide editorial guidance with support from Matt Carlstrom, Samuel Velasco, Simone Barr, and Michael Caniogolo. Samir Patel is Quanta's editor-in-chief.
Starting point is 00:40:27 Our theme music is from APM Music. The episode art is by Peter Greenwood, and our logo is by Jackie King and Christina Armitage. Special thanks to the Columbia Journalism School and the Cornell Broadcast Studios. I'm your host, Jana Levin. If you have any questions or comments for us, please email us at quanta at simonsfoundation.org.
Starting point is 00:40:51 Thanks for listening. from PRX.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.