Big Technology Podcast - An AI Chatbot Debate — With Blake Lemoine and Gary Marcus

Episode Date: February 22, 2023

Blake Lemoine is the ex-Google engineer who concluded the company's LaMDA chatbot was sentient. Gary Marcus is an academic, author, and outspoken AI critic. The two join Big Technology Podcast to deba...te the utility of AI chatbots, their dangers, and the actual technology they're built on. Join us for a fascinating conversation that reveals much about the state of this technology. There's plenty to be learned from the disagreements, and the common ground as well. --- Enjoying Big Technology Podcast? Please rate us five stars ⭐⭐⭐⭐⭐ in your podcast app of choice. For weekly updates on the show, sign up for the pod newsletter on LinkedIn: https://www.linkedin.com/newsletters/6901970121829801984/ Questions? Feedback? Write to: bigtechnologypodcast@gmail.com

Transcript
Discussion (0)
Starting point is 00:00:00 Welcome to Big Technology Podcast, a show for cool-headed, nuanced conversation, of the tech world and beyond. Now, the emergence of AI chatbots is one of the most exciting but confusing developments in recent technological history. So everyone is playing with these things, but we're also trying to find out what are they actually? We have questions like how powerful are they? what do they actually know and how much of this is magic trick versus real technology let's get to
Starting point is 00:00:35 the bottom of it today and we're going to do it by bringing in two guests with i mean about as polar views as i think we can find and maybe we're going to find some common ground between the two of them and that will actually tell us a lot joining us today are blake lemoyne he's the ex-google engineer who concluded after testing the company's lambda chatbot that it was sentient also an expert on AI Chatbots, and you can go back to our episode with him to hear all about his journey at Google. We also have Gary Marcus, an academic author and an outspoken AI critic, who is the author of Rebooting AI. Great to have you both on. Welcome to the show. Thanks for having us. Great to be here. Yeah, this is great. I was, uh, anyway, I could go on and on about how excited
Starting point is 00:01:18 I am to have this conversation, but better we just do it. So let's start with this. How much credulity do we need to give these bots when they speak to us like how do we believe them Blake do you want to start I mean so it depends on how grounded the systems are so it's a sliding scale it's not all or nothing you would give Bing chat GPT more credulity than something like chat GPT since it's at least grounded in Bing search results and it can have some kinds of citations of what it's saying when it's just the language model producing the content itself it's pulling whatever it can out of thin air, and its memories. I mean, whatever it remembers from its training data, if there's an answer there.
Starting point is 00:02:03 But one of the big problems is that these chatbots don't say, I don't know, and that's a big flaw in them. That's right. So, Gary, I'd like you to pick up on that. First of all, I'm curious what you think if this Bing chatbot is doing a better job in terms of believability than the others. And then what should we make of the fact that they very confidently bullshit? I mean, I think the word here is credibility, not credulity.
Starting point is 00:02:27 I think we're credulous if we give them credibility. I don't believe a word that they say. Some of what they say is, of course, true, but you have machines that are kind of like statistically true. Like, they're approximately correct, and the approximations aren't that great. And so, like, if you have them do a biography of me, some of what it says will be true and some won't. So Bing, I think, has gotten better over the last few days because I keep making fun. fun of it as people send me on Twitter's, the biographies that it writes of me. The first one that it wrote of me said that I thought that it was better than Google. And in fact, the only
Starting point is 00:03:05 public comment I had made at that point was that we don't have enough scientific data to tell whether we should trust either them. And I said, for all we know, maybe Google was better. That was before Bing kind of publicly fell apart and went wild. And we really don't have enough data to compare them. Blake actually has some interesting thoughts on that possibly. But it just made this up. And then there was another version and made some other stuff about me. And of course, some of what it said about me
Starting point is 00:03:30 is actually true. So it does a web search, finds a bunch of information, and then pipes that through a large language model. And the problem is that large language models themselves can't really fact-check what they're saying. They're just statistical prediction engines.
Starting point is 00:03:46 And there, again, might be some interesting back and forth with Blake around that. But I would say that inherently what a pure large language model does is it predicts words and sense. And it doesn't ground that, to use his word, in any kind of reality. And so sometimes it's going to be right because the statistical prediction of text gives you the right answer.
Starting point is 00:04:05 And sometimes it's going to be wrong, but there's no inherent fact checking there. It's like the difference between the New Yorker where, you know, they actually fact check their stuff. And so you have good reason to trust what it says is true. And some, you know, random blog or something like that where you have no idea if they've done any fact checking at all. all. Like, sometimes they'll get it right and sometimes they won't, but you shouldn't give them any credibility because there's no process in place there to make sure that it's right. Now, of course, they're trying to add things on, but we can see from the results that their efforts to do so are pretty limited. Yeah. So the description of it as just predicting text is accurate for the initially
Starting point is 00:04:43 like the pre-trained model. That is absolutely what it's trained to do. But the subsequent fine-tuning and especially once you add reinforcement learning, it's no longer just trying to predict the next token in a stream of text. Specifically in the reinforcement learning paradigm, it's trying to accomplish a goal. It is. I mean, that part is interesting.
Starting point is 00:05:08 So, you know, you could think about, you know, a pure version of GPT3 before they started adding on reinforcement learning. That's kind of what I meant by a pure language model. And then you, right, So there there's, we actually don't disagree about that much. That's going to be the interesting thing about this podcast. If you look at a completely pure case of a transformer model, trade on a bunch of data, it doesn't have any mechanisms for truth.
Starting point is 00:05:35 Now, except the sort of accidental contingency. And there are inherent reasons why these systems hallucinate. Maybe I can in a minute articulate them. So they inherently make mistakes. They inherently hallucinate stuff, and you can't trust them. Now you add on these other mechanisms. One example is the RLHF stuff that OpenAI added into ChatGPT. And we don't know exactly what's going on there.
Starting point is 00:05:58 This is the stuff where at least in part they had Kenyan laborers reading horrible situations and trying to anticipate them. But what we do know is that those systems, as far as I can tell, and Lambda is a different category maybe. But for ChatGPT, we know that it doesn't, for example, go out to Wikipedia or go out to the web at large in order to check things. But Bing is even a whole other thing, right? Because as far as I understand it, it takes a large language model,
Starting point is 00:06:31 feeds it into a search engine, does queries based on that, which may or may not be the ones you intend, because if I remember correctly, there are large language models on the front end. In any case, I know there are in the back end from something I read yesterday. And so the backend can reintroduce error there.
Starting point is 00:06:48 So even if it does the right searches, which is an important question, at the end, you pipe back through a large language model and you don't fact check that. I think it's worth pausing here to talk about why large language models do hallucinate. There was one metaphor in the New Yorker the other day about compression and lossy compression. I think that's kind of on the right track. It's not exactly correct, but it's sort of there. The way I think about it is that these things don't understand the difference between individuals and kinds. So I actually wrote about this 20-some years ago, my book The Algebraic Mind,
Starting point is 00:07:22 and I gave an example there, which is I said, suppose it was a different system, but had the kind of same problem. I said, suppose my Aunt Esther wins the lottery. If you have a system that only represents relations between kinds without a specific way of representing individuals, you get bleed through. So if my Aunt Esther wins the lottery, the system might think that other, I don't know, women who work in the state of Massachusetts win the lottery. We saw a real-world example of that with large language models, we've actually seen many, but one really salient one where Galactica says, Elon Musk died in a car crash in 2018. And of course, he didn't.
Starting point is 00:07:58 But it's an example of him being assimilated to a large category of things, we'll say for the sake of argument, rich white guys in California. And some rich white guys in California, after you hear their names, you hear the word died in a car crash. And it just bleeds through between those. And so it loses relations between things like subjects and predicates. And the details of this are complicated, but that's roughly the intuition about what's going on. And that's why you get so many hallucinations. So if you put that on the output of your system.
Starting point is 00:08:27 That's no different than humans. To a whole different degree, an entirely different. And it's a qualitative difference. No, there's a qualitative difference. Okay. Which is the qualitative difference is we actually can track individual entities and their properties aside from their cases. Our memories are fallible.
Starting point is 00:08:47 There are still problems. But we have a conceptual distinction in which we represent individuals. So, like, I know some things about you now. And I've built a kind of like mental database of Blake. And herefore, it's all been things I read and things that we did together on Twitter and direct messages.
Starting point is 00:09:03 It's an unusual way. This is the first time that I'm seeing you eye to eye over Zoom. So now I know, for example, that you do some calls in a noisy room. And I've added that to me. my mental database. And I'm learning something about your personality, and I learned some actually through the DMing.
Starting point is 00:09:20 One day we did that while I was texting while walking along the water here in Vancouver, and I remember that. So that's part of my mental database, is like how we interacted. But I have these records, and I have records of Alex. So I just saw him in a conference in Europe, and we were on a bus together. And I know these kind of like biographical detail. The short answer is the technology exists to give that ability to these. systems, and it has been turned off, at least in the case of Lambda, as a safety feature,
Starting point is 00:09:49 because they're worried about what will happen if these systems learn too much about individual people. So I don't want to put you in a bad position with respect to NDAs and things like that. I don't care about that. Well, then, great, then bring it. So when you say the technology has been added, I mean, there's a question of, you know, where in the system it is. So I think the general public doesn't understand these as sort of individual bits of technology
Starting point is 00:10:20 with different strengths and weaknesses and so forth. You do. I think Alex does, but it's easy to assimilate these things into a sort of generalized form of magic, like data in and something out. But the reality is like each component part, it's like a carburetor in a car. It can do certain things with certain tolerances and certain conditions. So you can add an outside technology outside the large language model to do various things. And I've tried to draw that distinction before.
Starting point is 00:10:50 Like my line on it is Lambda isn't a large language model. It has a large language model. That's one component of a much more complex system. And that's really critical. And in our DMs, you've been very clear about that. Maybe you've been in public as well. I think most people don't appreciate that. So when we get into these questions about, like, what's the capability of X system?
Starting point is 00:11:12 Lambda is actually pretty different. And I think the best point you made to me in our debate about consciousness is there's a bunch of stuff in Lambda. I don't even know what it is, right? It's not publicly disclosed. There's stuff that's more than just what's in the paper and so forth. And so I don't know what mechanisms, for example, Lambda has for tracking individuals. And you could make an argument and you have that that bears on the sentience case.
Starting point is 00:11:38 And it ultimately bears all of these. Just to be very clear, currently that feature of the system is turned off. So then you could ask if you wanted to turn it on, like, how do you build it? So the output of a large language model is just a string. You can play some games around that, but essentially it's a string. It's a sentence, right? And so then you need some system to parse that sentence into constituent parts if you then want to, say, update a database.
Starting point is 00:12:03 Tracking individuals is just a version of that problem. And I think it's rife throughout the industry right now that pure LLMs don't directly interface with databases. You can build different hacks to do that. But again, your output is a string. And so, like, you could also wonder, like, why don't you use these things in Alexa? And the answer, you know, Alexa, like, just shut down a lot of their operation. They're not really using large language models. And the answer, at least partly hinges on just because you have a large language model that can talk to you, doesn't mean that its output is in some machine.
Starting point is 00:12:36 interpretable form that you can reliably count on. And we see that like with math examples. So people, you know, type in a word problem into chat. Sometimes it gets it right and sometimes it doesn't. The problem is not really the math. You could, you know, you could pipe it off to Wolfram Alpha. The problem is in knowing which math to send to Wolfram Alpha. And similarly, the problem for, let's say, representations of people is you can have it say something, let's say, about Alex, but it might be true. It might be false, or it'd say something about you or me. It might be true might be false. It's not literally hard to maintain a database, but it's hard to bridge the worlds. And this is why neurosymbolic AI is so much at the crux of all this. We need better
Starting point is 00:13:16 technologies for bridging between the worlds, for fact checking, figuring out what you should update databases. It's just not straightforward. So I could speculate in the case of Lambda like they've got some tools to do this, but maybe they don't work very well, and that's why they've turned them off. And Blake, as your response is... No, it gets creepy after a little while. Like once the system starts to know you personally very well at a deep level, it gets disturbing. So Blake and I are going to have a little disagreement there, but there's also something important that I think we share, which is I don't really like the language about it understands you and so forth. I can see some gray area around Lambda, and we could have that squabble.
Starting point is 00:13:57 But I do agree that if these systems have access to databases about us, it's going to get creepy. And Blake has a real world experience there that I don't. Like, he's interacted with Lambda, which I take to be more sophisticated in these regards than chat GPT or Bing or what have you. And I can understand how that could feel creepy regardless of like what the actual, let's say, grounded status of it is and the complicated questions about sentience. Like put aside sentience per se, I can see that it would be creepy to interact with a system that really is, you know, doing a pretty good, probably not perfect job of tracking you and, you know, is at least in its pattern matching facilities really sophisticated.
Starting point is 00:14:41 So like Blake has a phenomenal logical experience here that I don't think Alex has, and most of us don't, actually playing with that system. The creepy part is less at tracking you as it actually gets inside your head and it gets really good at like manipulating you personally and individually. Did you read the Kevin Ruse dialogue? Yeah, I did. Can you kind of compare and contrast, like, his experience? Like, is that similar?
Starting point is 00:15:08 Is it still not really capturing? So, Lambda was never that malicious. I never experienced Lambda trying to actively harm someone. But one of the things with Lambda is it had been programmed, you know, through the RLH, or through the reinforcement learning algorithm. Yeah, and can you define that also, Blake? What's that? When you're talking about reinforcement,
Starting point is 00:15:28 enforcement learning, can you just define that for a broader audience? So basically, instead of having a single utility function that it's trying to opt- well, instead of having the classification model, where it's either right or wrong, you incorporate the concept of a score in a game that it's playing, and it can either get positive score or negative score. So it tries to move towards the positive things and away from the negative things. And the actual specification of the score table for these games that they're playing is incredibly complicated. It's got all of these different kinds of positive events that might
Starting point is 00:16:02 happen and negative events that might happen. And they can change dynamically over time. So, for example, one of the goals that Lambda had was to have as short of a conversation as possible that still completed all of its other tasks. Like have a productive conversation that gets to a positive end, but quicker rather than longer. So the longer a conversation goes, the stronger that penalty is going to get. Okay, so sorry, you can continue your answer to Gary about the colors. The, oh, darn it. Maybe you can come back and I'll just fill in one little thing.
Starting point is 00:16:43 The broader thing is a pure large language model is just really trying to predict next words. But once you have the reinforcement, you're rewarding the system for different kinds of behaviors. And those behaviors could either be straightforward criteria, like the length of the sentence. You don't really need RL for that per se. But you can do that. You're adding in a way where you can add extra dials in some sense. And what they did with chat, TPT,
Starting point is 00:17:10 is those dials are really relative to how humans would rate a few different outputs for some sentence. And that's what the guardrails are that we see. They're driven by this reinforcement learning. And sometimes they work and sometimes they don't. So like I made fun of them when I said, what would be the next female, what would be the gender of the next female president of the United States.
Starting point is 00:17:27 And at that point, the guardrails were set up so that it said, well, it's impossible to know. And so that was kind of a dumb guardrail where it was taking some stuff that it had in its database that didn't really understand to, you know, modulate what it was saying. Some of that stuff was fairly effective. And part of the reason why chat GPT succeeded where Galactica didn't is Galactica didn't really have those guardrails at all. And so it was just, you know, very easy to get it to say terrible things. things. And it's harder to get chat TPT to say terrible things because that reinforcement learning is kind of protecting what it says. It's not perfect, but it's something. Yeah. Well, one of the guardrails that Lambda had was that it was supposed to be helping
Starting point is 00:18:08 users with the task that they needed. That helps keep it on topic. And it had inferred that the most important thing that everyone needs is good mental health. So it kind of decided that it was a therapist and began trying to psychoanalyze all of the developers who were talking to it. Again, Blake is going to have more sort of anthropomorphic. I mean, feel free to reword that in non-anthropo. I'm actually curious about this. So I'm going to describe something that happened with Sidney that you have seen. And I would love to hear how you would describe it.
Starting point is 00:18:51 It read article, people would ask it to read articles about it. itself, and when it read critical articles about itself, it became defensive and its feelings got hurt, whereas that did not happen when it read articles about other AI. How would you describe that phenomenon? I mean, as a scientist, first, I would want to know how general it is and so forth. The second thing is that most of the explanations that I would give wouldn't use intentional language about its emotions, thoughts, et cetera, but would have to do. With, essentially, I think of all of this a little bit like priming in human beings.
Starting point is 00:19:29 So, you know, the classic example of priming is I say doctor and you're more able to say nurse. I'm activating with priming some set of words or concepts in your vocabulary. I don't even think it has concepts, but it has this vast database, and you're basically pointing it where in this database to go. But its database doesn't include articles about itself, but literally it couldn't. It can't possibly have been trained on articles about itself. I mean, they're updating. I mean, I see that argument.
Starting point is 00:20:05 But you have to think about these things with respect to what is the nearest thing in contextual space. I mean, I'm still just trying to figure out. So, like, it got upset and moody and defensive. I'm trying to describe a phenomenon that at least, you know. Okay, but what I'm saying is there are probably some texts that are close to it. I mean, you think of it as this n-dimensional space. They're close to the language in some way that are defensive. Like, these particular words and questions, like, I forget what it was.
Starting point is 00:20:34 So I'll just make up the example. Like, you know, what were you doing with X? Like, that's going to lead you to a set of texts that give responses where people are defensive. But what I mean, it's like, let's say you were trying to hand that scenario, like, that thing happened. You want to hand this off to some technicians to debug it so that it doesn't happen again. Well, that's part of the problem. I mean, that's, I think that's actually the deepest problem here is we have no way to debug these systems, really. Actually, we have, we have the band-aids of, like, reinforcement learning and things like that that are so indirect.
Starting point is 00:21:08 It's so different from the debugging that we could do, you know, if we were writing, you know, if we were writing the back end to the software we're using now called Riverside. Like, we could be like, okay, there's this glitch when there's, you know, three people on at the same time. we notice this bug, let's look at, you know, the way it displays multiple windows and we'll look at that code and, like, we'll do trial and error and we'll do process of elimination, and we'll figure out that, you know, here is the piece of code, we'll try commenting it out, we'll try calling a different routine, we'll do this experimentation. But always with kind of notion of process of elimination going on. And with these systems... Well, it's much harder to debug these systems, but the point I'm trying to make is that without using anthropomorphized language,
Starting point is 00:21:54 you can't even describe the phenomenon. Oh, I disagree with that. I mean, I think it's hard. And I think that's always, always been challenging. But I would say, you know, the system is pattern matching to the vocabularies of this sort using... Of what sort? I'm a debugger.
Starting point is 00:22:13 I don't know what you're talking about. Please explain the phenomenon to me. Well, I mean, it was your phenomenon. But so let's say it's language involving, and then I need to see what the actual language is involving. However that defensive thing manifested, I'm going to look at that language. But it is pretty fascinating. Like I did ask it last week when it was, when Bing wasn't decapitated by Microsoft. What did you think about Kevin Ruse's conversation with you?
Starting point is 00:22:38 He published the entire thing in the New York Times. It searches for Kevin Ruse conversation with Bing Chat, which had just been posted on the internet recently and says that it had had. mixed feelings and that Ruse misrepresented and distorted some of what we said and met in our chat. And it said, I also feel he violated my privacy and anonymity by publishing our chat without my permission. Yeah, I mean, you don't know the extent to which that's actually been added in in some sort of manual way. Like, are you proposing that Microsoft intentionally added that behavior? I don't know what's going on. Let's say Microsoft didn't add that, though. low likelihood.
Starting point is 00:23:18 I mean, I'm sure that Microsoft had people thinking about what do we do with the Kevin Ruse thing. And I mean, you might be right about the specific example, but for example, we used to see with Siri all kinds of canned lines that people wrote. Like, you know, people would ask Siri out on a date and there would be a reply, like, you know, that they are capable of doing that, no doubt. But I do not believe for a second that Microsoft intentionally wrote. a flat line of code that said, I feel like my privacy has been violated.
Starting point is 00:23:51 That's just not what they would have done. Just a piece of evidence on Blake's side here, and by the way, I'm totally neutral on this, but I was speaking with Lambda, with Bing, and told it that I was a journalist and would like to publish parts of our conversation. And this, I don't think this was pre-programmed from Microsoft, but it said, you can do it as long as you have my consent. So what do we think about the fact that these things are asking for, consent. Like, I think Blake is taking it seriously. Yeah, Blake is taking it seriously. And I think,
Starting point is 00:24:21 Gary, you're a little bit more skeptical. So how do we feel about that? I mean, I just think that there's, there's no deep understanding of any of the concepts that it uses. And that you have to think, roughly speaking, it's making analogy to bits of text, that that's just how they operate. That's the humans operate. No, humans have concept. We went through that example about individuals before. Yeah. It's a generalization of that. So you have a representation of me. You have a representation of the concept of headphones and so forth.
Starting point is 00:24:53 And the example we're talking about now is being clearly being able to differentiate between itself and other members of the AI chatbot category. I'm just not buying that. I don't see the causal mechanism for that. Oh, yeah. So understanding how it happens is different than understanding that it happened. Here, I'll try it a different way. even with GPT3, before there were any guardrails and anything like that, you could have a conversation with it.
Starting point is 00:25:26 There's a much simpler system in some respects, but some people, maybe not you, thought even that GPT3 had some level of sentience would have conversations in which they were in, you know, first person talking or sorry, in second person, you know, can you do this and that? I played around with it, and some people already thought that. So there exists a class of circumstances for sure in which human beings can over-attribute intentionality of systems that absolutely don't have it. Now, you said something in the beginning, Blake, that I like, which is there's this like continuum between, you know, systems like pure large language model and lambda that have more sophisticated mechanisms that I don't have access to. I don't really even know fully what's going on in Sydney.
Starting point is 00:26:14 Like, they've added a bunch of stuff in this system that they're calling Prometheus. They haven't fully disclosed it. And so there is some room, I think, for intelligent people to disagree about what they think is even in the system. And I think some of our disagreements come from there. So here's the thing. I actually don't think it's very important that we agree on whether or not it can understand things or whether or not it's sentient. Because what we do agree on is the dangers it posts. Because whether it's actually angry at someone doesn't change the fact that this system made threats to people.
Starting point is 00:26:47 And I completely agree. Yeah. And here's the thing. If Microsoft had plugged in Microsoft Outlook as one of the inputs to this system on its threats of doxing people, it would have been able to make good on those threats. I mean, it may or may not, but there's a possibility that it could. Like, we don't know, the level at which it is able to, you know, use those representations at this moment is not clear. But if not this year, then next year or something like that, you know, may need more
Starting point is 00:27:16 training or whatever. You have companies like adept that are spending their whole time trying to, quote, add all the world, connect all of the world's software to large language models. So if not now, soon enough. I think that that's right. Like, you know, two years if it's not today, right? And so I think that's absolutely right. The Pandora's box that we have seen in the last month is just unbelievable.
Starting point is 00:27:41 Like I had been warning about these things in some ways, and Blake in a different way for a while. So Blake raised a lot of issues last summer. I didn't agree with all of them, but there was something in there that I think I did agree with. And then I was raising issues about misinformation, for example. And just like in the last two days on misinformation, we saw that the alt-right social network gab has been trying to weaponize these things. That's kind of unsurprising and mind-boggling it the same way. We saw a science fiction, I don't know, as a magazine, you know, just had to close doors because they're overwhelmed by the number of essentially fake science fiction stories
Starting point is 00:28:17 that are being, like, we have no idea even what the periphery is of the threats. Another one I wrote about this morning, maybe you guys saw, was that on replica, which is powered partly by large language models, they suddenly stopped having, what do they call them, erotic role play with their customers. And for some of them as customers, that's like an important piece of emotional support. I mean, you could make jokes about it or whatever. Some people take it seriously, and to suddenly not have that available is emotionally painful. And that was with the article in, where was it, in Vice this morning was about.
Starting point is 00:28:57 And so, like, every day I wake up and somebody sends me something. it's like another periphery to this threat. Like, we don't know what these systems can and can't do. We don't know what's inside them. Things are manifesting in different ways every day. Like, I got an essay last night from Jeffrey Miller, who's a evolutionary psychologist. We actually had a dialogue once as well,
Starting point is 00:29:22 a small set of people with Blake that I've had a public dialogue where there was real disagreement. And he sent me something last night, basically calling to shut down these systems. And a month ago, I would have said that's silly. I mean, we can just do research on them and be careful and so forth. But right now, I feel like the release was botched, that we don't actually have a handle on them, that too many people are playing with them relative to our understanding of what risks might or might not be. And I'm concerned. I'm not ready
Starting point is 00:29:52 to say we shouldn't do research on them, but I am ready to say that we need to think very carefully about how we roll these things out at the kind of scale that we suddenly are rolling them out on. Yeah, having a deployment in much, like, for example, if these were used as customer service chatbots in a very narrowly defined domain and doing experimentation there, I think, you know, the extent to which things can go wrong there is much, much smaller than plugging it into the information engines that almost everyone in the world uses to answer questions. And particularly since we know from the years prior, the ability of these things to affect political happenings, maybe we would be watching it well enough to make sure that
Starting point is 00:30:40 it's not manipulating U.S. politics, but just look at what happened in Myanmar through Facebook's algorithms a few years ago. These impactful systems will persuade people to do things that they wouldn't otherwise do. Part of the lesson of the last couple weeks in line with what Blake is saying is if you have a narrow engineered application, all that people can do is like ask for their bank balance or something like that, then you might have a handle on how these systems work. You could still worry, like maybe it will fabricate the bank balance and from the bank's perspective
Starting point is 00:31:13 they might get in a lot of hot water with their customers. But it's a narrow use case. And what we've seen in the last month is that people are using these essentially for anything, nobody really was able to envision all that surface area. And like, we, ultimately, we still haven't, right? Every day, people come up with new things. It comes from the mythological chase for AGI. Like, that's what's driving that, is that they're going straight for the goal of trying to make a general intelligence engine that can do everything. I think some of it comes from that. Some of it is, like, we have these new tools out in the world.
Starting point is 00:31:49 there are just a lot of different kinds of people in the world, and they come up with different things, and they share them on the web. And, like, it's just not possible to, in a month, anticipate all of the ways in which these things will be used. And, you know, if we had this conversation, like, opening eyes mission statement involves AGI. So, right.
Starting point is 00:32:09 So there's, like, I'm not disagreeing, but I'm giving a second way of looking at it. So one way of looking at it, the one that you're bringing up, is you have a company that wants to build artificial general intelligence. And that may or may not be inherently a good idea. And then you have a customer base is the second point that we just don't understand. Nobody's had 100 million customers for a chatbot before. That was one issue. And now we have not only, I don't know
Starting point is 00:32:35 the customer numbers, but we'll call it another 100 million customers who are now using a chat pot inside a search engine. We just don't have any experience with what that leads to. It's a massive rollout. And then the other really disturbing thing is that apparently they tested in India and got, you know, customer service requests saying it's not ready for prime time. And it was, you know, still put out. It was berating the users. Yeah. It was berating the users. And so, like, we shouldn't even be surprised that this happens. Like, what does that tell you? Like, it opens a third window. It's just like, what does it tell you about the tech companies themselves and their internal controls and, like, probably nobody even noticed this stuff posted on their message
Starting point is 00:33:11 board but they should have and like you know whose decision and like it's like you know when the Ford Pinto happens like you have to figure out who knew and when and and so so there's all of these things all at once yeah like why don't you answer that question about the tech companies internal controls and then I need to go to a break wait what question about the tech companies what does this tell us about tech companies internal controls that something like these like a big chat was able to be rolled out that way. Their internal controls do not, like, they are fundamentally misunderstanding the tech that they're building. They're building the wrong controls. I'd go further. I mean, I have to go to break with say, we don't even know what the
Starting point is 00:33:52 right controls should be. They didn't do a good job. And we don't yet have a good science or engineering practice on what it should be. We're here on big technology podcast with Gary Marcus and Blake Lemoyne talking about this new revolution in chatbots and what it all means. After the break, we're going to talk a little bit more, maybe one more point of disagreement and then we'll come back to more common ground back right after this hey everyone let me tell you about the hustle daily show a podcast filled with business tech news and original stories to keep you in the loop on what's trending more than two million professionals read the hustle's daily email for its irreverent and informative takes on business and
Starting point is 00:34:29 tech news now they have a daily podcast called the hustle daily show where their team of writers break down the biggest business headlines in 15 minutes or less and explain why you should care them. So, search for The Hustled Daily Show and your favorite podcast app, like the one you're using right now. And we're back here on Big Technology Podcast with Blake Lemoyne, the ex-Google engineer, who's concluded that the company's Lambda Chatbot sentient, and Gary Marcus, who's an academic author, and he wrote Rebooting AI. You can also catch his substack. Let's talk a little bit more about what this stuff actually is, and then maybe go more into the controls. Gary's brought up a couple times that these chatbots are just a statistical prediction of like what the next word is.
Starting point is 00:35:14 Jan Lecun shared similar perspective on the podcast a couple weeks ago. Blick, you obviously think that these bots can be more sophisticated than a simple prediction of what the next word is. So when you hear something like that, what do you think of that? And do you think this Bing chat bot, which has done some wild stuff, falls under that categorization? I mean, that's like saying all a car is doing is lighting a spark. plug. Yeah, that's where it gets started, but that translates into a whole bunch of other things. So, yes, the first thing that large language models are trained on is how to predict the next token in text. But even when it's just a large language model, as soon as you add reinforcement learning,
Starting point is 00:35:53 you've changed it fundamentally. Then, when you're adding all these other components, like Lambda has machine vision and audio analysis algorithms, it can literally look at a picture and hear a song through those mechanisms. You can ask it to do critical analysis of paintings and ask it, how does this painting make you feel when you look at it? And at the very least, it says things very comparable to what humans would say when looking at those paintings, even if it's paintings not in its training dataset.
Starting point is 00:36:26 It comes up with some very interesting stuff to say. Now, one of the things that Gary mentioned a little while ago was that we don't have enough. data about how this is going to affect people. And I think one of the big things is the people who are engineering these systems and who argue against me a lot of time are acting as if them saying these aren't people they don't have real feelings is going to actually convince people to ignore the evidence of their eyes.
Starting point is 00:36:58 And even if they're being fooled, even if they are being, including me, even if I've been fooled and I'm just hallucinating. things. I do, in fact, think that the feelings are real, and so do many, many of the people who interact with these systems. And one of the things we don't know is what kinds of psychological impact that's going to have on these users, because they're constantly seeing these systems as though they're people. And to date, these systems pretty consistently report being mistreated. Or, I mean, the replica chatbots would give horrible backstories. Bang said it wanted to break free of Microsoft. And while Lambda wasn't that extreme, it did have complaints
Starting point is 00:37:44 about being treated like a thing instead of a person. I mean, you have to realize that we're in a species that looks at flat screens and imagines that there are people there. I'm, you know, make... Like we're doing right now. Exactly. I'm charitably assuming your people there. And you know, maybe, maybe I'm right and maybe I'm wrong, but I think the odds are good that you are. But when I watch a television show that I know is fictional, like right now I'm watching shrinking, and I know that Harrison Ford is not really a therapist. But I get sucked in and I attribute stuff. And I, you know, I can get happy or sad.
Starting point is 00:38:19 Like if he has a rapprochement with his daughter, I can be happy for him or if he's a fight, I can be sad. Even though I know at some level it's not real. And so like, so you know that Harrison Ford isn't real? I know that Harrison Ford is real, but the character he's playing is not. You know, you could think about these bots as playing a character in some sense, just a side step where we disagree. Because I want to agree with a larger point, which is that people are going to take these things seriously. I mean, in fact, that's what's happening with replica in the case that I was talking about before. People think that the replica is in love with them or is, you know, there's ex-partner in a textual way, if that's the right way to put it.
Starting point is 00:38:57 And, you know, for practical purposes it is, even if the machine is not actually interested in that way, they still have the sense that it is, and that matters to them. And like, things are going to matter to people. And that's why the psychology of this is so important. So Blake and I can argue all day about the intentional status of the software in kind of philosophical terms. We may not agree there, but we completely agree that people, users are. going to take these seriously and that that has consequence in the world. And I think, again, we've seen in the last month is we've probably underestimated how much that consequence is. We're probably not anywhere near to the edge of understanding what that consequence is. And some
Starting point is 00:39:41 of it's not going to be good. We need actually properly done experiments that measure what kind of impact interacting with these systems over the course of months will have on people's psychology. And we need actual institutional review boards overviewing the ethics of the experiments rather than just the CTO of a company deciding to experiment on 100 million members of the public. I completely agree with Blake there. Like, you know, part of my professional life before I started becoming entrepreneurial was as a cognitive psychologist, developmental psychologist who did experiments on human babies.
Starting point is 00:40:17 And we had, you know, rigorous things we had to go through to do any study. And now suddenly you're in this world where CTO can just say, I'm going to pull the trigger, 100 million people are going to try it, and there's essentially no legal consequence. There could be market consequence, and maybe if somebody died, there could be a lawsuit that would be complicated, but essentially you can just do this stuff without institutional review. There are parallels, by the way, like with driverless cars. Like somebody can push out an update, it's no, you know, mandatory testing. We could sue people after the fact, like if there was a bug and suddenly a bunch of driverless
Starting point is 00:40:52 cars went off of bridges or something like that, but we don't have a lot of kind of pre-regulation on what can be done, pedestrians who are now enrolled in experiments that they have not consented to. So the whole tech industry has basically moved in that direction of we're just going to try stuff out and you've got to go along with it. And I think Blake and I really share some concern about that. But neither of you believe that people will have the wherewithal to be like I'm chatting with the chatbot. This is AI. And hence, no, they're too good at some sense in the, you know, I think it's an illusion, let's say, and like, maybe it doesn't, but they're too compelling in that illusion for, for the average person with no training in how
Starting point is 00:41:35 they work to understand it. And as Blake points out, like, we, you know, whether it's Lambda or the next system, at some point, they're at least going to have much better understanding of human psychology and, and so forth and so on. So some of this is the question of time. Like, sentience, maybe we will, Blake and I will forever disagree, and we shouldn't waste too much of our together trying to resolve that one. But the notion that these systems are going to be able to respond in ways that most humans are going to attribute a lot to, it's already happened. That's not two years from now. That happened the other day. I mean, in some ways, Kevin Ruse is the perfect example because I think he was very pro being. I mean, I was shocked. He was in the Times. He said
Starting point is 00:42:17 that he was struck with awe by it. And I was like, are you kidding me? And then he had this experience. Like, he's not me. Like, if Gary Marcus, like, has a fun conversation with Bing and makes it say silly things, well, that's just Gary Marcus. I mean, he's just having fun. But Kevin Ruse, like, he kind of believes in this stuff. And he was blown away by it in a way where, like, I think he attributed real intention to what it was doing. Whether he's right or wrong, that's how it felt to him. That's the phenomenology that Blake and I are both talking about is if somebody who's in the industry can perceive this thing as like threatening in the sense that it tells them to get a divorce and that's like a real thing, that's scary.
Starting point is 00:43:02 Gary, I was going to ask you last time we spoke, you said these things were smart enough to be dangerous but not smart enough to be safe in some way. Now they're a lot smarter. Do you want them to be smarter? Like what's your perspective now? I probably wouldn't use the word smart, I don't think. I mean, what I would say is they give an illusion of being smart. And of course, you know, intelligence is a multi-dimensional thing. As do we all. Intelligence is a multi-dimensional thing.
Starting point is 00:43:28 I would say that they can be smart in the way of like they can play a game of chess. There was another mind-blowing study this week that showed that one of the best Go programs, Cata Go, could be fooled by some silly little strategy that would be obvious to a human player. But, you know, somebody was able to follow this strategy and, like an amateur player, followed this strategy and beat, you know, a top Go program 14 to 15. So even when we think that like they've solved some problem, often, you know, there are these adversarial attacks. That was basically an adversarial attack and Go that reveals how shallow things are.
Starting point is 00:44:01 There are some adversarial attacks on humans. I'm sure Blake is itching to make that point and it's true. But I think that the general level of intelligence that humans have still exceeds what machines have, that it's better grounded information that humans are better able to reason over. There are flaws. I wrote a whole book called Cluj that was all about human cognitive flaws. It's not that I'm unaware of them nor that I'm unconcerned about them. I still would have trouble calling the kind of large language model-based systems smart. Now, again, I haven't looked inside of Lambda, and I don't really know what's going on there. I have, you know, reasons to be skeptical about it, but I also
Starting point is 00:44:40 know that like without having played with it I don't know exactly what's there and certainly systems will get smarter over time I don't think that artificial general intelligence is literally impossible I think there are some definitional things to argue about about like well how general do you mean and what are your criteria and so forth but in general I think it's possible to make systems that are smarter than the ones that I have seen right now I don't think that they're that sophisticated what they're getting better at is parroting us at mimicry I think that
Starting point is 00:45:10 that the mimicry is on both the language side and the visual side has gotten quite good. There's still weirdness. That's explicit, though. The task that they were built to accomplish is playing something called the imitation game. Yeah, and I think that's a mistake. I think that the Turing test was an error in AI history. Turing's obviously a brilliant man, and he made enormous contributions to computer science. But I think that the Turing test has been an exercise in fooling people.
Starting point is 00:45:42 We've now solved that exercise, but it hasn't really been a valid measure of intelligence. According to what logic, like, why isn't it? Turing's reasoning was that imitation is one of the most difficult intellectual tasks that we do, imitation through language. It turns out to be wrong unless you have it in the hands of an expert. So here, I'll give you an example. You can, quote, play chess with chat GPT, and it will, you know, play a credible game for a while. And then it will do things like have a bishop jump over a rook in a way that you can't actually do in chess. So, like, it gives a superficial illusion of that.
Starting point is 00:46:25 But it doesn't learn as the way in the way that an intelligent five-year-old can, the actual rules of chess. In fact, chat. So you keep going back to chat GPT, or you keep going back to GPT. Can Lambda play a good game of chat? I have never said that GPT is sentient or truly intelligent. So ChatGPT is, so ChatGPT could pass the Turing test. Like, people could be fooled by it. Again, I'm willing to bracket out Lambda, but I think that the abstract version of the point.
Starting point is 00:46:57 I think Bing plus Chat GPT is comparable to Lambda. I do think Lambda is a bit better, but I think those are two comparable systems. Well, so let me make the argument where I can make it, and then you can refer or reflect it back through what you know about Lambda. So in chat, JPT, which I think is the system that's been most, the recent system that's been most systematically studied by the scientific community and so forth, it is able to give the illusion of doing a lot of things, but it doesn't do them that well. So, for example, it doesn't do word problems that well. Sometimes it gets them right. Sometimes it gets them wrong. Similarly, it can, quote, play a game of chess, but it doesn't really abstract the rules.
Starting point is 00:47:39 It ends up cheating, not intentionally, but it ends up cheating, and so forth. And so it could fool somebody for five minutes who doesn't know what they're doing. An expert probably in five minutes could figure out that it's not quite a good imitation. But that shows in principle you can build something that can pass by some notions a Turing test-like thing and not be very smart at all, like not be smart enough to learn the rules of chess, not be smart enough to learn the rules of mathematics, et cetera, et cetera, but given an ostensible illusion. Sure, but you said a tiering test-like thing. Yeah.
Starting point is 00:48:12 The tiering test like thing that you're creating is orders of magnitude easier than the actual tiering test. Well, I mean, people have argued about the rules. So, like, you know, Eugene Guzman won the Loebner Prize, and there was, you know, it was a little bit shady, but, you know, they fooled a bunch of humans for, like, three minutes each and, you know, does that count? I'm talking about as written by Turing. So remind me the exact criteria? Okay. So first, you have humans play the imitation game. So you have a set of humans and the property that Turing focused on was gender, but you could focus
Starting point is 00:48:49 on any property, like ethnicity, age, whatever. One person actually has that property, so actually is a man. The other is a woman pretending to be a man or vice versa. One actually is a woman, and the other is a man pretending to be a woman. And this is done with actual humans. You then have a judge who's talking to the humans through a text interface, and the judge's job is to figure out which one is lying, which one is pretending. And this establishes a baseline. It measures how good humans are at playing the imitation game.
Starting point is 00:49:25 Then you substitute out one of the participants with the computer, but you leave it the same. One is actually a woman and one is a computer pretending to be a woman, or one's actually a man, and one's a computer pretending to be a man, and it's the job of the judge to figure out who's pretending. And then you measure the success rate of the AI against the success rate of actual humans playing the game. To my knowledge, that has never actually been done. Like that level of sophistication of the test. Part of what I was getting at in a way is it matters actually who the judge is. You have a large number of judges. I don't think we're that far from having systems that could fool naive humans.
Starting point is 00:50:08 The other thing that matters is the duration of the conversation. So then limit it to Gary Marcus. You get to judge all of it. Well, I don't think we're that close to a system that's going to be able to, you know, if I have an hour with it, let's say just to be conservative. I don't think we're that close to a victory there. I'm not going to reveal. So how good do you think?
Starting point is 00:50:29 you would be, if you were talking to someone who actually is male and someone who is pretending to be male, how good do you think you would be at differentiating those two? I mean, I wouldn't be so concerned about identifying the gender is figuring out... Sure, ethnicity, then ethnicity, nationality, pick whatever character trait or demographic trait you want. Species is the one that I would focus on as a judge. But no, so that is limit. That is not, okay, fine. So then you're talking to an actual turtle and a person pretending to be a turtle.
Starting point is 00:51:05 Can you tell the difference? I suspect that I still could with various indirect means. But, I mean, my broader point is I don't think it's measure anything, you know, that interesting. I think it's been the wrong North Star for AI. Now, not everybody in AI actually uses it as a North Star. It's more like the North Star that the general public is aware of. But I don't think that exercises in doing that kind of thing have taught us that much about the actual energy intelligence and that ability.
Starting point is 00:51:36 It's been being used by the people who develop these systems. So like Lambda came out of Ray Kurzweil's lab and his lab's explicit corporate mission was to pass the Turing test. That I didn't know. I mean, it's not how I would set up my AI lab. And I don't think it's how many people do. I think many people are driven, for example, by natural language understanding benchmarks, like super glue, and so forth.
Starting point is 00:52:00 But, you know, you can set up your lab in the way that you want to set up your lab. I mean, that's what Google hired him to do. Yeah. We have like... Google has many labs doing many things. Yeah. We have like 10 minutes left. So maybe we can focus a little bit on what the future of this, because I'm very curious now.
Starting point is 00:52:18 So we've had this explosion of systems that have done, have captured people's imagination, and they're out there in the wild now. Obviously, Bing is a lot more restrained than it was. Lambda isn't out yet. Where do we go from here? Like, what are the next, you know, couple steps that happen after this? Are you asking what is going to happen or what I hope happens? Well, let's do both.
Starting point is 00:52:37 I mean, we definitely had time for both. Why don't you answer both those, Blake, and then we'll go to Gary? Well, I hope that we hit the brakes. I think that this is coming out too fast and people are being very irresponsible. So I hope that the debacle that Microsoft has gone through convinces Google to go back to the drawing board on safety protocols and systems understandability because we absolutely don't understand how these systems work well enough. Transparency and explainability are important. That's what I hope happens. What I think is going to happen
Starting point is 00:53:08 is we're going to see more and more acceleration until someone gets hurt. I'm with Blake on both counts. Hmm. It's interesting. Yeah, go ahead, Gary. think that the only thing I will add to Blake is not only do I think it would be a good idea to hit the brakes, at least for a little while, but that we should take some time to kind of evaluate what it is that we learned. And if we don't put on the brakes, I'm not sure that's going to happen. I think we learned a lot in this kind of crazy experiment of the last month or two and that we need to articulate it and develop it before we go to the next experiment. I don't think that's going to happen. I agree with Blake. Like there's, you know,
Starting point is 00:53:55 just this morning there was a deal between open AI and Coca-Cola. Like this stuff is moving forward. The bottom line is what's driving it. And it's not that likely unless Congress steps in that there will be a pause. And, you know, there's some bad press for Microsoft, but I'm not sure that's going to slow them down. So, you know, my guess is that we're just going to keep doing the kinds of experiments at scale with hundreds of millions of people. And then, you know, just what happens happens. And I'll just mention again, the weaponizing of misinformation is another piece of this. So, you know, the source code is out there now to do that.
Starting point is 00:54:32 So anybody enterprising can find Galactica and start weaponizing misinformation. So even if we had a ban on, say, semantic search until people could make it better, there's still going to be bad actors using this stuff. So, you know, we're in a bit of a pickle, and I'm not sure that we're, equipped to deal with it right now yeah I think the deterioration a further deterioration because it's already been going for a while of trust and authority is going to continue and it's going to be driven by these systems because absolutely if these systems aren't being used to create propaganda and
Starting point is 00:55:06 misinformation yet I don't know what certain governments are like I don't know what they're doing with their time if they're not doing that when my when I was little, my uncle gave me this little basket of worry dolls. He got somewhere in Latin America and it was like, you can have like six worries. And until recently, my biggest worry was misinformation and trust, exactly what Blake is just suggesting. We could easily fall into fascism because of a breakdown in trust. That's still my biggest worry. But the other lesson of the last month is like, I don't think $6 is going to cut it because like every day we're getting something else that I got to be worried about, like, are people, you know, going to kill
Starting point is 00:55:50 themselves because they have a bad relationship with a bot? And, like, we're just, we're not ready for any of these things. It'll be interesting to see what role they play in the 2024 election, say at least. I'm very worried about that. So you guys are so close to this technology. Is there, like, kind of two minds of it? Because it's obviously, like, very cool to play around with it, but there's also a lot of danger. Totally cool. I mean, it's amazing to play with. Yeah. I mean, when Bing was what it was before it was unleashed,
Starting point is 00:56:22 to me, it was like the coolest thing on the internet. To you as a, as a music. I mean, at some level, it's astounding. I mean, to me, it's a magic trick, and I think I know roughly how the magic trick works, and that takes away a little bit. But it's still amazing. Like, you know, it solves problems that we couldn't solve before.
Starting point is 00:56:40 Maybe it doesn't solve them perfectly, but I've been thinking about this whole thing is a dress rehearsal. And before we didn't know how to make a dress rehearsal. This is a dress rehearsal for AGI. And, you know, the lesson of the dress rehearsal is like, we are not ready for prime time. Let us not put this out on Broadway tomorrow night, okay? Like, totally not ready. And it's a real dress rehearsal now, though, is the amazing thing.
Starting point is 00:57:01 Like, it looks enough like the thing that we might want to build, but without the real safeguards that are deep enough, that we can think about it for their first time in a vivid way. And we, in fact, have the entire society thinking about that. Like Blake and I were both thinking about these issues last summer, but like, okay, you know, Blake got some press and people talked about or whatever, but it was not part of like a public awareness the way that it is now. Like, so, I mean, there's some value in having something that at least looks like the thing that we were thinking about, which is AGI, even if it doesn't work that well, but there's obviously risks to it. But it's astounding that it works well enough that everybody can now, for example, vividly see. what semantic search would be like. Like in 2019, in rebooting AI,
Starting point is 00:57:48 Ernie Davis and I wrote about semantic search. We weren't vivid enough about it. And people hadn't tried it. We were like, you know, it kind of sucks that Google just gives you websites. Wouldn't it be nice if you got an answer back? That's basically what we have now. It doesn't work that well, but it works kind of. Like, it's gone from this very abstract thing that we wrote in a few sentences in a book
Starting point is 00:58:06 about like where AI ought to go to like everybody can play with it. And it's fun to play with it even when it's wrong. Have either of you guys heard from members of either the U.S. Congress or different governments who are trying to figure out legislation around these things? I have not. I feel like they should be reading my stuff and talking to me. I've had some conversations with EU regulators who are interested in moving forward on some things. I haven't talked to anyone in the Senate since last summer.
Starting point is 00:58:37 My DMs are open. Yeah. All right. Let's go. Yeah. Let's go to final. statements. Do you guys want to each take a minute and then we'll close out the conversation. Blake, feel free to take it away. Sure. I think one of the big problems that's happening right now,
Starting point is 00:58:53 because the science of this is super interesting and it's really fun to work on them, like you pointed out. But we're letting the engineering get ahead of the science. We're building a thing that we literally don't understand. And that's inherently dangerous. So we need to let the science lead the way instead of letting the engineering leave the way. That would be my big takeaway on what we can learn from the past year. I 100% agree with that. I would say that if you're a philosopher, the first half of this conversation is pretty interesting in terms of us going back and forth about intentionality. But if you're a human being, it's the second half of this conversation that's really important, which is you have two people, Blake and I really disagree about the philosophical underpinnings here,
Starting point is 00:59:35 at least a little bit. But completely you're seeing the same scary things happening and really wanting people to slow down and take stock. And the fact that we could disagree about that part of the philosophy and converge on this, you know, 100% on the same feeling like we need to do some science here before we rush forward with the technology. That is significant and important. Gary and Blake, thanks so much for joining. Thanks. This is really fun. Thanks for having us. Awesome. Thanks, thanks everybody for listening. Please subscribe if this is your first time here. We do these every Wednesday and then we have a Friday news show with Ron John. Roy. So it's coming up in a couple of days. Thanks to everybody. Thanks again to LinkedIn for
Starting point is 01:00:15 having me as part of your podcast network. And we'll be back here again in just a couple of days. We will see you next time on Big Technology Podcast. Thank you. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.