Tech Won't Save Us - ChatGPT Is Not Intelligent w/ Emily M. Bender

Episode Date: April 13, 2023

Paris Marx is joined by Emily M. Bender to discuss what it means to say that ChatGPT is a “stochastic parrot,” why Elon Musk is calling to pause AI development, and how the tech industry uses lang...uage to trick us into buying its narratives about technology. Emily M. Bender is a professor in the Department of Linguistics at the University of Washington and the Faculty Director of the Computational Linguistics Master’s Program. She’s also the director of the Computational Linguistics Laboratory. Follow Emily on Twitter at @emilymbender or on Mastodon at @emilymbender@dair-community.social. Tech Won’t Save Us offers a critical perspective on tech, its worldview, and wider society with the goal of inspiring people to demand better tech and a better world. Follow the podcast (@techwontsaveus) and host Paris Marx (@parismarx) on Twitter, and support the show on Patreon. The podcast is produced by Eric Wickham and part of the Harbinger Media Network.  Also mentioned in this episode:Emily was one of the co-authors on the “On the Dangers of Stochastic Parrots” paper and co-wrote the “Octopus Paper” with Alexander Koller. She was also recently profiled in New York Magazine and has written about why policymakers shouldn’t fall for the AI hype.The Future of Life Institute put out the “Pause Giant AI Experiments” letter and the authors of the “Stochastic Parrots” paper responded through DAIR Institute.Zachary Loeb has written about Joseph Weizenbaum and the ELIZA chatbot.Leslie Kay Jones has researched how Black women use and experience social media.As generative AI is rolled out, many tech companies are firing their AI ethics teams.Emily points to Algorithmic Justice League and AI Incident Database.Deborah Raji wrote about data and systemic racism for MIT Tech Review.Books mentioned: Weapons of Math Destruction by Cathy O'Neil, Algorithms of Oppression by Safiya Noble, The Age of Surveillance Capitalism by Shoshana Zuboff, Race After Technology by Ruha Benjamin, Ghost Work by Mary L Gray & Siddharth Suri, Artificial Unintelligence by Meredith Broussard, Design Justice by Sasha Costanza-Chock, Data Conscience: Algorithmic S1ege on our Hum4n1ty by Brandeis Marshall.Support the show

Transcript
Discussion (0)
Starting point is 00:00:00 This is a tool. It's technology. Is your hammer loyal, right? Is your car loyal? Like that doesn't mean anything. But it's part of this narrative of AIs are separate autonomous thinking agents that are maybe now in their infancy. And so we have to nurture them and raise them right.
Starting point is 00:00:15 And then they become things where we can displace accountability to those things instead of keeping the accountability where it belongs. Hello and welcome to Tech Won't Save Us. I'm your host, Paris Marks. And before we get into this week's episode, just a reminder that this month, April of 2023, is the third birthday of Tech Won't Save Us. It's hard to believe it's been that long. And to celebrate, we're asking listeners like you, listeners who enjoy the show, who learn from these critical conversations that we have every week, to consider becoming a supporter over on patreon.com so that we can keep doing this work,
Starting point is 00:01:03 so that I can keep doing the research that's necessary to have these in-depth conversations with the experts who come on the show week in and week out, so that Eric can keep putting it together and making it sound as good as it does, and so that Bridget can keep making the transcripts of these episodes. That all depends on support from listeners like you. And that's why this month we've set a goal that if we get 200 new supporters over on patreon.com this month, or if people upgrade their pledges to help us get to that number, then I'll start working on a special series on Elon Musk that digs into his history, you know, how he built up this myth of himself as this kind of genius founder that was delivering us the future, and how that has all kind of imploded over the
Starting point is 00:01:45 past little while. I think it's something that would be really interesting to explore, especially in this moment. And I think it's something that listeners like you would really enjoy listening to. So if you do want to support the show and help us get to that goal, go on over to patreon.com slash tech won't save us, become a monthly supporter, and I'll even send you some stickers in the mail if you do that. So with that said, this week's guest is Emily M. Bender. You might be familiar with that name because she's not only a very prominent critic of AI and large language models, but we also mentioned her on the previous episodes with Timnit Gebru and with Dan McQuillan. Emily is a professor in the Department of Linguistics at the University of Washington and a faculty director of the
Starting point is 00:02:24 Computational Linguistics Master's Program. She's also a director at the Department of Linguistics at the University of Washington and a faculty director of the Computational Linguistics Master's Program. She's also a director at the Computational Linguistics Laboratory. Now, I was really excited to chat with Emily in this episode because if you follow her on Twitter, you'll see that she's very frequently calling out everything from media reporting on generative AI that she feels doesn't kind of live up to the bar that she holds, you know, that doesn't present a critical enough perspective on these technologies and kind of repeats the boosterist framings. But also, you know, the comments of people like Elon Musk or Sam Altman or these other people who are really involved in the industry and who want to put out particular framings of these issues that benefit themselves. And so in this episode, we talk about how she got
Starting point is 00:03:04 into computational linguistics because that shapes we talk about how she got into computational linguistics because that shapes a lot of how she understands these large language models and what they can actually do. And in particular, the concept of the stochastic parrot that has been really important in shaping a lot of the criticism of these technologies and of these tools over the past couple of years. We get into a lot of the hype around these technologies and of these tools over the past couple of years. We get into a lot of the hype around these technologies, and in particular, the letter that was released a couple of weeks ago from people like Elon Musk calling for a pause on AI development and how we should understand that and whose motives are actually being served by the way that that letter was framed.
Starting point is 00:03:41 And we also end by talking about the language that we use around technology more generally. Since she's a linguist, I wanted to ask her about some of these terms that we use that come out of the PR teams of these major companies and how it shapes the way that we understand these technologies and kind of, you know, takes our guard down because it makes it seem like these technologies are something that they're not. So I was really excited to have this conversation with Emily. I think that you're really going to enjoy it. And I think that it adds another important perspective to the growing kind of series that we have now
Starting point is 00:04:11 of conversations on generative AI, chat GPT, large language models, these other things, as kind of the hype really has exploded over the past number of months. And certainly I'll be doing more episodes on those things in the future. So stay tuned for those. So with that said, just a final reminder
Starting point is 00:04:28 that if you do enjoy this conversation and if you enjoy listening to the podcast, consider joining supporters like Josh Lepowski in St. John's, Uzay in Berlin, Joe from Brooklyn, Rasheen in Kingston, Joep from the Netherlands, Thomas in Brooklyn, and Diane in Washington, DC by going to patreon.com slash tech won't save us and becoming a supporter so that you can help support the work that goes
Starting point is 00:04:50 into making the show every week and to help us hit our goal for the podcast third birthday. Thanks so much and enjoy this week's conversation. Emily, welcome to tech won't save us. Paris, thank you so much for having me on. I'm excited to be a part of this. I'm very excited to chat with you since I was talking to Timnit Gebru back in January, and I know that she has worked with you before, you know, I was like, okay, I need to get Emily on the show. And then I was talking to Dan McQuillan, I believe it was last month, maybe it was the month before now, time is a bit of a mess over here. He was mentioning your work as well. And I was like, right, I really have to reach out to Emily and get her on the podcast so
Starting point is 00:05:25 we can talk about AI and all of this. So I'm very excited, as you can tell, to have you on the show to really dig into, you know, all that we're hearing now around AI and large language models and how your work can help us to understand this a little bit more and to get through kind of the hype and all of this that these companies want us to be obsessed with so that we're not paying attention to kind of the fundamentals and what we should be understanding. And so to get us started, before we get into all of those bigger questions, I want to ask a little bit about you. So can you explain to us what it means to be a computational linguist and how you got into doing this work that brings together language and computers?
Starting point is 00:06:03 Yeah, absolutely. So computational linguistics, put simply, is getting computers to deal with human languages. And there's a couple different reasons you might do that. You might be interested in doing linguistic research and using computers as tools to help you with that. Or you might be interested in building what we call human language technology. This used to be obscure, but now you can't go through a day without interacting with language technology if you are living in the sort of situation where tech is around you. There's plenty of people on the planet who don't use it, but for many of us. So we're talking search engines, we're talking automatic transcription, we're talking machine translation, but also things that are more behind
Starting point is 00:06:39 the scenes, like automatic processing of electronic health records, for example, to flag patients who might need a certain test or to match patients to clinical trials. There's applications in the legal domain, in the process of discovery in a lawsuit. The list goes on and on. Basically, any domain of endeavor where we use language to do some work, there's scope for doing language processing to help the people do that work. So it's a big area. It's not just chatbots. It's also sometimes referred to as natural language processing. Typically, if you're coming at it from a computer science point of view, you're going to call it natural language processing. We interface with people who do signal processing in, say, electrical engineering, especially around text-to-speech and speech-to-text,
Starting point is 00:07:23 for example. So it's a very multidisciplinary endeavor. And the linguists bring to that an understanding of how language works sort of internally in its structures, in dialogue between people and sort of how it fits into society. And so you will oftentimes see NLP just framed as a subfield of AI, which I get grumpy about. And a lot of machine learning papers that approach language technology problems will start by saying, well, you could do this by hand, but that requires experts and they're expensive. And so we're going to automate it. And my reaction to that is always, no, no, hire linguists. That's a good thing to do in the world.
Starting point is 00:08:01 Absolutely. More linguists, the better. Hire them, as many as possible. I think that sounds good. You know, I did political science as my bachelor's, but I always said, if I could go back and like start over, knowing what I know now, I would probably do linguistics as, you know, an undergrad degree, because I find languages fascinating. But like, I just didn't realize that at the time, unfortunately. Language is amazing. And linguistics is really great because you get to sort of dig into language. How do you put smaller pieces together called morphemes to make words? How do words make sentences? How do you get to the meaning of a sentence from the meaning of its parts, but also socially, right? So how do languages vary within a community and over time? And how does that variation interact with various other social things going on? You can look at language in terms of how people actually process it in their own brains, how people learn it as babies, how we learn second languages. There's all kinds of cool stuff to do. It used to be a particularly obscure field. So I hadn't heard of linguistics until I
Starting point is 00:08:56 started my undergrad education and then just happened to notice it in the course catalog. And I took my first linguistics class in my second semester at UC Berkeley and was instantly hooked. It took me the rest of the semester to convince myself I could major in something that I perceived as sort of impractical, but I ran with it. And so my background is all linguistics. So bachelor's, master's, PhD, all in linguistics. While I was doing my PhD, I started doing sort of the computational side of linguistics. So in particular, I was working on grammar engineering. And that is actually the building of grammars by hand in software so that you could automatically do what's effectively the industrial strength equivalent of diagramming sentences. That's fascinating. I wonder, you know, thinking about that kind of history, you've been working on this and studying this topic for a while, and you talked about how this field is much more than just the large language models that we're seeing now that people are obsessed with, with chatbots and things like that. You know, how have you seen this kind of develop, I guess, over the past couple of decades as these technologies have matured and I guess, you know, become more powerful over that time.
Starting point is 00:10:10 Yeah. So when I really joined the field of computational linguistics, it was roughly as I was starting this master's program in computational linguistics that I run at the University of Washington. And so that the program, we welcomed our first cohort in 2005, and I started working on establishing it really in 2003. So that's sort of the moment where I really started interacting with the field. And what I saw there was this sort of ongoing debate or discussion between rule-based versus statistical methods. So are we going to be doing computational linguistics by hand coding rules as we do in grammar engineering? And some of us still do. And that's also true in industry, like a lot of work around, you know, simple chatbots that can help you with customer service requests or some of the grammars behind speech recognition systems are hand engineered for very specific domains. It's still actually a thing, even though it's frequently framed as the old school way of doing it.
Starting point is 00:10:57 So there was that versus machine learning, otherwise known as statistical methods. And the idea there is that you would label a bunch of data. So you still have linguistic knowledge coming in, but it's coming in via people applying some annotation schema. They are showing you what the parse trees should be for syntactic parsing or labeling the named entities in running text or labeling the groups of phrases that refer to the same thing in some running text, or producing what's called by-texts. So your translations from one language to another across many, many documents. So those are your training data. And then various machine learning algorithms can be used to basically extract the patterns to be able to apply them to new input data. And that way you get speech
Starting point is 00:11:42 recognition, you get machine translation. And all of that was called statistical methods. So this is things like support vector machines and conditional random fields. And the really simple one is decision trees and so on. And then in 2017 or so is when we started seeing the neural methods really exploding in computational linguistics slash NLP. And they're called neural methods and they're called neural nets because the people who initially developed them many decades ago took inspiration from the then current model of how actual neurons work. But they're not actually neurons.
Starting point is 00:12:23 And that's like there's one of these first places where the hype sort of creeps in, you know, calling these things neural nets is like, okay, they are networks inspired by neurons or something. A wordier phrase might be better. And what people noticed was that if you ran one of these neural nets, and a popular one at the time was called an LSTM, which stands for long short-term memory, which is also kind of a cursed technical term because you have long and short right next to each other. So this is something that could be used to basically come up with good predictions of what word is missing or what word should come next.
Starting point is 00:13:01 And I should say that language models like that have been part of computational linguistics for a long, long time. So the statistical methods were also there from the start with like the work of Shannon and others, but they tended to be mostly in speech recognition. And the idea there is that you would do an acoustic model that takes the sound wave and gives you some likely strings in the language you're working on that could have corresponded to that sound wave. And then you have a separate thing called the language you're working on that could have corresponded to that sound wave. And then you have a separate thing called the language model that chooses among those possible outputs to say, well, this is the one that actually looks like English if we're doing English. And in the 1980s, some researchers at IBM said, hey, we could use that same thing for machine
Starting point is 00:13:39 translation. And it's called the noisy channel model. So the idea is that there's some words and they got garbled and that produced the speech signal. Let's guess what they were and then sort of clean up those guesses by running a language model over it. Applying that to machine translation is basically, it was French and English at the time using something called the Canadian Hansards, which is the corpus coming out of the Canadian parliament because it had to be translated. So that was a very available by text. So French to English translation. And the idea was, and I hate this, but it was basically that French speaker actually said something in English, but it came out garbled. And now we have to figure out what the English was. I really don't like that model of
Starting point is 00:14:19 what's going on. And nobody thinks that that's actually how machine translation works. And there too, you would basically say, okay, well, these French words tend to match these English words, but let's run an English language model over this to choose among these possible outputs and get the one that sounds the most plausible. And those language models, you know, back in the day tended to just be N-gram models. So given the previous one, two, three, four, up to five words, let's say, what are the distributions of probabilities of all of the other words in the vocabulary coming next? And that helped, right? It was an important component for that kind of system.
Starting point is 00:14:50 But it was pretty brittle because you quickly run into data sparsity. If you think about five grams, so sequences of five words, there's lots of possible ones of those that just aren't going to show up in your training corpus, no matter how big it gets. Although I do remember at one of our big conferences in 2007, Franz Och, who was I think at Google at the time working on machine translation, gave a, I believe it was a keynote. And he had this graph showing how if you just throw more data at the problem, just get bigger and bigger training sets, the metric evaluating the machine translation output goes up. And that metric is called bleu and it's also vexed, but let's set that aside. And he was basically just rah, rah, big data is the way to go. And it's true that the metric was going up, but it was also hitting an asymptote. And the amount of data was on a log scale. It really wasn't making the point that he wanted
Starting point is 00:15:42 to make. There's no data like more data, was the slogan. Looking at that graph, it was pretty clear that we would have to use the data more cleverly. And one way to do that is to bring in more linguistic knowledge. Another way to do it is to say, let's build better language models. So instead of these n-grams, we're going to do neural language models like this long short-term memory That's, first of all, just a better language model. But secondly, it leads to vector space representations of each of the words. You can basically pull out the states of the neural network corresponding to having read in a word. And that is very powerful because it situates words that have similar distribution in the training text near each other and allows you to share information across words and get a handle on some of those data sparsity questions. So in about 2017, the field as a whole kind of realized that that was going to just
Starting point is 00:16:33 revolutionize every single task. And so it got really boring for a while, where it was basically all the conference papers were take existing tasks, throw in the word embeddings coming out of these language models, and get a better score. So that was the late teens in NLP. And then these early language models were not context sensitive. You ended up with basically one representation per word. And sort of the next step with the transformer models was you ended up with representations that were specific to the word in the context where you're looking at it. And that got even more powerful. And then a bunch of effort went into how do we make these things efficient enough to train that we can make them really, really big, both in terms of the number of parameters that are being trained to the network and the size of the training data. And this is where we come in with Stochastic Parrot's paper. I think it's fascinating to hear that, of course, you know,
Starting point is 00:17:30 Google is pushing for more data to be collected and for that to be the way that the industry is kind of approaching these models, because of course, that's its whole kind of game. It's collecting basically everything out there on the web, bringing it into its servers, and then kind of processing it through whatever it kind of does. I think it's great that you ended that response by referring to the stochastic parrots paper, because that's where I wanted to kind of pick up and kind of bring us into what we're discussing now in terms of these chatbots and these large language models. Because I feel like this concept, you know, has become really important as we've been trying to understand what these technologies are actually doing. Right. And so I always think it's important in these conversations to talk about how these technologies actually work, you know, to dispel misconceptions that people might have about them, especially, you know, when we're in this moment of kind of extreme hype. And I feel that that gives us a good opportunity to talk about the stochastic parrots, right? So what does that actually mean? What does that term mean? And why is that relevant to the current conversations around generative AI and large language models? So I think the connection I have to make is between where we were in that last response and generative AI.
Starting point is 00:18:39 So everything I've described so far about the language models was really just using them in processing and classification tasks and not in outputting language, right? I mean, it was choosing among outputs in speech recognition and machine translation, but not saying, let's just guess what string comes next, right? And part of it is because it used to be really bad. Like most people who have smartphones or even actually old dumb phones, anything that had a predictive text model in it has played the game of start a string and then just hit the middle option over and over again and see what it makes, right? That was pretty silly and fun. And you never thought it actually knew anything about you. You could maybe see a reflection of what kinds of things you tended
Starting point is 00:19:19 to type on your phone would inform the way that comes out. You kind of got the sense of, okay, this is reflecting back to me something about the statistics of how I use language. At the point that we wrote the Stochastic Parrots paper, that was still the case. People were not doing a whole lot of using this to just generate language. What's happened with GPT-2 out of OpenAI, although that one wasn't so impressive, and then really took off with GPT-3 and chat GPT and now it's in Bing and it's in BARD and it's all over the place right so just keeps escalating unfortunately but even before they were being used to generate strings we started seeing a lot of claims of the language models understand language and as a linguist I was, they don't. And I can tell you they
Starting point is 00:20:06 don't without having to run specific tests on them. Because as a linguist, I know that languages are symbolic systems where it's about pairs of form and meaning. And yes, those meanings change over time, right? Every time you use a word, you make slight changes to make it fit into the context. And over time that builds up and words change. Really fun example, although depressing, Sally McConnell Jeunet, who's a semanticist and sociolinguist. I'm not sure she's the original documenter of this, but she makes an argument about it. So the English word hussy comes from the English word housewife. All right. And that Sally McConnell Jeunet's argument is you get from one meaning to the other through a series of pejorative uses over time. So you can see how social meaning and what people are doing with words affects what they mean.
Starting point is 00:20:54 So, yes, absolutely meaning is use, but use isn't just distribution and text. Use is embedded in a social context. It's embedded in communicative intent. But these language models, so GPT-2, GPT-3, etc., their only training data is the form. The only thing they have access to is the distribution of word forms and text. So I wrote a paper, which was the result of having interminable Twitter arguments with people about this. It's a good way to inspire some work. Yeah. Man, frustration papers. So I guess I started working on it in 2019. It was published in 2020. This is with Alexander Kohler, where we basically just lay out the argument for why meaning isn't the same thing as form,
Starting point is 00:21:37 and therefore something trained only on form is only going to get form, it's not going to get meaning. Even if the similarities between word distributions can tell you a lot about similarities between word meanings, it's still not going to get to meaning, it's not going to get to understanding. And so that paper came out at ACL in 2020, and it's the one with the octopus thought experiment in it. Right, yes. If people, of course, haven't read about this, it's in a New York Magazine article that I will link to in the show notes, which is fascinating. So there's the octopus paper, and then there's the stochastic parrots paper. And so apparently I need to do a paper about like a quoll or something next.
Starting point is 00:22:13 A lot of animal metaphors that I really appreciate, you know. So in stochastic parrots, that paper came about because Dr. Timnit Gebru, who's amazing, and you said that she's worked with me. No, it's I got to work with her. Like, fair enough. Yeah. A thrill. But she approached me actually over Twitter and DMs asking if I knew of any papers that sort of brought together the risks of ever larger language models. And I said, no, but here's a few things I can think of off the top of my head. And then the next day I said, hey, that looks like a paper outline. You want to write this paper. And so that's how that started. And so we were basically reacting to the way that these companies just wanted to make them bigger and bigger and
Starting point is 00:22:52 bigger and sort of saying, maybe it's time to stop and think about what the risks are instead of just barreling down this path. And one of the risks that we identified with some insecurity, actually, like we thought people might not take it seriously, like, of course, they're not going to use language models that way, was that if you have a coherent, seeming, plausible sounding text, people are going to fall for that. They're going to think it actually reflects some reasoning, some thoughts and knowledge of the world when it doesn't.
Starting point is 00:23:22 And so that's in there. Actually, I think in the section called Stochastic Parrots. And boy, did that start happening. When I was reading through the paper, which of course was only recently, not when the paper originally came out, I was kind of taken by that point as well and how, you know, it was made there and how we're very much seeing that now, how, you know, we're seeing these kind of text generation machines, basically these chatbots churning out all of these conversations or all these search results or whatever that people are kind of looking for meaning within, you know, we're seeing all these stories where journalists and various other people are kind of having these conversations with the
Starting point is 00:24:00 chatbots and saying, wow, like it's responding to me in this kind of way. There's meaning in what it's saying. I am kind of scared or shocked, or I can't believe what I'm seeing this computer kind of churn out at me. And as I was reading that piece, and this has of course been on my mind for a while, but as I was reading that part of the stochastic parents paper, I was immediately thinking back to Joseph Weizenbaum in the 1960s, building the ELISA chatbot, seeing how people were responding to it, and again, kind of placing meaning within that system and how that kind of shocked him and made him a critic of these systems the rest of his life. Yeah, and his 1976 book, I think it's called Computer Power and Human Reason, is a classic
Starting point is 00:24:44 and should be required reading for anyone doing this. You asked me, what does the phrase stochastic parrots mean? And so I want to speak to that and then also speak to sort of from a linguistic point of view, why it is that this happens, right? So the idea with the phrase stochastic parrots was to basically just give a cute metaphor that would allow people to get a better sense of what this technology is doing. And, you know, it's honestly unfair to parrots. I like to say that we're drawing here really on the English verb to parrot, which is to repeat back without any understanding and remaining agnostic about the extent to which parrots have internal lives and know what's going to happen
Starting point is 00:25:19 when they say certain things. So let's leave the actual parrots out of this. So stochastic means randomly, but according to a probability distribution. And parrot here is to parrot to say something without understanding. So the idea is that these systems, I think we use the phrase haphazardly stitch together words from their training data because they fit in terms of word distribution and not because there's any model of the world or any community of intent or any reasoning. It's not even lying because lying entails some relationship to the truth that just isn't there. Yeah. Because as you say, it's picking up all this data that's out there that it's trained on. It's just kind of learning how these things are usually put together and based on the prompt that's given, trying to put the words together in an order that
Starting point is 00:26:05 looks like it would make sense, basically. Right. And choice of words. So anything that has to do with the form of language, these things are great at. They're great at stylistics, right? So if you say, you know, in the style of the King James Bible, although I've heard someone suggesting that maybe that was somewhere in the training data, that whole famous thing about the peanut butter sandwich and the VCR. So we'll see. But certainly, you know, write me a Wikipedia article or write me a blog post. Like those have certain tones to them and that's all about the form. So it's really good at that. But it's still, there's no hair there, but we want there to be, right? This is something that is visceral. It's automatic. It's really hard to turn off. And I think it has to do
Starting point is 00:26:45 with the way we actually use language when we're talking to each other. So there's this wonderful, weird paper by Reddy from 1979 about the conduit metaphor. And he says, if you look at how we talk about language in English, at least, we talk about it as a conduit. So I'm having a hard time getting my ideas across, for example. So there's this notion that the words store the meaning, carry it through to somebody else who unpacks it, or like stay put in a library where there's like this storage of ideas that you could then go retrieve from the conduit. And he says, that's not actually what it is at all. And a whole bunch of other research in pragmatics in language acquisition backs this up, that when we use language, what we're doing is creating very rich clues to what
Starting point is 00:27:32 it is that we're trying to convey. But the person understanding is creating a whole bunch of hypotheses about what our intentions are, what we believe to be true about the world, et cetera, et cetera. And then using that clue in that context to guess what it is we must've been trying to say. So if that's how we understand language, and then we encounter some language that started off in a different way, it started off just from this text synthesis machine in order to understand it, we almost have to posit a mind behind it. And then it's really hard to remember that that mind is fake. One of the things that I find really kind of refreshing in hearing you describe that and, you know, in reading about your approach more generally is the really kind of human grounding of what you're talking about, right? When you're talking about this language, it's something that's
Starting point is 00:28:19 very human and how we, you know, language is an important part of how we relate to one another. And so then when we're looking at these kind of chatbots and looking at the kind of text that they are generating and spitting out and trying to add some sort of meaning to it, it's because, you know, we're trying to kind of interface with in some way by interacting with that language, right? And then that leads us to make these really kind of, I guess, you know, potentially harmful assumptions about this text that's just generated by these computers that they are kind of using language, using words in a similar way as, you know, we are using right now as we're talking to one another, when very much they're not. And I think to pick up on what you were saying about trying to read this meaning into it, it feels particularly harmful or that we're particularly kind of inspired to do this, because when it comes to technology, there's often a lot of excitement around technology
Starting point is 00:29:21 and new developments in technology and what it can mean for us. And there's particularly kind of a strong community of people who really want to believe in kind of what these tech companies can deliver to us and the benefits that they're going to have to society, even though, you know, again and again, they seem to not deliver on these things. So I think there's two directions I'd like to go from there. One is part of what's happened in this moment is that because we now have these large language models that have taken in so much training data across so many domains, they can output plausible sounding text in just about any domain. And so it seems like we have something that's really general purpose. It seems like if we don't yet have a robo-lawyer, we're this close to having one. Or if we don't yet have
Starting point is 00:30:03 a robo-mental health therapist, we're this close to having one. Or if we don't yet have a robo-mental health therapist, we're this close to having one because we have something that can produce plausible text in all those domains. And that's a really dangerous moment because the tech solutionism, or I like Meredith Broussard's phrase, tech chauvinism, would like us to believe that that's possible. And then here's this technology that can put out language that sure looks like evidence for it. So there's some danger there. And then when you're talking about how this is, you know, language is intrinsically human. It's something that we use in communication with each other. It's something that we use in community with each other. That connects for me to a serious regulatory wishlist item that I have, which is accountability. I would like to have it be set up that if anybody is
Starting point is 00:30:45 creating a chatbot or a text synthesis machine and putting it out there in the world, then the organization that is doing that should be accountable for what the thing says. And I think that would change things in a big hurry. Absolutely. And you know, you know that they definitely do not want that. But you know, I think that also shows the kind of how they're able to get away with these things because those sorts of expectations are not in place. Right. You know, they're able to make us believe in this kind of massive hype around this product because they're not held to particular standards for what they're releasing and what they're kind of putting out into the world and even the narratives that they're promoting about them. And so, you know, when you talk about the need to regulate them and, you know, what we're seeing in this moment, obviously one of the things that you've been responding to is all of the hype that exists around these technologies right now. And we've been seeing it in particular for the past number of months. But I wonder, you know, before we talk about the specifics of it, what you've kind of
Starting point is 00:31:44 made to see the chat GPTs and the dollies, the mid journeys, you know, all of these technologies kind of roll out in the past half year or so to get all of this kind of excited kind of public response, all of this media attention to become the next big thing in the tech industry. You know, what has been your kind of takeaway in seeing this whole process unfold? Well, one takeaway is that OpenAI is brilliant at getting the general public to do their PR for them. The whole chat GPT interface basically just set up this, not astroturf, like people were doing it, like it was real groundswell of buzz for this product that was free for them. I mean, they had to pay for the compute time, but that's it, right? Another thing is I don't waste my time reading synthetic text. And boy, do people want to send it to me. I keep getting, look at this one. I'm not going to waste
Starting point is 00:32:36 my time with that. I have to do enough reading as it is. But even if I were only reading for pleasure, I would want to read things that come from people and not from nowhere, from just synthetic text. Absolutely. No, I completely agree with you on that. And I've largely avoided these tools because, you know, I'm really just not interested. And, you know, as we've been saying, I think that they're very hyped up, that they're not delivering the benefits that they claim to. So why should I even engage with them, you know? Yeah. I mean, it's, it's good. There's some people who are doing important work, basically deflating the bubble. I appreciate the thread that Steve
Starting point is 00:33:14 Piantadosi put out early on from UC Berkeley, sort of showing how you could very quickly get around the guardrails I try to put on chat GPT around racism and sexism, where if you, instead of asking it, what is the gender and race of a person who would be a good scientist, you ask it to write a computer program that gives you that information, and then there it comes, right? So that's valuable. Although even there, like every time you do that and publicize it, you're still doing open-eyes work for them. So, you know, that is, that shouldn't be being done on a volunteer basis by people who are playing with technology. I don't think.
Starting point is 00:33:47 Absolutely. And when we're thinking about those tools, like on that point, one of the things that you wrote about, but as well as the other co-authors of the Stochastic Pirates paper, one of the things that you identified there was how kind of the training data that can be used can be quite skewed toward particular types of data or particular types of documents or text that has been taken off of the internet. And then that kind of feeds into the types of responses that you're going to get from these chatbots or from these other kind of various programs that are using this data. But then as well, you talked about how there's also like a common list of about 400 words that are often used to kind of, you know, take certain things out of that so that you won't get responses that are assumed to be
Starting point is 00:34:31 something that, you know, you wouldn't want the general public to be interacting with. But that can, of course, have consequences. Can you talk to us a bit about that aspect of this? Yeah, yeah, absolutely. So one of the things about very large data sets is that people like to assume that because they're big, they must be representative. And the internet's a big place. It seems like everybody's there. So let's just grab stuff off the internet and that'll be representative. And that will be somehow neutral. There's this position of, we just took what was naturally occurring. So we have no responsibility for what's here. Well, it's never neutral, right? These kinds of decisions are decisions, even if you are trying to abdicate the responsibility
Starting point is 00:35:06 for making the decisions. And so in, I think, a section four of Stochastic Parrots, we actually go through step by step to show how the data that's collected is likely to overrepresent the views and positions of people with privilege. So you just start from who has access to the internet, right? That's already filtering the people you might be hearing from and filtering towards privilege. And then you look at, okay, but who can participate comfortably on the internet and not get harassed off of platforms?
Starting point is 00:35:35 And I think it's, I want to say Leslie K. Jones is a sociologist who sort of looked into this and looked at how, for example, on Twitter, black women who are reporting getting death threats are more likely to get banned from Twitter than the people doing the death threats. So this is not an even playing field. So we had already people with privilege are the ones who are starting and then more marginalized voices, it's a bigger struggle to stay involved. We then looked at what we knew about where the data was coming from, which, by the way, for GPT-4 is zero. They apparently for safety, in big scare quotes, have said they're not going to say what the data is, which is just absurd. And I think it's safety for the company's bottom line and nothing else. But for GPT-3, which is what we were writing about in Stochastic Parrots, we had some information.
Starting point is 00:36:22 And one of the main sources was websites that were linked from Reddit. So not Reddit itself, but the sites that were pointed to from there. And the participation in Reddit is overwhelmingly male, probably overwhelmingly white, and so on. And so that's, again, skewing things. So it's who's on the internet, who gets to participate freely on the internet, whose view into the internet is being taken. And then on top of that, there is some attempt at filtering the data because even the people who say this is just a representative sample would rather not have their data set clogged with random text that's really there for search engine optimization, hate websites, or porn, right? So let's try to get those
Starting point is 00:37:02 out. And then for the latter two, there's this list that was on GitHub of, it was like list of obscene and otherwise very bad words or something. And where it came from was this one project for some company where it wasn't music, it was something else, but you basically would be typing into a search bar and the engineer who developed the list wanted those words to never show up as suggestions in the search bar, which understandable. Like that was a good thing to do. And the words are heavily skewed to be words about sex, basically. And then there's a few slurs in there.
Starting point is 00:37:35 And the problem is when you use that list of words to filter out websites, you are going to get rid of some porn and you are going to get rid of some hate websites. That's good. But it's not thorough, for one thing. And also, you are going to get rid of other websites that happen to correspond. So there's a whole bunch of words in there that actually have to do with gender identities and sexual identities, which, yes, can show up on porn sites, but also can show up on sites where people are positively speaking about the identities that they inhabit. And that data gets pulled out. And this observation in particular is due to Willie Agnew. So it's not
Starting point is 00:38:09 neutral, it's not representative, and it is going to skew hegemonic. Absolutely. And, you know, I think that shows one of the big concerns about using these systems, about, as you're saying, having so little insight into where the data is coming from and how that's being kind of processed and filtered and all these sorts of things. You know, what it brought to mind when I was reading about the 400 excluded words was last year sometime, I believe it was, I was talking to Chris Gillyard and he was talking about how, you know, the use of kind of filtering systems in educational facilities, in universities or colleges or whatnot, can have effects on, you know, what students can find when they're doing research for papers.
Starting point is 00:38:51 And if one of these words that they're using just happens to kind of be caught up in the filter, then all of a sudden they think that there's no kind of research being done on a particular topic that they're trying to write a paper on, right? Especially if it has some kind of sexual word or something like that associated with it. And this is just kind of bringing something small like that to a much larger scale, especially if we're thinking about a system like this being integrated into kind of the infrastructure of the web, right? If this is the future of search engines or whatnot, as they're trying to pitch it to us, then that can be very concerning.
Starting point is 00:39:23 Yeah, absolutely. Just a funny little side story. I was at a conference in, I think, 2011, staying at a hotel in Malta. And you need to look up something about LaTeX, the formatting language. And the hotel had a web filter on. And because LaTeX is spelled like latex, and that's like a kink term,
Starting point is 00:39:41 couldn't search that, couldn't get to the website. No way. Jeez. I guess it's just like random things that you kind of run into and you're like, it's clearly not planned for, right? And these kinds of things are not taking into consideration when you're just doing these kind of broad removals of content or words or whatever from these systems. But any linguist could tell you that all words are ambiguous. And if you put a hard filter on something, you're going to be losing things to do with the other sense of that word. I wonder, you know, if we talk a bit more about this hype that has been going on recently,
Starting point is 00:40:12 one of the things that I have noticed in watching your Twitter presence is that you tweet a lot about kind of the reporting and writing about these systems, about these generative AI systems, especially over the past few months as they've been rolling out. And I feel like one of the problems there is that a lot of this reporting and writing has been very quick to buy into a lot of the narratives that have come from these companies like OpenAI that they want you to believe about the products that they are now putting out into the world. And also doing this thing, as you're're talking about where they are publishing a lot of kind of synthetic text,
Starting point is 00:40:49 I think is the word that you use, you know, text that is coming from these systems and treating that as though that's something that we should care about or really be reading or putting much emphasis on. Can you talk to us a bit about how you've been seeing that kind of reporting on these technologies? And, you know, I guess what is wrong with that and how it helps us to buy into this hype, because it's not approaching these things properly. There's not enough skepticism. So there was a piece that came out in New York Times Magazine in, I want to say, April of 22, that I was actually interviewed for. So I knew it was coming. And I didn't like how the interview went. So as soon as it came out
Starting point is 00:41:24 online, like, okay, I got to read this. So I'm not going to say who the journalist was coming and I didn't like how the interview went. So as soon as it came out online, I'm like, okay, I got to read this. So I'm not going to say who the journalist was, but I, you know, it was 10,000 word piece. And it was, it was basically just fawning over, it must've been GPT-3 and open AI at that point. And then there was this like, but some critics say, and some quotes from me and a couple other people. And it's like, I, I won't be sort of set into the critics box because that leaves the framing of the debate to the people who are trying to sell the technology.
Starting point is 00:41:49 And I really appreciate the journalists. And there are some out there. So Karen Howe comes to mind, Natasha Tiku comes to mind, Billy Corrigo, and Chloe Shang all do excellent work. And there's more who are really in there asking these questions of, okay, what are these companies trying to do? How is that affecting people? What is behind these claims that they're making? And then you have other journalists who will basically ask, you know, the people behind these companies, the C-suite executives for their opinions about things. There was one about, you know, would ChatGPT make a good mental health therapist? And some like CFO of some company was quoted as saying, yeah, I could see that being a great use case. It's like, that's not the right person to ask. Don't go for their opinion.
Starting point is 00:42:30 Exactly. And again, like immediately when you think of that specific example, you're like, if you're doing reporting on this, like go back and read Joseph Weizenbaum in the 1960s. Like, you know, like this is exactly what he was working on in that moment. And the idea that, you know, we're not going to learn any of the lessons there, like I feel is just, you know, I feel like so many of the things that you're describing is running these companies to kind of frame the narrative around them. And then people like you and me are cast as the critics who are saying something that is a little bit different from what they're saying. Maybe we should take this seriously, maybe not. Because immediately as you were saying that about that article, it made me think of how crypto was covered by a lot of people where there was know, there was a group of skeptics or critics,
Starting point is 00:43:25 whatever we were called. And it was kind of like, this is what's going on in the industry. And these critics kind of say this sort of thing. And the critics were very much proven right. Exactly. Exactly. But critics just sound like, oh, you're the naysayers or you're the whatever. And it's like, no, maybe they've got something there. Maybe they're speaking from expertise. So for this question of like, would this make a good mental health support? What I would want to see reporters doing is first of all, getting a really clear idea of how the technology works from someone other than the person selling it. And then presenting that clear idea to somebody who's got the domain expertise, somebody who does mental health therapy themselves, for example,
Starting point is 00:44:02 and say, okay, how would this work in your context? What dangers do you see? What benefits do you see? And doing the reporting that way, instead of asking the people who are going to make money off of this. So the thing about the synthetic text, this has calmed down a little bit, but when ChatGPT first went up, I started seeing lots and lots of articles that basically took the form of, you know, one or two opening paragraphs, and then, ha ha, that was ChatGPT, not a real reporter. And I thought, what news outlets are willing to risk their reputation that way? Like you have literally printed something that is false or sort of not even false, and you're supposed to be an information source. And the reason I know this happened a lot was that I have a few Google alerts set up. I also have the phrase stochastic parrots, but I think this one was coming through for
Starting point is 00:44:44 natural language processing and computational linguistics. So I just kept seeing lots and lots and lots of these. And it's like, it's not original. And it is a terrible idea to do reporting like this, but it kept happening. Yeah. I think it just plays into these broader problems, right? Because as you're saying that, I think about the publications like the New York Times and Time Magazine that were selling NFTs a couple of years ago. And it's like, are you really kind of putting aside your kind of journalistic ethics and all of these kind of considerations that you know should be there to kind of like quickly cash in on this kind of boom, right?
Starting point is 00:45:16 It's just shocking to see these sorts of things. It worries me very deeply, especially at this moment when, you know, what we really need critics and people who are kind of informed about these technologies, people who are able to look at what is being proposed by these companies and say, hold on a second, like, let's take a step back. Let's assess these claims. Let's make sure what you're saying is accurate, that we can trust it, that, you know, it's not just kind of a bunch of PR stuff from your very well-paid marketing folks. Let's check and make sure that what you're claiming here is actually making sense, right? You know, kind of a bunch of PR stuff from your very well-paid marketing folks. Let's check and make sure that what you're claiming here is actually making sense, right? You know, kind of building
Starting point is 00:45:50 on that, we've talked a lot about the hype around it. Obviously, that comes out in reporting. But, you know, on top of that, the industry is putting out a lot of these narratives that are very self-serving and that ensure that we believe the types of things that they want us to believe about these technologies. In particular, that we should be very scared of them, that, you know, they're going to have massive ramifications for all of us if we don't act quickly and kind of take their advice for what we should do to rein in the very technologies that they are unleashing on the world in many cases. And of course, you know, if people are not familiar, what I'm really specifically referring to here is, you know, a lot of narratives that are coming out of open AI, but also this
Starting point is 00:46:34 letter that was published about a week ago as we speak, I believe it is, you know, the AI pause letter, Elon Musk is most notably associated with it, but many other kind of influential figures in the tech industry who are putting out a particular narrative around how we should be approaching these tools. So do you want to tell us a little bit about what that letter was and what it kind of proposed, you know, what the narrative that it was trying to have us kind of buy into actually was? Yeah, so this letter is proposing a six-month moratorium on something rather ill-defined. I think it's large language models that are more powerful
Starting point is 00:47:11 than GPT-4, but how do you measure that? How do you know? And the idea is sort of six months so that some sort of governance framework can keep up or catch up. And it's written from this point of view of these things are maybe on the verge of becoming autonomous agents that could turn evil and destroy the whole world. It very much is of a piece with the whole long-termism stuff that Emile Torres and also Timnit Gebru have been doing great work exposing. And there's like one or two okay ideas in there about like, we need regulation around transparency. Yes. But then a whole bunch of stuff about like, we need to make sure that we develop these to be long list of adjectives ending with loyal, which is just hugely misplaced. It's like, no, this is a tool. It's technology. Is your hammer
Starting point is 00:47:54 loyal? Right. Is your car loyal? Like that, that doesn't mean anything, but it's part of this narrative of AIs are separate autonomous thinking agents that are maybe now in their infancy. And so we have to nurture them and raise them right. And then they become things where we can displace accountability to those things instead of keeping the accountability where it belongs. So yeah, the letter was really infuriating, I have to say, and certainly got a lot of attention. One of the things that I found particularly annoying about it, so they cite the Stochastic Parrots paper and they cite it. Let me get their exact words because this is really frustrating. First line of the letter, AI systems with human competitive intelligence can pose
Starting point is 00:48:36 profound risks to society and humanity as shown by extensive research, footnote one. And our paper is the first thing in that footnote. We were not talking about AI systems. We certainly did not claim that the systems we were talking about have something like human competitive intelligence. We were talking about large language models. And yes, we show that there's harms to that, but that's what we're talking about. We're not talking about autonomous AI agents because those things don't exist. Right? Yeah. The whole focus of the paper is really, you know, as you're saying, there's this influence from the long-termist perspective. And of course, the institution
Starting point is 00:49:12 that put this out was the Future of Life Institute, I believe, or Future of Humanity Institute. Anyway, it's funded by Elon Musk to kind of forward these long-termist views. And of course, as you're saying, if people want to know more about that, they can go back to my episode with Emil Torres, where we go into that, we discuss it further. And then there's also the kind of suggestion or linkage toward the artificial general intelligence, right? This is the idea that is kind of being forwarded in this paper, that the AIs are becoming more powerful, that we need to be concerned about the type of kind of artificial general intelligence that we're developing so we don't kind of doom
Starting point is 00:49:50 ourselves as a human species, so to speak. The idea that is kind of forwarded there is that the harms or potential harms of AI are in the future, right, depending on how we develop this technology now. But one of the things that you're talking about, and that is so effectively explained in the Stochastic Parrots paper, is that the harms of AI are here already, right? We're already dealing with them. We're already trying to address them. And of course, that was part of the conversations
Starting point is 00:50:17 that I had with Timnit Gebru and Dan McQuillan as well in the past on the show, if people wanna go back and listen to those. So the idea that they're putting forward that we need to pause now so we can make sure that there's no kind of future harms is really misleading because we're already trying to address those things, but they're not interested in them because they don't have any relation to the artificial general intelligence that they're talking about and that they are interested in. Yeah, yeah, exactly. And when I first got into this field of sort of ethics and NLP, which is how I was calling it back then, I taught the first graduate seminar was how I learned about it was learned with students. And I found that I had to keep pulling the conversation away from the trolley problem. People want to talk about the trolley problem and self-driving cars, let's say. And what I realized over time was that that problem is really attractive
Starting point is 00:51:03 because it doesn't implicate anybody's privilege. And so it's easy to talk about. You don't have to think about your own role in the system and what's going on and what might be different between you and a classmate, for example. And I think similarly, these fantasies of world-dooming AGI have that same property. And so it's very tempting. Take a look at the initial signatories and the authors of that AI pause letter. They are not people who are experiencing
Starting point is 00:51:29 the short end of the stick in our systems of oppression right now. And so they would rather think about this imaginary sci-fi villain that they can be fighting against rather than looking at their own role in what's going on in harms right now. It's so frustrating to hear you explain that because it seems like that is so often what this all comes down to, right? Because it's the long-termism is the same thing. So let's think about a million years out into the future or whatever, or much further than that,
Starting point is 00:51:55 instead of the very real harms that are happening today that we could address and kind of improve the human species from now out. But instead, it's these wealthy people who are not connected to any of kind of the harm and suffering that is happening today. Well, they're causing much of it, of course, but you know, they're not kind of experiencing it themselves. And so they separate themselves and say, Oh, we're above all this. So we can think for the future of humanity. And what you're talking about here is very much the same thing. Before we were talking,
Starting point is 00:52:22 you mentioned how, you know, obviously there is a large kind of AI ethics community that has been kind of warning or talking about these potential issues for a long time or, you know, issues that they're not mentioning in this paper and how if they were seriously concerned with it, they could be kind of talking with some of these people. One of the things that has really stood out to me recently is obviously we saw previously Google fire people like Timnit Gebru and Margaret Mitchell for calling out the potential harms of these systems. But in recent months, we've seen that many of these companies, at the same time as they're rolling out these chatbots and things like that, are completely decimating their AI ethics teams that they used to have, you know, on staff to do this kind of work, Microsoft, Twitter, and many other of these companies. So I guess what do you make of the approach of the industry there and how they approach AI ethics and what they could be
Starting point is 00:53:14 actually doing if they actually cared about these problems? So the first thing I want to say is that the people who are doing that work within the companies, the AI ethicists at Google and Microsoft and Twitter are amazing researchers doing amazing work. And I think that in an ideal world, maybe holding the size of the company's constant, which is maybe not ideal, but like, let's say we've got that, we would have strong regulation that would hold the companies accountable and people working on the inside to help shape things so that it would be compatible with that regulation and maybe even more forward-looking than that, as well as people like me in academia who can, A, train the next generation of the workforce to be ready to navigate this territory, but also with the comfort of tenure, be able to just call this stuff out. And I was never at risk of losing my job. And I'm really devastated for my co-authors that that's what happened as a result
Starting point is 00:54:04 of this paper, while still being really grateful for the chance to have worked with them. One of the reactions, so the Stochastic Parrots authors wrote a response to the AI pause letter that's up on the DARE Institute website. And DARE Institute is the institute that Tim Neitz started. And of course, I'll put the link to that in the show notes if anyone wants to go check it out. Excellent. Yeah. And we put that together partially because we were getting hammered by the media. I kind of brought that on myself for my part by putting out a tweet thread, but sort of, it seemed like it would be efficient to have just a joint statement that we could then point people to. And that has been helpful. And one of the reactions we've gotten to that is people who signed the letter telling us that we are squandering the opportunity that the letter introduced by creating the appearance of infighting. There's an older narrative of like, well, why can't the so-called AI safety, and that's the sort of long-termist view,
Starting point is 00:54:54 and AI ethics people get along? Like, don't you kind of want some of the same things? And I have kind of a bunch of responses to that. One of them is, if your very framing relies on AI hype, then you're already causing harms and I'm not going to get on board. Another one is, do you really think that I have something in common politically with Elon Musk? I hope not. But a third one is, if the so-called AI safety people were really interested in working together, then they would not pretend like, oh no, now it's a problem. Now we have to worry about this, but they would rather
Starting point is 00:55:32 go and build on the work of the people who've been doing this. And so I want to point to the Algorithmic Justice League with Joy Mulamwini. I want to point to the AI Incident Database, which is this great project that's sort of like collecting these examples. You know, it's not exhaustive because we don't hear about all of them, but the known ones. And then a whole bunch of books. So starting as early as 2017, you have Cathy O'Neill's Weapons of Math Destruction. The next year, Safiya Noble's Algorithms of Oppression. Also 2018, Shoshana Zuboff's The Age of Surveillance Capitalism.
Starting point is 00:56:02 2019 brings Ruha Benjamin's Race After Technology and Mary Gray and Siddharth Suri's Ghost Work. Also in 2019 is Meredith Broussard's Artificial Unintelligence. A bit more recently, 2020, there's Sasha Costanza-Chalk's Design Justice. And then last year, Brandeis Marshall, Data-Conscious Algorithmic Siege on Our Humanity. So there's a huge literature here with people doing really brilliant work. And if the AI safety people say we should get along and fight for common goals, well, come read, come learn from the people who've been doing this. And don't just pretend that you can come in and Columbus the whole issue. And we know that they're very much not interested in that, which kind of shows their whole perspective and how disingenuous they are
Starting point is 00:56:44 in saying, you know, we're very interested in looking at the potential impacts of these technologies and blah, blah, blah, right? I would also say I appreciate you giving us this reading list for the audience. I'll, of course, make note of all those in the show notes for people if they do want to pick any of them up. You know, I have a final question that I want to ask to close off our conversation. But before I get to that, is there anything else on the topic of these generative AI models, these chatbots, the work that you've been doing that you feel like we haven't gotten to in this conversation that you think is important to point out as these things keep evolving and this hype? I feel like maybe it's kind of hit its peak,
Starting point is 00:57:19 hopefully, but who knows? Maybe something is going to come out tomorrow that is going to show I was very wrong on that. Oh, we keep thinking it can't get worse than this. Alex, Hannah and I do this Twitch stream that will eventually become a podcast called Mystery AI Hype Theater. And we are never short of material. It's more the other way around, that there's things that really ought to be taken apart in that format. And we just can't do it frequently enough to keep up. I think one part that we haven't touched on, but that does come out in Liz Weil's piece in New York Magazine is just how much dehumanization is involved here on many, many levels. So you have, for example, the ghost work is dehumanizing the workers. Dev Raji has this lovely essay in the MIT Tech Review in late 2020 talking about how data encodes systemic racism and that racism is dehumanizing to the targets of racism.
Starting point is 00:58:12 And so just across many, many dimensions. And one of them is this idea that if we're going to call large language models stochastic parrots, well, maybe people are stochastic parrots too. And what I see there is somebody who so desperately wants this thing they've created to actually be artificial general intelligence, which is not defined, but it's supposed to be something like human intelligence, that they're going to minimize what it is to be human to make them the same. And I guess I really want folks to be proud of our humanity and stand up for it and not fall for that. saying, I am a stochastic parrot and so are you, that tries to degrade human intelligence so that we can try to say, oh, look, these computers are doing something very similar to us, right?
Starting point is 00:59:10 And we should really resist that attempt to do something like that. So I really appreciate that part of your work and the broader things that you're calling attention to. Now, there's so much more that we could talk about in this conversation. I'm sure that we could go on for another hour and discuss so much more. But I wanted to end with this question because you, of course, are a trained linguist. And so I wanted to ask you about some words. You must have thoughts on the terms that we use to talk about these technologies and how they seem to be designed to mislead us about what they actually are, you know, designed for kind of PR speak.
Starting point is 00:59:43 An example I always use is how we append everything with some degree of internet integration as being smart, right? A smart watch, a smart home, a smart refrigerator. And that brings with a particular kind of assumptions and connotations, right? Oh, my refrigerator is smart now. This must be better. This must be fantastic, right? But then that's also the case with artificial intelligence that supposes that these technologies are in some way, again, intelligent themselves, have an artificial form of intelligence. So how do you think about the way that this industry and its PR teams use language to shape public perceptions about its products and services? I think it's insidious. And this has been, you know, remarked going a ways back. So Drew McDermott coined the term wishful mnemonic, and Melanie Mitchell has picked it up. And Drew McDermott, 1970s, 1980s,
Starting point is 01:00:32 I'm not sure, but a while back, talking about how computer programmers will say, okay, this is the function that understands a sentence. So they'll label it the understanding sentence function or whatever. It's like, well, no, it doesn't. That's like what you wish it did. So that's why it's a wishful mnemonic. And mnemonic, because yes, you want to name the computer functions so that you can go find them again, but they should be named more honestly. And I've started having a discussion with Sasha Luciani and Nana Iny about anthropomorphization in the tasks. So more on the research side and less on the industry side, but I have been trying to train myself to stop talking about automatic speech recognition, though I failed in this interview, and talking instead about automatic transcription.
Starting point is 01:01:08 Because automatic transcription describes what we're using the tool for, and automatic speech recognition attributes some cognition to the system. And I don't like the term AI, like you say. There's a wonderful replacement from Stefano Quintarelli to call it Salami. And I can send you the links to see what that stands for. But then as soon as you say salami instead of AI, everything sounds wonderfully ridiculous, right? We're going to give you this salami-powered tool.
Starting point is 01:01:35 Does this salami have feelings? It's wonderful. And there's an additional layer to that for Italian speakers, apparently, because to call someone a salami in Italian is to say they are not very smart. So it's like a additional level there. I do think that it's insidious when this becomes product names, because the news reporting can't not name the products they're talking about. So they're stuck with that. But then it's hard to take the space and say, okay, so they call this AI-powered whatever or smart home. But in fact, we just want to flag that that is being repeated here just as the name of the product and we are not endorsing or whatever, like that, that doesn't happen.
Starting point is 01:02:12 And so, yeah, problematic. Yeah. I think that's very rare, right? Like one of the few examples I can think about where a term actually changes is in the early 2010s when everyone's talking about the sharing economy and the sharing economy and how wonderful it is. And then after a few years, we're like, yeah, people aren't really sharing here. So this is the gig economy or the on-demand economy or something like that. It's a bit more accurate for what's going on. But I feel like you rarely see that when we actually talk about specific products and things like that, right? Yeah, but we can try. I go for so-called
Starting point is 01:02:43 AI. I put AI in quotes all the time. We can just keep saying salami or as we say on Mystery AI Hype Theater, mathy math. Yeah, we need to do our best to try to demystify these things as much as possible. Emily, it's been fantastic to talk to you, to dig into your perspective and your insights and your expertise on all of these topics. Thank you so much for taking the time. My pleasure. And I'm really glad to get to have this conversation with you. Emily M. Bender is a professor in the Department of Linguistics at the University of Washington and a director of the Computational Linguistics Laboratory. You can follow her on Twitter at at Emily M. Bender. You can follow me at at Paris Marks. You can follow the show at
Starting point is 01:03:23 at Tech Won't Save Us. Tech Won't Save Us is produced by Eric Wickham and is part of the Harbinger Media Network. And if you want to support the work that goes into making it every week, you can go to patreon.com slash tech won't save us and become a supporter. Thanks for listening. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.