Tech Won't Save Us - Don’t Fall for the AI Hype w/ Timnit Gebru

Episode Date: January 19, 2023

Paris Marx is joined by Timnit Gebru to discuss the misleading framings of artificial intelligence, her experience of getting fired by Google in a very public way, and why we need to avoid getting dis...tracted by all the hype around ChatGPT and AI image tools.Timnit Gebru is the founder and executive director of the Distributed AI Research Institute and former co-lead of the Ethical AI research team at Google. You can follow her on Twitter at @timnitGebru.Tech Won’t Save Us offers a critical perspective on tech, its worldview, and wider society with the goal of inspiring people to demand better tech and a better world. Follow the podcast (@techwontsaveus) and host Paris Marx (@parismarx) on Twitter, and support the show on Patreon.The podcast is produced by Eric Wickham and part of the Harbinger Media Network.Also mentioned in this episode:Please participate in our listener survey this month to give us a better idea of what you think of the show: https://forms.gle/xayiT7DQJn56p62x7Timnit wrote about the exploited labor behind AI tools and how effective altruism is pushing a harmful idea of AI ethics.Karen Hao broke down the details of the paper that got Timnit fired from Google.Emily Tucker wrote an article called “Artifice and Intelligence.”In 2016, ProPublica published an article about technology being used to “predict” future criminals that was biased against black people.In 2015, Google Photos classified black women as “gorillas.” In 2018, it still hadn’t really been fixed.Artists have been protesting AI-generated images that train themselves on their work and threaten their livelihoods.OpenAI used Kenyan workers paid less than $2 an hour to try to make ChatGPT less toxic.Zachary Loeb described ELIZA in his article about Joseph Weizenbaum’s work and legacy.Support the show

Transcript
Discussion (0)
Starting point is 00:00:00 I think at the other end of it, it's kind of the same issues we have with social media companies, the spreading of disinformation, trying to centralize power, moderation, and exploiting people, like a labor force that's sort of second-class citizens around the world. So I think that's kind of the path we're on. Hello and welcome to Tech Won't Save Us. I'm your host, Paris Marks. Just a quick reminder that we are running a listener survey this month, so if you have about five minutes, you can find the link to that in the show notes if you want to fill it out for us. Definitely appreciate if you can do that. Now this week, my guest is Timnit Gebru. Timnit is the founder and executive director of the Distributed AI
Starting point is 00:00:53 Research Institute, or DARE. And she's also the former co-lead of the ethical AI research team at Google before being fired in 2020. And that might be a story that you will remember quite vividly. It won't be news to you that we are in the middle of a hype cycle. You know, there's a lot of excitement right now around particular AI tools like chat GPT or stable diffusion, churning out writing and images based on things that they have scraped from the internet. As a result of that, there are a lot of big promises being made right now around what potential future these AI tools might bring, what our society might look like because of
Starting point is 00:01:32 all of these things that these tools can churn out based on all the data that they have taken from the web. But are these things exaggerated? Are we likely to actually realize a lot of these things that people at OpenAI and other boosters are saying we might? I think that there's reason for us to be skeptical of these things. First of all, because a lot of these promises have been made in the past and not followed through on. But second of all, because it's very beneficial to a lot of these people for us to buy into it and for us to believe that these AI tools are the next big thing after the failure of cryptocurrencies, after the failure of the metaverse. Tim Neat is also throwing some cold
Starting point is 00:02:11 water on these visions of the future being put forward by these AI hype masters. And I think it's a perspective that we should listen to. You know, Tim Neat is obviously much more familiar with all this than I am. She previously worked on large language models, and it was her work on that that ultimately got her fired from Google because the things that she was saying and trying to publish about these models, the reasons that we should be critical of them and thinking more critically about them, are things that Google didn't want us thinking about and didn't want to be brought to the public's attention as these tools could make up a significant piece of the future business model of companies like Google itself. So in this conversation, we talk about all this hype,
Starting point is 00:02:51 we talk about what these AI tools might mean. But of course, we also talk about Tim Neat herself and her own history, because I did want to know a bit more about what happened at Google and how it came to the point where, you know, she was ultimately fired by this company that many people see as, or at least have traditionally seen as, you know, a positive force in the valley or in the tech industry. And obviously that has started to shift in recent years, especially as, you know, the do not be evil slogan was officially kind of taken off of the company. So I hope you enjoy this conversation. I certainly did.
Starting point is 00:03:23 If you like it, make sure to leave a five-star review on Apple Podcasts or Spotify. You can also share it on social media or with any friends or colleagues who you think would learn from it. And if you wanna make sure that we can have more of these critical conversations on AI, which is a topic that I think we'll be returning to
Starting point is 00:03:37 quite a lot this year, as it seems to be the next big thing that this heck industry is trying to sell us, your support is definitely appreciated so that we can do that. And so you can join supporters like Randy Weinstein from New York, Patrick from Chicago, and Ben Finch from Boston by going to patreon.com slash techwontsaveus if you want to do that. And with that said, enjoy this week's conversation. Tim Neat, welcome to Tech Won't Save Us. Thank you for having me. I'm a huge fan of your work, and I am very much looking forward to reading your book.
Starting point is 00:04:09 But given all that's on our plate, it's like really hard to figure out how to do things that we need to be doing, like read books. I completely understand that. I have like a million books on my reading list. You know, I get sent books all the time. And I'm like, where am I going to find the time for all these? So yeah, no pressure. But I have like a million books on my reading list. You know, I get sent books all the time. And I'm like, where am I going to find the time for all these? So yeah, no pressure. But I have bought it.
Starting point is 00:04:29 Yeah, I appreciate that. When you get around to it, though, you know, you can certainly let me know what you thought of it. Definitely looking forward to it. I have obviously been following you and your work for a while as well. You know, I think a lot of people will know you and when you came to their attention. Admittedly, I didn't know your name before you were fired from Google, but at least we got to know you. Many more people did at that time. One upside to that whole situation, I guess. And I'm really excited to talk to you today about, you know, artificial intelligence,
Starting point is 00:05:00 all this hype that has been around it lately, how we should actually be thinking about these technologies and what they might actually mean for us, you know, going forward and into the future, especially as, you know, with the failure or loss of interest, I guess, in kind of Web3 and crypto and metaverse, seems like AI is going to be one of these technologies that the industry is refocusing on and trying to, you know, build this hype around into the future, right? And so, as I said, you know, we haven't talked very much about AI and machine learning and these kind of topics on the show previously. Certainly, they've come up. But I was hoping that you could give us a general understanding of what it
Starting point is 00:05:35 actually means. What is AI or artificial intelligence? What does that refer to? Is machine learning the same thing? Is that different? How should we understand these terms and what they actually mean? So I want to start with a comment that, you know, I got my PhD from a lab that's called the AI lab at Stanford. I have my PhD in electrical engineering, but the lab is in computer science department. And I find myself asking the questions that you're asking right now. So I was to start with that, right?
Starting point is 00:06:07 And so as a researcher, actually, I had never described myself as a quote unquote AI researcher until that term became like the mainstream term. And it was like super hyped up. I remember about, I would say about 10 years ago is when the hype started. And so everything has been rebranded as quote unquote AI. And so I first and foremost think of it as a brand, like as a rebranding, more of a marketing term. And some people, I think there was an article by Emily Tucker from Georgetown Law Center for Privacy and Security. The title is Artifice and Intelligence and kind of talking about how it's a misnomer. We should not call it artificial intelligence.
Starting point is 00:06:52 So I'm just that's just an aside. But so the way I look at it, I look at artificial intelligence like a big tent, a big field with subsets of things inside that field. So I would say my understanding of the field is that you try to create machines or things that can do more than, you know, what's been programmed into them, right? And so this could mean natural language processing, where people are more interested in analyzing text and speech. And so then they can create things like speech to text transcriptions in an automated way or automated machine translation from A to B, from one language to another language, et cetera.
Starting point is 00:07:34 It could be computer vision techniques where, you know, you want to look at what's in an image and kind of analyze that image more so than just taking it. Maybe you can do some statistics. Maybe you can say, you know, there's a chair analyze that image more so than just taking it. Maybe you can do some statistics. Maybe you can say, you know, there's a chair in that image. There's a hat, a person, a house, whatever. And that would be computer vision. And these are all under subsets of, at least at Stanford, they're under subsets of the AI lab.
Starting point is 00:07:57 There's robotics where people are working on robotics. And then machine learning is more of a technique. Machine learning would be a technique that many times people try to use for many of these various things, right? So you could use machine learning techniques in natural language processing, you can use machine learning techniques in, you know, computer vision, etc. So nowadays, because a specific type of a machine learning technique is one that is almost everywhere, people use many of these things interchangeably, right? They use AI, machine learning, deep learning interchangeably.
Starting point is 00:08:34 But it wasn't always like that. Actually, the deep learning people, I remember Jan LeCun, I think it was in 2012 or something, literally wrote an open letter to the academic community, the computer vision community, being like, you are not accepting my papers and you guys don't like us, the deep learning people and whatever. We're going to persist, you know. And now it's like they're the dominant paradigm. It's as if there was no other paradigm. So they're not all the same things, but they're often used interchangeably. Yeah. A really big tent, I guess, of a lot of things that are going on there in a term that we hear more and more often, I guess, in the way that you're describing it there
Starting point is 00:09:09 as more of a marketing term. Do you also feel that it's kind of misleading us and making us believe that these technologies can do things that they can't? Like, I think a bit as well, when we talk about like, smartwatches and smart gadgets and all this kind of stuff, it's a particular term that's created by PR people at these companies, I'm sure, that make us believe that, you know, this is a technology that's really smart, right? It has this term that we obviously think is very important that's associated with it, instead of calling it something else that wouldn't give it these particular connotations. Is the term artificial intelligence making us believe that these
Starting point is 00:09:41 technologies operate in ways that maybe aren't actually reflective of what the term would have us think? Yeah, absolutely. So many times there is a bunch of things going on. I don't only blame the corporations or other entities that might have a vested interest in marketing terms. I also blame researchers who definitely feed into the hype. So first is the naming of areas of study, aspirational naming, right? It doesn't mean that things are the way they are, but it's aspirational. So neural networks is one, right? So maybe for some people, the brain might be an inspiration, but it doesn't mean that it works like the brain. And so some neuroscientists are like, why are you coming to our conference and saying this thing is like the brain? Come on, you know, like, but that's aspirational, right? Computer vision can be full of hype then there is also the conflation of a bunch of things so whereas i look at ai as you know a field of studying with a bunch of stuff set like i told you there are a bunch of people whose interest and whose goal is to build what's called artificial general intelligence, AGI, which I personally don't even know what it is. It seems like a god to me. It's going to be here in a few years, though, don't you know?
Starting point is 00:11:12 Are we already there? I'm not exactly sure. I mean, it's this all-purpose, all-knowing god, where if you build it the right way, you build a god. If you build it the wrong way, you build the devil. So it's also the single existential risk for humanity, but we're building in. I just, I'm lost, you know? So there's also that. So because there's that segment of the population that is the loudest and has the most money right now, they also influenced the discourse around any type of AI discourse, because what they try to do is make it look like everything they're building is kind of that, AGI, kind of has AGI characteristics, right?
Starting point is 00:11:52 That's another kind of hype angle. But then when we go to the corporations and other vested entities, another reason where this hype helps is that if you make people think that you're building something that has its own agency, who knows what it's going to do. And it's so capable of doing this and that people are more focused on the thing because they think it has more agency than you as the builder of the system, the deployer of the system, the entity, the regulation, whatever. Because now responsibility is sort of not your responsibility, but it's like, who knows
Starting point is 00:12:34 how the machine will behave? So everybody sort of starts thinking like that rather than holding relevant entities accountable because it's an artifact that people are building and deploying. So at both ends of the spectrum, you have people building it and deploying it and people, you know, being harmed or benefiting. So when you move the conversation away from that, it makes it easy to evade responsibility and accountability. And then also here we have Hollywood, which really just has not been helping, right? So anytime people hear the term AI, the number one thing that comes to mind is either Terminator or some other, you know, I don't know, whole Hollywood-y thing. And so again,
Starting point is 00:13:18 that's also what's in the public consciousness. So none of these things are helping us in kind of really having a real understanding of what's real and what's hype. Yeah, they're really good points. And, you know, especially the makers of the technologies being able to use this to kind of get the blame or get accountability pushed to some other place, right? Like, oh, we couldn't have known it was going to do this when we developed it and trained it on these particular data sets and what have you, right? I want to keep exploring this thread, but I want to detour a little bit to give us another path into it, because I want to talk a bit about your career as well. You know, I'm interested, obviously, there are many different routes you could have taken into technology, into the tech industry, different things that you could have explored. Why was this area of technology of interest to you when you were doing your studies, your PhD, and then working in the industry? So, you know, it's really all I've ever wanted to be is an engineer, a scientist. I'm interested
Starting point is 00:14:16 in building things. I'm interested in learning about things. I'm just, you know, and that was really my goal for a long time. Since I was little, I was interested in certain subjects, etc. So that's really all I tried to do, right? But then during that kind of just trying to live my life, you have various experiences. You know, I had to leave because of war. Then I came to the States and just, you know, the racism was just like in my face since day one. I went with teachers telling me that I can't take classes, that I would fail if I took the exams these people take and all of that. That experience persisted for a long time. So I was, you know, I had a very, I would say,
Starting point is 00:14:57 you know, a kind of visceral understanding of some of these issues. However, what's really interesting is that I still hadn't really connected the tech, you know, so I would understand how I'm facing sexism in the workplace, right, while building a product or whatever. But I never really had an understanding of how the tech itself was perpetuating these things or how it was being used by powerful entities or the militarization of like the you know these universities like MIT and you know Stanford and Silicon Valley whatever that all came honestly way way way way way way later and so I was chugging along doing my engineering stuff and I was like oh I'm gonna start doing what's called analog circuit design then I did that
Starting point is 00:15:42 then I veered off into something else then I veered off into something else. Then I veered off into something else. And I somehow arrived at computer vision, which is a subset of AI, right? And so I was like, oh, this is super cool. Let's do this. And there was just three things that happened simultaneously. One is that I was, you know, shocked by the lack of Black people in the field. So, you know, graduate school was
Starting point is 00:16:06 even much, much, much more shocking than an undergrad. So in graduate school at Stanford, I realized that they had literally only graduated one Black person ever with a PhD in computer science ever. Now it's two, by the way. Yeah, since the inception of the department ever, right? 50-something. Yeah, exactly. And then you go to these conferences. And I remember before I started Black Nei, I counted like five Black people out of 5,500 people attending these international conferences, right? From all over the world. So that was one, just like dire lack of Black people. Then again, a fluke. Secondly, I started watching Joy Boulamwini's talk. She told me I don't say her name right. And we keep on saying each other's name,
Starting point is 00:16:51 Boulamwini. And she told me she doesn't say my name right. It was hilarious. We were like, just repeating each other's name to each other. But anyways, I saw her talk. And she was talking about how these open source face detection tools wouldn't detect her face unless she puts on a white mask. And they would detect, you know, her friends and other people's faces, but not her face. Right. And so I saw that around again, like around 2015, a very similar time. Like I said, when I wrote about this, just counting five black people out of 5,500 in the AI conferences, that was 2016. And then again, in 2016, I read the ProPublica article on crime recidivism. And so
Starting point is 00:17:35 there was this article that talked about a particular startup that purported to have a model that can tell you the likelihood of someone committing a crime again. And, you know, judges and courts were using this data with other kind of input to deter it, to set bail, to determine, you know, how many years you should go in prison for or whatever. And like, I had no idea that this kind of stuff existed because that's when I put all of these things together. I'm like, I know the people who are building these systems. I go to school with them. I work with them. I go to conferences with them. I know how they think. I know what they say when I talk to them about police brutality. You know, I know. And so these are the people building this. Are you kidding me? And then it's going to be used like just I already could tell,
Starting point is 00:18:25 you know, could imagine the issues, right? And also could see how other people were not really caring about these potential issues. Around the same time, the Google gorillas fiasco happened, where they were classifying people as gorillas, right? Black people as gorillas. So all of these things sort of happened at the same time. And then the final piece was the hype. Basically, you know, I had been going in and out of my PhD for a long time. I tried this thing, dropped out, this other thing, dropped out. And then so when I went back, finally, working on my PhD in computer vision, it was literally at the inflection point before and after of the hype in AI so before it was there was no hype I started my PhD when there was no
Starting point is 00:19:11 hype just in like two years it started exploding and so then around the whole Google guerrillas fiasco thing open AI was announced and oh my god I wrote this piece just for myself because I was just so angry. I was going to submit it as an open letter to somebody. My friend was like, everybody's going to know it's you. So it's not anonymous. Like, so it's just, I kept it to myself. And I was so angry because they were billed. So at that time they were a nonprofit. They changed into a for-profit, but they were a nonprofit and they basically were saying that they were going nonprofit. They changed into a for-profit, but they were a nonprofit and they basically were saying that they were going to save humanity from the dangers of AI. It was going to be an AI safety first company. They were worried that these large corporations who are profit
Starting point is 00:19:58 driven, we don't know, they're not going to give us the share their data and stuff are like controlling AI, which is very important and it's going to give us the share their data and stuff are like controlling AI, which is very important, and it's going to change the world. And it's like these eight white guys, and one white woman and one Asian woman in like Silicon Valley, all with the same expertise in deep learning, plus Peter Thiel and Elon Musk. I was just like, you gotta be kidding me, right? So these are the situations that led to me kind of starting to focus, pivot a little bit and try to like, you know, learn more about the potential harms and things like that. I appreciate you outlining that because I would imagine that a lot of people who are in the tech industry who listen to this podcast or don't have
Starting point is 00:20:41 kind of, I don't know, had an awakening about the politics of AI, you might say, you know, have a similar story. Sure, there might be some different experiences in there that they have experienced personally in seeing these things. But I imagine they went into it not thinking so much about the impacts of these technologies in the world. But then after working on it, after learning a bit more about, you know, what they are working on, they start to realize more about the impacts of these technologies and how they are kind of part of this system that is creating that, right? I want to fast forward a bit, you know, obviously, I can't not ask about it, but your time at Google, when you were the co-lead of the ethical AI research team, as far as I'm aware, you know,
Starting point is 00:21:22 there were a lot of people who recognized that this was a really diverse team that had been put together to really dig into these issues, to provide a critical look at what was going on in the AI space and at Google in particular. And then, of course, we all know the story of your firing in 2020, I believe it was, and, you know, just everything that went down there in the press, you know, what was that whole experience like? And what was it that Google found so objectionable about the work that you were doing to lead it to take such extreme action against you? So those are two years, but it was like so much in those two years. So by the time there was no question who I was at this point, right?
Starting point is 00:22:06 Joanna had published this paper showing the disparities in error rates and face recognition systems among people of different skin tones and genders. And I had, you know, co-founded Black Nei. We had our first workshop. I was very vocal about the issues of racism and sexism in the industry, etc., etc. So I joined Google September 2018. And I was already very, very scared. When I was about to join, I was just nervous. I'm like, what am I, where am I going? Like, I talked to a bunch of women who had a lot of issues of harassment who sat me down and telling me to think twice about it.
Starting point is 00:22:52 And so then Meg Mitchell, who was there at the time, and, you know, Meredith Whitaker was also there at the time I knew both of them. And that was during the whole Maven, you know, kind of protest. And so actually, to me, it was like the fact that there were people like that was a big deal. Because it means, okay, I can, you know, you know, there's dissent anywhere. But like, if you don't look like you have it, it means, you know, it's, it's a, it's bad. It's a bad place. So at least I can have that. And then I was thinking, you know what, Meg, you know, is someone I can really work with. She had started the ethical AI team at the time. And it was just like a couple of people. She asked me if I could co-lead the team with her. And I was like, well, at least I have this one person I could work with. And I joined the team.
Starting point is 00:23:26 And so right off the bat, November, there was the Google walkout. And right off the bat, I was seeing just the systemic issues, sexism, racism. Now, this is kind of not exactly about the work itself, right? But just kind of it impacts the work, the organizational dynamics of who gets to make decisions, who's valued, who's not obviously directly impacts what the output is. Totally. It's the wider context of the company that you were entering into, right? Absolutely. So I already I started making noise immediately. And after a few months, I was certainly not the friend of HR.
Starting point is 00:24:00 You know, like clearly I was not the friend of like the higher ups. Like, you know, it's just, you know, so that that is how it started. Then I really started fighting for our team, our team to have a voice, to be at different tables, decision making tables, to grow our team, to hire more people. And we had to fight for literally like every single person we were trying to hire. Right. We hired social scientists as research scientists for the first time. Dr. Alex Hanna was, she's now director of research out there. She was the first person who we hired as a social scientist to be a research scientist because we were like, you know, you just, you need different perspectives.
Starting point is 00:24:38 Is that what, if that's what we were really trying to do. So our work was, you know, we were primarily a research team, but we also, we had lots of teams coming to us for questions. If they were gathering data, you know, how are they going to annotate it? What are some of the issues? We worked on things like, for example, we had a paper called model cards and this was Meg spearheading it saying that, you know, any model that you put out there has to be accompanied by a set of tests, you know, and kind of think about it with tolerance. Like I did this, you know, as an engineer, like there's no way you would put out any products out there without extensive documentation and testing what is supposed to be used for. What are the standard operating characteristics?
Starting point is 00:25:29 What are the ethical downstream considerations, what was your task in mind when you did it, that kind of stuff, you know? So we were doing all this stuff. A bit of the opposite of the move fast and break things ethos, right? Absolutely. And you know what's really interesting is in hindsight, it's very clear that even something as simple as that. So I published a few papers like that where I didn't really have any kind of opposition from people in corporations or engineers, because I'm saying these are like engineering principles. You know, I'm not saying, you know, politically, you have to do this and that.
Starting point is 00:25:58 But that's super political, because what you're saying is that instead of making $100 in like one week by doing this, whatever thing you're doing, first of all, spend one year instead of one week because do these additional tests, etc. Secondly, like hire additional people to do these things. So I'm saying make less money per thing, put more resources per thing. And sometimes you may not even want to like release this thing. So we had to figure out in our team, we want to get people rewarded, like promoted and all of that, not punished. Right. So how do we say, Hey, we stopped you from releasing five products, you know, because they were harmful. You know, how do we like, how do we talk about it in these
Starting point is 00:26:42 terms? So that was sort of what I was doing. That was my job. And, you know, I was extremely tired, exhausted from all the fighting of, you know, sexism, racism, whatever, trying to, you know, get our team at various tables. But we grew our team, and it was definitely a very diverse team in many, in many different ways. So then came 2020. So at this point, there was the Black Lives Matter protests. And then there was OpenAI's GPT-3 that was released. And at that time, you know, it's not like this was the first language model that was released. But they were the first people who were basically talking about this as like this all-knowing, all-encompassing, just hyping it up, like ridiculous amounts of hype. And that seeped into the entire industry.
Starting point is 00:27:29 So all these researchers everywhere, like chats are like on the chat be like, oh my God, this is so cool, it's so cool. And then the, you know, higher-ups be like, why are we not the biggest? Why don't we have the biggest model? Why don't we? And I'm just like, oh my God, why?
Starting point is 00:27:42 What's the purpose of having the biggest, what kind of pissing contest is this, right? That's how it felt. And then at the same time, a bunch of people at Google were asking our team, when we're thinking about implementing infrastructure for large language models and things like that, what are the ethical considerations we should have? So at some point I was like, look, we got to write a paper or something. So I contacted Emily Bender, who is a linguist. She's written about curating data sets, documenting them, etc. I'm like, voices like hers need to be heard prominent. So I contacted her and said, hey, all these people at Google are talking about large language models. They're asking me questions. Some of them are just want to do bigger and bigger. Like, is there a paper I can send them? And she said, no, but why don't we write one together?
Starting point is 00:28:29 So that's what happened. We wrote one together. I was like, oh, excellent. Oh, we have a little emoji as a title of the paper. That's going to be cool. That was it. It was not a controversial paper, right? So we outlined the risks and harms of large language models. And so she coined the term stochastic parrots. You know, there was like other members of our team, obviously Meg Mitchell, who got fired three months after me. You know, we go through some of the risks that we see. One was environmental racism, because you can use lots of compute power and training and, you know and using these models. And the people who benefit and the people who pay the cost are different, right?
Starting point is 00:29:10 So that's environmental racism. We talk about all sorts of risks of bias and fairness and things like that because these models are trained with vast amounts of texts from the internet. And we know what's on the internet. So we spend a lot of time on that. Unfortunately on the internet. So we spent a lot of time on that. Yeah, we spent a lot of time on that. We talk about how just because you have so much data doesn't mean that you have quote unquote diverse viewpoints. We talk about how there are risks of people interpreting outputs from these models as if it's coming from a person. Of course,
Starting point is 00:29:44 that was on the, you know, like that happened. It was on the news. The other risk that we talked about was the risk of putting all of your resources and research direction in this thing and not and not other things. And so that was it, you know, was under review, whatever. And then randomly, I was told that after it went through all these internal processes that we had to retract the paper or, you know, remove our names. And first it was about retraction. And I was like, well, we have external collaborators.
Starting point is 00:30:12 And I also don't trust what you're going to do if we just retract the paper. It's not like you're discussing improvements or specific issues. So then they said, OK, you can just retract the names of the Google authors. And so even that, I was like, you know, I'm not comfortable with that, doing that without a discussion. I want to know what kind of process was used because I can't be doing research. I'm not in the marketing department, right? I'm in the scientific research department. If it was like PR slash marketing, cool.
Starting point is 00:30:40 Yeah, do whatever you want. If you're having me as a scientist to write peer review papers, now this is different. And so of course, then, you know, long story short, I was basically fired in the middle of my vacation. And then it was a whole, you know, public outcry. And then I had a lot of like harassment campaigns and threats and all of this. And so that's what happened at Google. I appreciate you outlining it in such detail for us, because I think it gives us a really important insight into the work that you were doing, right? And I do want to pick up on some of those pieces of the paper that you were talking about, right? Because these are really important
Starting point is 00:31:19 things that I feel like don't get the attention that they actually deserve, right? And that we should be talking about more when we do talk about these AI tools, these large language models, even beyond that, the environmental costs of this. It can be easy to ignore how these models are based on a lot of compute power, right? Having these massive data centers that require a ton of energy to train these models, to keep them going. You know, there are rumors now that chat GPT, even though it's free for you to go and use it online, that it's costing millions of dollars a day just to kind of keep the servers running, essentially, so that you can do that. But that doesn't get talked about as much.
Starting point is 00:31:59 And then, of course, how it's using a ton of data that's scraped from the internet. We know the severe problems that exist in the data that is actually coming off of the internet. As you were talking about, we know how these data sets can be very biased. This is not like a controversial or new thing that we're just hearing about. This is something that's been around for a long time and has not been effectively addressed. Of course, the opportunity costs in the research there, how there's all this hype around this particular type of AI, this particular type of large language model. So all of the energy, all of the research, all of the money is going to go into building on these
Starting point is 00:32:36 types of things. But what if there's a different type of an AI tool, a different type of a language model that is actually better suited toward the public, right? Toward having public benefits that might come of this, toward the types of things that we would be more concerned with as a people instead of just what's going to excite, you know, these people at the top of the tech industry make money for these large corporations. Those things might not be aligned, but when the hype goes in one direction, it gets distracted from anywhere else. So these are all really important points that come of this paper and the work that you were doing with your co-authors and the other people you were working with. Honestly, even though we were worried about the risks then,
Starting point is 00:33:14 I still did not expect it to explode in the way that it has right now, just in such a short period of time, right? All of the things that you outlined are about, to me, obfuscating what exactly is going on, right? The way we talk about things in the tech world. Oh, you're running it in the clouds. It's not a cloud. It's like there are data centers taking resources like water, people working in these data centers who are being exploited, you know, it's not cloud. So in AI, I just wrote an article recently about this, the exploited labor behind what so-called AI systems.
Starting point is 00:33:58 When people in the hype, when people talk about it, it's like, oh, my God, look at this machine. It's so cool what it can do. Well, look at the armies of exploited laborers that first whose data has been taken. Everybody's data has been taken. I think with the art, it's become much more clear because artists are galvanizing. But all of these systems have data sets that they scrape from the internet and data laborers that do some type of tasks. Some of them supply data sets or data. Some of them label them in crowdsource and micro tasks and all this stuff. And so that's what this whole field is predicated on. And that's when it started to become more popular with the advent of things like Amazon
Starting point is 00:34:42 Mechanical Turk, where you could like really crowdsource things and pay people pennies per little task and things like that. Now there are lots and lots of companies who are raising billions just based on that outsourcing kind of task, similar to content moderation and social media and things like that, right? It's again, the hiding of all of the different things that are going on. And so there's the theft, there's the data, the hidden cost, the hidden data labor, the hiding of all of the different things that are going on. And so there's the theft, there's the data, the hidden cost, the hidden data labor, the, you know, hidden costs of, like you said, you know, environmental costs. So that's actually the one thing I feel like we didn't really cover in our paper is the level of theft. So we talked about data curation, right?
Starting point is 00:35:21 And how you can't really just scrape everything on the internet and assume that something good is going to come out. You have to curate your data sets and document them and really understand them. And if your answer is, oh, it's too big to do that, then that means you just, you shouldn't do it. What other, you know, where else can you do that? Like, can you sell me a food item at a restaurant and be like, eat this food? I don't know what it's made of. Like, you know, there's some sugar. I know that there's some flour, whatever, whatever else is, you know, you're on your
Starting point is 00:35:52 own. Can't do that. Not allowed, right? We haven't understand. But in this industry is the most unregulated industry, right? And, you know, they can proliferate things because of that, right? And the other thing is, you know, look at proliferate things because of that, right? And the other thing is, you know, look at all of the different, I just learned today, you know, I was talking to some
Starting point is 00:36:09 people who know the AI Act in the European Union. There is a proposed legislation, the AI Act that, you know, there is still debating. And so people are trying to have, are lobbying for exceptions for what they call general purpose models. So chat GPT type thing, large language models would be general purpose models, right? Now think about it. OpenAI has been going around billing this thing as an all-knowing godlike thing, right? They're not saying, hey, you can, they haven't gone out of their way to tell you, hey, you can only use this thing in these limited ways, whatever. That's not how they're talking, right? When you see the press, when you see Sam Altman's tweets and all these people, they're like,
Starting point is 00:36:49 oh, this is AGI. However, they would be exempt from any sort of harm if they say, hey, they just have a one-liner saying, don't use this in any high-risk things. But they can go around saying, like, this is a god. But then you have these people who are talking about how they want to use these chatbots in Supreme Court. I don't know if you've seen this. You have these people who are using them in mental health applications, right? Unethically, very clearly. So now you have open AI who can make tons of money because it's not going to be free.
Starting point is 00:37:20 Obviously, it's not going to be free. They're going to start making tons of money from these kinds of people, but they won't be liable, right? But they created the hype. They created the race. They created what's in the public consciousness, you know, hiding all of the costs and hiding what's really happening, telling us they've created some sort of godlike thing. And now the people who are going to try to use it as such are going to be, they're the ones who are going to be liable, right? Even if we have any such are going to be there. There's the ones who are going to be liable. Right. Even if we have any sort of regulation, that's assuming that's the best case scenario. So, you know, I just even with the things we first saw in 2020, I just I still did not foresee it to be this prevalent just in the two years. And it's very worrying to see how much it has exploded.
Starting point is 00:38:04 Right. in the two years. And it's very worrying to see how much it has exploded, right? And how seemingly unprepared, you know, the discourse has been to reckon with what it actually might mean for us, right? You know, as I've been watching, say, chat GPT rollout or GPT-3 or, you know, the kind of art, quote unquote, art AI tools that have gained popularity over the past year or so, I guess, I've been a bit unsure how to think about it, right? Because on one hand, you have this kind of public narrative that is being promoted by these companies, you know, like OpenAI, where these tools are going to absolutely upend everything, right? All artists are going to lose their jobs, all writers are going to lose their jobs because these tools can make writing or images that look like a human has
Starting point is 00:38:52 done it. But that feels to me like something that we've heard about a lot of tech products in the past that have really been unable to live up to these really lofty kind of hype inflated claims that we've heard before. And then on the other hand, I wonder, you know, what might the real impact of this be when we get past the hype fueled hysteria that we're in now? And I'm unsure what to make of that. And I wonder what you think of it. That's a really interesting point because I'm like stuck in a cycle. So I, you know, I'm stuck in a cycle where I'm constantly being like, no, no, don't, no.
Starting point is 00:39:27 Yeah, where I have a hard time thinking past that, right? And I'm actually trying to force myself and my institute to carve out time to also work on our vision for that, like our affirmative vision for a tech future that is not dependent on what they're doing. But if we were to think about, you know, the actual impact of this, I do think centralization of power is a big, so let me give you one example, one example of an actual impact. OpenAI has some, you know, a speech to text model, transcription model, and a data set called like Whisper, right? And then Facebook Meta came up with a model called Galactica saying that, you know, it can write all scientific papers, whatever. But then they came up with another paper, which they called No Language Left Behind, saying that, oh, you know, we have this data set that
Starting point is 00:40:22 contains so many different languages, it's so accurate, whatever, and this and that. And there is one of our fellows. His name is Asma Lashteka, who created a startup called Lassan. It's a machine translation startup specifically for some Ethiopian languages, Ethiopian and Eritrean languages, a few of them. I mean, there's so many, first of all, Ethiopian and Eritrean languages, but just even a few of them. I mean, there's so many, first of all, Ethiopian and Eritrean language, but just even a few of them. So somebody who was thinking about investing in them, I believe, was like, hey, have you seen this paper from Facebook? Your startup is going to die. Basically, it's solved.
Starting point is 00:40:55 It's a solved problem. They say they solved it. Let's say it was true. The people who would make money are not people like him, right? The people who would make money are not people like him right the people who would make money are not people who actually speakers of that language and the people who get the money the workers of the startups and stuff are not those right it'll just be like this one company located in one location but it's not true right it's not true because he was looking at the the data set and it's so embarrassing. Some of the languages that are spoken by literally more than 100 million people, their data set was complete gibberish.
Starting point is 00:41:33 So I think really, whatever the path is, centralization of power is where we're headed. Yeah. The example that you give there seems really similar to what we're seeing. If you say that Facebook is saying that it had this model that was going to work for all the languages, but then if you look into look great. Like, you know, it looks like the AI is creating this incredible image or is writing this thing that makes a ton of sense. But as soon as you start to kind of tinker with it or try some different things that might be a bit less conventional, you know, you can see that hands don't generate properly in the images. And there are other kind of issues which show that the AI, so to speak, doesn't actually know what it's doing. It's just trying to combine these different images and styles in a way that you are looking for and the way that it's been taught happy to spit out complete lies and things like that, as long as you frame your question appropriately.
Starting point is 00:42:47 And that it's really unclear that it will truly be able to replace human writing, as we've been told. And one of the things that really occurred to me, as I've been seeing this over the past number of months, is that we had a whole similar narrative right around 2015 and 16, as you say, like in that moment, there was a narrative that, oh my God, all drivers were going to lose their jobs because self-driving cars were right around the corner. And we can see how that worked out. But at the same time, there were also narratives about how writers were going to lose their jobs because all these things could be automated. And it didn't happen in that moment. And I'm sure that these promises have been made before in the past as well. And so I'm always very skeptical when this comes up, like as we're seeing right now, as we're in the midst of this kind of hype-fueled hysteria, like what is actually
Starting point is 00:43:34 going to come out on the other side of this? And I'm very unconvinced that it's going to be what, you know, Sam Altman wants us to believe. Yeah, he said that the future is going to be unimaginably great because of their chatbots, I guess. I don't know how you create a text-to-image generator and you think you're going to create utopia, but this is the language of the long-termists and effective altruists and stuff, which everybody, some people, I am vindicated because some people used to think I was just obsessed with this niche group of people for some reason. But, you know, I've been around them for a long time and I just sort of used to roll my eyes, you know. But now it's just very much, you know, something that you can't ignore, right?
Starting point is 00:44:16 You know what it really reminds me of? I realized that when you were talking about stable diffusion, it reminds me of social media and the issues we're talking about in terms of social media, right? Because they're having issues of moderation too. So then you'll have issues of the moderators and what they're going to have to see and what they're going to be exposed to and have to moderate and the exploitation of these moderators because their entire business model is predicated on theft and not compensating people for stuff, right? And so if they had to, if they were forced to do these things, they just would decide not to do it because it's not worth it. It's not going to
Starting point is 00:44:55 make them money. So I think at the other end of it, that's what I'm seeing. It's kind of the same issues we have with social media companies, the spreading of disinformation, trying to centralize power, moderation and exploiting people like a labor force that's sort of second class citizens around the world. So I think that's kind of the path we're on. And while promising utopia, you know, which just like, I don't understand how you go from A to B. For me, we can never have a utopia that is, you know, based on the vision of some homogeneous group that, you know, has absolutely no ground in history or they don't understand art even. You know what I'm saying? Like, I'm, you know,
Starting point is 00:45:42 I, you know, used to play piano for a long time and I, you know, I'm saying? Like, I'm, you know, I, you know, used to play piano for a long time. And I, you know, I'm a dancer. And I don't, you know, I'm not a visual artist. My goal is not to watch a robot dance, even if they end up doing technically things that are, you know, whatever technique is just one aspect of it's like having words it's a communication mechanism between humans right it's a human expression mechanism people use art for social movements like people use i just i just you know what i'm saying so these are the people who want to like give us techno utopia and i don't know what it even means like a world where you just sit at your computer all the time. You never interact with other human beings. Everything you do is like with a chatbot and with a robot and with, I mean, like, why do
Starting point is 00:46:31 I want to live in that world? You know, honestly, so even their vision, which won't happen is dystopian. You know, I just don't see how this is utopia. Yeah, it seems fundamentally inhuman, right? Like, I'm going to connect a few threads here based on what you're saying. But, you know, when we think about the work aspect of this and what it might mean for work, one of the things that comes to my mind and that, you know, I reflect on a lot as I think about tech's impact on work and on the work that we do on the jobs that people can expect is that, you know, going
Starting point is 00:47:05 back to the gig economy and even before, you know, we often get promises that tech is going to automate things, right? Or is going to take away the need for human labor in various instances. And what we actually see is that it doesn't take away the need for the labor. It just kind of degrades the job, the expectations that the workers can expect when it comes to performing those sorts of tasks, right? So whether it's the gay economy or the people who are working, you know, to label AI tools and images and things like that, that you're talking about these people who are really poorly paid, who are treated very poorly. You know, we can see the rollout of technologies in like Amazon factories and things like that, how that's used against workers.
Starting point is 00:47:45 They haven't automated the whole factory, but they have instituted algorithmic management of these people. And then, you know, as you're saying, when we think about the future that is being positioned that these sort of AI tools like chat GPT or stable diffusion are supposed to create for us, it seems to be taking away like one of the most what I think many people would imagine would be one of the most human things is our ability to create art, right? To create culture and to try to replace that with tools, with computers. And we would just consume what the computers create rather than what humans themselves make. And if there's ever going to
Starting point is 00:48:19 be an argument for something that probably shouldn't be automated, it's those particular things, right? And then the final piece of this, because you mentioned effective altruism, and of course, its connection to long-termism, is that there seems to be a fundamentally kind of inhuman future being presented there too, right? Where we need to be concerned about the far future and humanity's future and all these sorts of things, but not like the issues that are really affecting people today, because the people who are pushing it, who are leading it, whether it's thinking up the AI models and the futures that they predict for us or promoting long-termism, they are incredibly disconnected from the actual
Starting point is 00:48:55 problems that people face in day-to-day life. And the kind of futures and the kind of solutions that they are imagining just do not deal with that or reflect that at all. Yeah, exactly. Absolutely. And it's kind of, you know, iterations of the same story each decade or each century or whatever it is, you know, even when you look at the rebranding of terms, you know, like big data is basically AI now. Like I think everybody says AI, they don't say big data anymore, right? Like it's just kind of, you know,
Starting point is 00:49:25 every so often you have a rebranding of things that sound cool. And I am someone who got my PhD in the field because I thought, you know, interesting things about it. I would like to focus most of my energy on just imagining like, what are, you know, useful tools, interesting tools you can build. But, you know, we end up, and that's another thing I resent these people for, right? Nobody said, go build us a god or try to build us a god. They decided that this is a thing that not only they should do, but it's like a priority for everyone to do. And so then they do that.
Starting point is 00:50:02 And then the rest of us have to like be stuck cleaning up rather than sort of trying to implement our vision of what useful things we can build without our skills and our little amounts of funding we have. We don't have billions of dollars that they do, right? So then we end up in this constant cleaning up mode, which means that we can never catch up. That's really the worry that I have, right? When I talk to legislators or like, you know, when I talk, for instance, we're still talking about, you know, regulating face recognition, right? We're still talking about the fact that we don't have any regulation for social media platforms, but the EU has some, California just came out with their privacy law. They have moved on to like the
Starting point is 00:50:46 metaverse and synthetic, you know, like stable diffusion type stuff, right? They don't even care about this stuff anymore, right? So I often talk about it as regulating like horse carriages when we're talking when actually it's like cars that they've created, you know, they've already paid off all the legislators like ruined public transportation, created roads and cities and all the infrastructure for cars. And we're still sitting and talking about, well, the horse carriage that you've created is not safe. And how can we regulate it?
Starting point is 00:51:14 Guess what? They've moved on, right? They moved on. And we're already going to be too late when we are. So I think it's really important for us to do both the uncovering, critiquing of the harms, but also investing in the people who can do something different, like alternative visions, because otherwise we're just going to end up trying to clean up and never catch up. Absolutely. It's essential, right? If we ever want to think of a different type of future or a different type of technology than what's on offer by OpenAI or by Elon Musk or by whoever else. I'm wondering, you know, as we're talking about this, of course, you are not at Google any longer. You have your own institute that you have founded and that, you know, some people who you worked with at Google have come to join you there. And
Starting point is 00:52:00 I'm sure some people who weren't at Google as well. You know, what is your focus in trying to do research on artificial intelligence now at this research institute? Like, what's the goal that you're trying to achieve without having to worry about what Google and the higher ups there are going to do? You know, what's the kind of lens that you're approaching this research through? So I think the first lens is interdisciplinarity, not just, you know,
Starting point is 00:52:26 in the research world, like sociologists and computer scientists and engineers, etc., but also like labor organizers, refugee advocates, activists, like people who haven't generally been involved in the shaping of this research, right? So I think that if you're talking about quote-unquote benefiting humanity, which, you know, that's not what I'm talking about. You obviously should start with the people who are already doing work to do that and incorporate their perspective. So our research questions and methodology are very much informed by that. We co-create, you know, together. Like, for example, Adrienne is, you know, working at our institute, she used to be a delivery driver at Amazon. And she just gave a talk at our one year anniversary about like wage theft. And so she was calculating, she was showing like her estimate of wage theft that was by Amazon and they, you know, for each worker, right? And I would have never thought to do that if we didn't have her, right? Or we have Mehron, who's a refugee advocate and also very much a victim of transnational repression. And so
Starting point is 00:53:30 we're doing this kind of project on analyzing the harms of social media platforms. And we have to work on natural language processing tools to do that and all that. And so it's really important to have her perspective. But it's really interesting, too, because it challenges your points of view, because Mehron, for instance, is very much into Bitcoin. And the rest of us are like, very, very skeptical, right? Like, I'm like, you're in the room full of skeptics. But I have to listen. This is someone who single handedly rescued 16,000 refugees from human trafficking, right? And has fought like, you know, against governments and all sorts of people. So I have to listen to her when she's telling me the ways in which she's
Starting point is 00:54:09 been using it to help refugees, the ways in which refugees have been using it, you know, so I have to then listen to her, right? So it's interesting. I think like, even to reimagine a different sort of future, it's very different to have these kinds of perspectives, rather than just like this techno utopic, whatever, you know, from some tech bro in Silicon Valley. So that really is the lens we're taking. And I think that what's really hard for us is basically what I just said to you is that because there's so much so many dumpster fires, we have to fight all the time, it's very difficult to just carve out space to be like, what is our positive vision, imagination for the future? And so we're trying to force ourselves to do that.
Starting point is 00:54:52 Like we have, we're starting this possible future series, which is like a few pieces to show like tech, it's not this deterministic march towards some predestined future, right? So the first piece is kind of talking about that. Second piece is about, like, what would an internet for grandmothers would have looked like? My grandmother did not read or write, did not speak English, was paralyzed for a number of years. So when we're talking about technology, you know, you never think of her as an early adopter or customer or whatever, right?
Starting point is 00:55:23 Yeah. So, you know, I think then speculating and then like then also bringing it back down to earth, it's like, okay, what are the components of such, you know? So I think we're just trying to create space for ourselves to think about these things and executing on them to the extent that we can given our expertise. Yeah, that's fascinating. And it reminds me of a conversation I had about, you know, people in the favelas in Brazil and how they were kind of utilizing technologies to serve their needs and rethinking about how they could be used in ways that
Starting point is 00:55:54 companies that were developing them probably never thought about, right? Yeah, it's so fascinating to think about those futures and those other potential uses and how other people might think about or approach technologies. I love that. And, you know, I'm sure I'll be paying attention to the work that you'll be doing at Dare, you and the team, of course. I'm wondering, as we close off our conversation here, where do you think this is all going in the near future? Certainly, there's been some critical coverage of chat GPT, stable diffusion, these types of technologies. But there's also been a lot of coverage. And I would say the majority of the coverage is really kind of uncritical, buying into the hype, you know, framing these things as though they are creating
Starting point is 00:56:35 text that you would never be able to tell the difference between a computer and a human and all these sorts of things. And it makes me worried to see that. And you talked about the company that was using ChatGPT to turn out responses for a mental health service that they were running. And it immediately made me think of Joseph Weisenbaum's Eliza chatbot that he made all the way back in the 1960s and the worries that he had around people thinking that this computer system, that this chatbot could recognize what they were saying actually had some degree of intelligence when it didn't at all, right? So, you know, that's a long way of asking, where do you see this going in the next few years? And what should we be paying attention to to try to have a more critical take and a more critical view on these tools as they get all this hype and attention. You know, similar to you, I don't know if it's just my imagination, but I have started to see a lot more critical approach to tech in general, much more so than like, let's say 10
Starting point is 00:57:39 years ago. I don't know if Steve Jobs died today, if he would have had the type of, I'm not exactly sure actually, the kind of reverence that he did have had the type of, I'm not exactly sure, actually, the kind of reverence that he did 10 years ago, right? Which was like a god. And when I was at Apple, I felt like I had joined a cult, actually. Like, whoa, my God, you know, the way they were talking about him. Did you wear a black turtleneck? Absolutely not. But it's just, I would say, but on the other hand, it's really depressing because it's the same story over and over again. Like if you look at Theranos and how the media just hyped it up, built it up all over and then like, oh my God, how did it happen? You know, or Elon Musk's like time person of the year, whatever. It's not like we didn't know. The rest of us being like, oh, come on, you know? And then the whole Sam Bankman-free thing, it's just kind of the same cycle over and over again. And with this hype, I mean, I am shocked that ChatGP,
Starting point is 00:58:35 I mean, the level of attention and hype it's getting is truly shocking. You do have to give it to them on the hype machine. They are so good at it. And so whatever kind of critique we have, it is not even comparable to the kind of hype that it is generating. And so to me, that's where I see it. I see the same cycle. If at some point a Theranos type situation happens, you know, it's not the media that's going to help us, right? Like, so it's legacy media. I think it's really not helping us because what they're doing is they're giving megaphones to the super powerful and making it even more difficult to raise awareness about
Starting point is 00:59:16 some of these issues and talk about reality. And so that's kind of what I'm seeing. That's where to me, it's going, you know, you have an explosion of startups. It's not even the big, just the big tech companies right now. Actually, Google is the one who didn't release a chatbot and they have to tell their researchers why not. I mean, they could have done that before they fired us. But like, you know, now and they have an explosion of startups, because this is the next thing that all the VCs in Silicon Valley are talking about and like pouring their money into. And so I just I think we're super behind in kind of really raising awareness about the potential harms.
Starting point is 00:59:55 And I just see like, I honestly see the hype continuing. Brando's talking about having chat, you know, I mean, this guy was saying that having a chat bot as a lawyer would level the playing field because poor people would have access you know to and it's really interesting because this goes into kathy o'neill's point in weapons of mass destruction where poor people are often much more interacting with automated systems than rich people. So now the fact that this guy is saying that automated systems are going to level the playing field for them, it's just obviously the idea did not come from someone who didn't have a good lawyer and therefore obviously it didn't come from a poor person who didn't have access to a lawyer.
Starting point is 01:00:44 So I don't see it slowing down anytime soon. Yeah. No, I share your concerns, right? Especially when we're in this moment when the tech stocks are down, when interest rates are higher, and the industry is clearly looking for the next big thing after Web3 and the blockchain stuff and the metaverse clearly haven't worked out. But what does give me hope is seeing how, you know, at least I feel that the criticism that was made of Web3, of cryptocurrency, of the metaverse, I feel like made a difference in trying to make sure that there were more critical perspectives of these technologies in the media, in, you know, the discourse that people were engaging with when it came to these technologies in the media, in the discourse that people were engaging with when it came to these technologies.
Starting point is 01:01:27 And so my hope is that if we're able to kind of hone a critique of these technologies early on, that we might be able to influence some of this coverage and hopefully not have the repetition of some of these kind of 2010s hype cycles around some of these technologies that we saw in the past. So that's what I'm crossing my fingers for. I hope so too. Yeah. I think it does make a difference.
Starting point is 01:01:49 When I talk to Joy and I'm like, look at face recognition. I thought we made progress and it's everywhere now. It's like exploding. And she's like, what I think about is what would have happened if we didn't do what we did. And so I do. You're right. I do think it's making a difference.
Starting point is 01:02:06 I'd like to see more of it, you know, I'd like to see more. Totally. Yeah. I completely agree. And I think that's a great place to end off our conversation. Timnit, thank you so much for taking the time. It's been fantastic to chat with you. Thank you so much for having me. Timnit Gebru is the founder and executive director of the Distributed AI Research Institute. You can follow her on Twitter at Timnit Gebru. You can follow me at Paris Marks, and you can follow the show at Tech Won't Save Us. Tech Won't Save Us is produced by Eric Wickham and is part of the Harbinger Media Network.
Starting point is 01:02:38 And if you want to support the work that goes into making it every week, you can go to patreon.com slash tech won't save us and become a supporter. Thanks for listening. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.