Your Undivided Attention - Mustafa Suleyman Says We Need to Contain AI. How Do We Do It?

Episode Date: September 28, 2023

This is going to be the most productive decade in the history of our species, says Mustafa Suleyman, author of “The Coming Wave,” CEO of Inflection AI, and founder of Google’s DeepMind. But in o...rder to truly reap the benefits of AI, we need to learn how to contain it. Paradoxically, part of that will mean collectively saying no to certain forms of progress. As an industry leader reckoning with a future that’s about to be ‘turbocharged’  Mustafa says we can all play a role in shaping the technology in hands-on ways and by advocating for appropriate governance.RECOMMENDED MEDIA The Coming Wave: Technology, Power, and the 21st Century’s Greatest DilemmaThis new book from Mustafa Suleyman is a must-read guide to the technological revolution just starting, and the transformed world it will createPartnership on AIPartnership on AI is bringing together diverse voices from across the AI community to create resources for advancing positive outcomes for people and societyPolicy Reforms Toolkit from the Center for Humane TechnologyDigital lawlessness has been normalized in the name of innovation. It’s possible to craft policy that protects the conditions we need to thriveRECOMMENDED YUA EPISODES AI Myths and MisconceptionsCan We Govern AI? with Marietje SchaakeThe AI DilemmaYour Undivided Attention is produced by the Center for Humane Technology. Follow us on Twitter: @HumaneTech_

Transcript
Discussion (0)
Starting point is 00:00:00 Hey, everyone, this is Tristan. And this is Aza. Your undivided attention is about to have its second ever Ask Us Anything episode. So are there questions that you'd love to ask me or Aza about the show? Or more broadly about our work at the Center for Humane Technology and how we want to tackle these questions of AI? So here's your opportunity. Go to humaneTech.com forward slash ask us. Or, and we love this option, record your question on your phone and send us a voice memo at
Starting point is 00:00:27 undividedat humane tech.com. So that's humanetech.com forward slash ask us or undivided at humanetech.com. We hope to hear from you and we are really looking forward to continuing this dialogue. So I think the challenge for the next century is going to be what we don't do rather than what we do. And in the history of our species so far, progress has been a function of what we do. Progress has never been a function of what we say no to. And now, the strange reality is that we actually need to learn when and how to say no collectively. That's Mustafa Suleiman.
Starting point is 00:01:10 He was one of the three co-founders of DeepMind, one of the very first companies whose stated goal was to create artificial general intelligence. DeepMind was acquired by Google and is now owned by Alphabet. Recently, Mustafa co-founded Inflection AI, a public benefit corporation that launched a chatbot name, named Pi for personal intelligence, which aims to create a personal AI for everyone. Now, we don't usually interview CEOs of AI companies on this show, but Mustafa's position as both a creator and a critic of AI makes him unique. He sits at the center of the Silicon Valley race to develop AI. And as you'll hear, he also argues that containing AI is absolutely necessary. Now, the reason why we started with this quote from Mustafa is he's saying something
Starting point is 00:01:56 deeply surprising that progress is defined by what we say no to versus what we say yes to and that's different than every other era. To say no to something takes a kind of maturity to see a power and to not put your hand out to grasp it. And it's unclear through human history how many times humanity has been able to exercise that kind of maturity. I met Mustafa back in 2014 when I walked into his deep mind office in London. And at that time I was a design ethicist at Google, thinking about how we could govern the unethical outcomes we were seeing from social media and the attention economy, arms race. I was looking for allies, and I found almost literally no one who was really thinking about these issues as deeply as Mustafa was. But before that,
Starting point is 00:02:40 we should mention that Mustafa worked in conflict resolution and human rights, so the lens he brings to ethical questions with AI is a deeply thoughtful one. Right, and he has a new book coming out just this month, and it's called The Coming Wave, Technology, Power, and the 21st Century's greatest dilemma. Now, Mustafa is basically arguing these two simultaneous kind of contradictory truths. On the one hand, containment of AI is impossible. On the other hand, containment of AI must be possible, that we have to prevent runaway AI proliferation. Now, Tristan and I see some of these things differently from Mustafa, but we're all asking ourselves the same questions right now. Is it possible to contain AI? And what would it look like? And we wanted to bring some of these
Starting point is 00:03:24 answers to you. So without further ado, here's Mustafa Suleiman. Thank you so much for that kind welcome. It's incredible to be here, I guess, almost a decade on from us both angsting over the same things, debating strategies for intervention, and it's been amazing to journey with you both. So the argument about whether it's possible to contain AI runs through your book. And before we go deeper into that, what does it mean to contain a technology. When we think of containment, what exactly are we trying to do? I mean, the way I see it is that we're trying to monitor, control, and potentially even close down completely technologies that are either in development or they are in deployment. And being able to do
Starting point is 00:04:17 all of those components represents a threshold of containment that I think should give us comfort that these technologies can be managed in a safe way. It doesn't mean that they will inherently be safe, right? You could have something that is unsafe and contained, but at least then we have a framework for governance and decision-making which ensures that they remain accountable to us. Now, people often thought that in the early 50s, that nuclear power, particularly the weapons component of nuclear power,
Starting point is 00:04:48 would itself inevitably proliferate. And the very good news is that we use traditional force and traditional threats and incentives to suppress the widespread proliferation of those weapons. In fact, three countries actually relinquished nuclear power over the last 20 years. And there are some things to learn from that, but fundamentally that isn't a cause for complacency or comfort because, you know, nuclear is incredibly expensive. it's hugely capital time intensive. Handling the material is extremely difficult. And so it doesn't have any of the characteristics of the coming wave. So I think the challenge for the next century is going to be what we don't do rather than what we do.
Starting point is 00:05:38 And in the history of our species so far, progress has been a function of what we do. Progress has never been a function of what we say no to. So it's this absurd paradox that we now have to, you know, put the genie back in the bottle, right? The engine of progress in the last 2,000 years has been invention, science, creativity, commerce, this massive evolution of ideas, right? Which has just turbocharged health and happiness for everybody on the planet for so many centuries. And now the strange reality is that for various, frankly, technical reasons, we actually need to learn when and how to say no, collectively, right? This can't just be unilateral.
Starting point is 00:06:26 Could you paint a picture for just a second before you get into, like, the why it'll be hard to say no for this technology in particular, but what are we saying no to? I think that there are a number of capabilities which we can demonstrate with extremely high confidence, potentially even mathematical provability. Then we should stand back from experimenting with those tools because they're potentially very, very dangerous. So here are some examples. The first is recursive self-improvement, right? So if an AI system has the ability to update and improve its own code and the way that it
Starting point is 00:07:05 works without human oversight, with access to very large amount of resources, potentially unlimited resources, where it's operating on the open web or it's interacting with humans in an unconstrained way, then that recursive self-improvement could be something that might be hard for us to monitor and hard for us to limit. And if you don't just provide it with the ability to update its own code,
Starting point is 00:07:29 but in fact to change its own goals, right, those two characteristics to me could produce potentially quite dangerous systems. So some people want to be hearing that and saying, okay, well, but it's just code running on a computer. It's a blinking cursor. I go to chat.com.
Starting point is 00:07:44 I hit return. That doesn't move atoms and things around the world doesn't create January 6th is what do you mean specifically when you say through this autonomy and generality it could you know start to impact and make vulnerable the actual physical world that we all live in so the way to think about the coming wave is that everybody is going to have an intelligent assistant or an intelligent agent in their pocket it will be able to take a natural language instruction and carry out actions in the digital and physical world. It means that it could generate a brand new website. It could interact
Starting point is 00:08:20 with other AIs in order to persuade them or negotiate with them or establish a contract with them for physically constructing something, building something. It can send emails. It will be able to make phone calls. It will ultimately see what you see and hear what you hear. So it's going to be sort of living life alongside you and interacting in the world just in the same way that you interact with the world today. So, Mustafa, to recap what you said, lots of people are going to be getting a personal AI assistant, an AI that's there helping them all the time. And that sounds like it's going to be a really big boost to personal productivity and overall
Starting point is 00:08:57 economic output. And a lot of people might be wondering then, well, that sounds great. What's the problem? The problem is one of proliferation. Everybody is going to get access at the same time within the same five-year period. Think about it like this. Today, whatever you earn, if you live in the modern developed world, you almost certainly have the same cutting-edge smartphone
Starting point is 00:09:23 as the very richest people on the planet. We all use the same high-quality laptops. And that's because of proliferation. Over the last 60 years, chips have become a critical commodity in our economy, and everybody has the very best quality. hardware. That is an incredibly meritocratic achievement, if you think about it. It's amazing. Two or three billion of us are within five percent of one another in terms of our access to the best hardware ever invented in the history of our species. The same trajectory is going to
Starting point is 00:09:59 unfold with respect to access to intelligence. Everyone is going to have a team of elite AIs in their corner that help to pursue their objectives. And that's the conundrum. It's both incredible from a productivity perspective, but also incredibly chaotic from a civilization perspective. How do you imagine that impacting markets, nation states, jobs, terrorist groups? This seems like it has massive, massive implications. I love you to just, like, give us a tour of what that means, geopolitically down to an individual. I think the first thing to say is that we are already in a state, of huge international anxiety, right? I think in the Western world, there is a fear of climate
Starting point is 00:10:48 change, there is a job insecurity, a fear that democracy is under threat and under attack, a lack of confidence in the political process. On all fronts, there's this kind of general anxiety, which is, I think, creating fragility in our civilization. And that's an important context and backdrop. The real question is going to be who gets access to a technology like that? How much will it cost to deploy a system of that power? Is it going to be open source or is it going to be centralized? Is that capability going to spread to everybody? That will change everything from the cost of labor to the structure of companies. I mean, it really is nothing short of a seismic transformation in what work means.
Starting point is 00:11:40 And so, you know, it's hard to say exactly where that leaves the future of the nation state. But I think that has the potential to cause huge disruption. And then this brings back this question of containment because of the whole premise of this conversation is that there are places where we might want to say no in the name of progress. That progress actually, I love your frame. You know, the progress in this new era of human history is going to be about our ability collectively to say no rather than to say yes to things. Let me say one thing to complicate the picture first, because I think it informs the answer. And the complication is that this is going to be the most productive and fruitful decade in the history of our species. What I've just described to you with artificial capable intelligence, these ACIs, means that many, many millions of people, if not billions of people, will have access to the best lawyer, the best doctor, the best job, the best job.
Starting point is 00:12:37 the best chief of staff, the best coach, the best confidant, the best creative aid, the best project manager. And so the upside here is going to be so sensational that knowing what to stop is just part of the problem, right? Actually persuading people to hold back is going to be the major challenge. These technologies are inherently omni use, which means the same. core technology is massively valuable in every possible domain. And we aren't talking about the simplistic two-dimensional framing of tools or weapons, right? That was the old sort of, you know, dual-use characteristic of military technologies like nuclear, right, which basically has two uses. But these language models, by their definition, have a unique use for every single
Starting point is 00:13:33 individual. So they're not just dual use. They're infinitely omni use because intelligence and language is the technology of life, right? That is how we communicate with one another. It's how we think, it's how we create, is how we do. And so distilling language and intelligence into something that can be reproduced, scaled up, that's what's both going to turbocharge us as a species, but also lead us to run into one another because we already run into one another. It's simply going to set light to the existing conflicts that we have with one another. The differences in culture and politics and commercial competitive constraints. All right. So getting back to the question of how we contain this Omni-use technology, I'm interested in whether you see this
Starting point is 00:14:20 is something that just happens at an elite political level and inside the big tech companies, or is this something that everybody listening to the podcast can somehow be involved in? Yes. Everybody has to get involved now. Defining the standards of what is acceptable is going to shape the future of our species. And that means participating in protest, in resistance, participating in creating alliances, participating in the political process. It means creating and making these tools. Because if you don't know how to really use these tools and you don't make them, you'll never truly understand how good they are at certain things and how bad they are at other things. And that will often be very surprising.
Starting point is 00:15:03 and they're very accessible. They're really not that complicated. You can wrap your mind around it in two or three hours of watching videos online where there's tons of really accessible tutorials. And I think that's a great place to start when it comes to containment. First, stop fearing it as this impossible to understand enemy. And second, deeply immerse yourself in the tools of holding these models accountable and making them and manipulating them
Starting point is 00:15:31 so that you can see the boundaries of their... rather than just imagining their sort of super-intelligences, which they're totally not. The bigger the models get, the more controllable they are. You know, that's an amazing outcome that we are creating very nuanced, very precise, very specific behaviors in these models. That's a great story because that regresses us to the traditional governance question, which is whose values, what should those behaviors be, and who gets to decide? I've known you for a long time of stuff. I see you as someone who has taken on a tremendous burden and worked exceptionally hard to try to figure out how do we ethically govern the impact of technology on society.
Starting point is 00:16:18 And if I recall correctly, when DeepMind, the most powerful AI company in the world at the time, it was acquired by Google as part of the agreement which you helped construct, Google's use of DeepMind had to be governed by this ethics board that you put together. And so just to sort of say, you have been thinking about how do we get this right? for a very long time. And that was back in, what, 2010, 11 or something. And then later on, you know, building the Google applied ethics board and doing a number of things to try to create different incentives for the dynamics that normally govern how companies would deploy AI in ways that may not always perfectly match up with the best of humanities
Starting point is 00:16:54 plus, you know, public interest. You know, you're totally right that it's the incentives that really make a difference here. And how can we create a new set of incentives? I've really been obsessed by that question since founding DeepMind. When we founded DeepMind, our strapline in our business plan was building safe and ethical AGI for the benefit of everybody. And that really influenced the culture of the company in a very fundamental way.
Starting point is 00:17:20 We attracted a certain type of person as a result. And I think it had a non-trivial effect on the way that OpenAI formed and then Anthropic 2. When we were acquired in 2014, you're right, we made it a condition of the acquisition that DeepMind's technologies could never be used for military purposes. And, you know, we also had, as you said, an ethics and safety board with actually six independent members, outsiders, which were jointly appointed by us at DeepMind, me and Demis and Shane, and, you know, Larry, Sergey and Sundar at the time.
Starting point is 00:17:53 And that was the first steps towards trying to bring in some outside oversight over how we were developing these technologies and how they were governed. And over the course of five or six years, whilst I was a deep mind, I really tried hard to get Google to add an ethics board of its own. For a very brief time, it did have an ethics board. It had its own internal charter, which I drafted the AI principles, the Google AI principles, which are still in operation today. And although there isn't an independent oversight, and unfortunately that all fell apart very quickly, the charter of AI principles is still in operation. and there's a big team now that are focused on responsible AI.
Starting point is 00:18:32 So although it's not glamorous and publicly recognized, there is an operational culture shift in Google's approach, which is now more precautionary, is more proactively attentive to the risks because there's a kind of bureaucratic process in place for scrutinizing
Starting point is 00:18:48 all of the launches that Google does to make sure they adhere to its AI principles. So it's not a complete solution, but it's a first step. I also co-founded the partnership on AI, where we brought together the six big technology companies, Apple, Microsoft, IBM, Google, Facebook, DeepM, Open AI, along with now I think 120 civil society groups from Oxfam to ACLU to everyone's now a member. And it's a forum, you know, it's a way to share best practices.
Starting point is 00:19:19 It's a place to meet other people in the field who are concerned about these issues. It's a place to organize and convene and debate. And they produce lots of good research and lots of. of good events now. So all these are kind of very small steps. You know, none of it in some solves it, but it's, you know, we have to be optimistic. You know, my phrase, containment must be possible because, yes, it does look like we default to mass proliferation and that fundamentally changes our civilization. But it has to be possible. And the way to do that is that we all have to take small steps and be active and participate. There's a question that's in my mind and I bet it's
Starting point is 00:19:58 in the mind of many of our listeners as well. You know, you're the CEO of this new company with Reid Hoffman, Inflection AI, and it's working on creating these personal AIs for everyone. How do you reconcile working in this space, participating in the race, and at the same time raising the alarm about it? You know, inflection was founded as a public benefit corporation, right? What that means is that we have a legal obligation
Starting point is 00:20:26 as directors of the company of inflection to attend to our mission, which is to build AI for the benefit of everybody in a safe way. So we have to think about the harms as a legal requirement proactively in our documentation. So it's not just about satisfying shareholder interest. It's about these other potential harms. And I think little changes like that
Starting point is 00:20:50 really make a difference over the long term. So given all of those attempts that you have tirelessly made over many, many years, to try to create governance of individual AI actors and deep mind and inflection and so on, there's still this question of what external system of governance would be needed to bind this global race. And that gets back to this question of containment, both between companies that are all racing, and maybe they individually have a long-term benefit trust the way Anthropic does or are incorporated as a public benefit corporation the way you are. But that's not enough
Starting point is 00:21:20 when it comes to the overall race dynamics and the pull that pulls us towards, if I don't do it, I just won't have a say in the future, so everyone has to keep building and increasing capabilities to match the speed of the other guys. What is your thought there? I think it needs governments that are prepared to concede and to cooperate because we've talked a lot about how there are going to be proliferation dynamics because people want access in open source and individuals and small organizations want to create their own forms of AI. But we haven't talk so much about the proliferation dynamics that result because governments and militaries are going to be chasing this like there's no tomorrow. And we're already seeing that.
Starting point is 00:22:05 You know, I've been tracking the lethal autonomous weapons negotiations at the UN for the last 15 years. And sadly, we've made almost no progress, I think it's fair to say, in the negotiations. We haven't even been able to define what autonomy means and what it means to have a human in the loop, right? Meaningful human control is the phrase that they debate over. And naturally, that's because autonomous weapons provide asymmetric advantages to small organizations or small nation states or militias. And nobody wants to be the first to say, I'm going to step back from the brink here. It might mean that war is cheaper. It might mean that there are fewer human casualties, at least on one side, if not both.
Starting point is 00:22:53 And so it's a very attractive potential for nation states. And of course, AI has the same potential. And so there's also this competitive race dynamic among nation states, which is driving the default to proliferation. So how do you address that? Well, at the end of the day, whilst these models feel like they're kind of evolving super fast because they're in software,
Starting point is 00:23:18 that software still runs on hardware and that hardware is physically located in a data center on territory which is subject to international law most of the time, right? And those chips are manufactured by a very small number of players. You know, Nvidia produces the vast majority of AI chips that train these models
Starting point is 00:23:39 that produce these experiences. And that represents a significant choke point which nation states can, you know, both track the location, of these data centers. They can impose regulations on people who operate the data centers to report risks or to comply with other kinds of safety standards. So that's going to become the kind of focus of negotiation in the future. And successful containment strategies are most likely going to be around sort of licensing, regulating and monitoring who gets to use what chips
Starting point is 00:24:17 and where they sit, which data centers they sit. What I hear you speaking to is what's also kind of discussed by Sam Malman and others is sort of this international regime for monitoring the flows of chips in the world, where does NVIDIA sending its chips, where are those aggregated, where do those sit in data centers, and we're going to need, whether it's licensing or auditing or tracking, and these kinds of things that are being discussed right now, legislatively in the U.S. and elsewhere.
Starting point is 00:24:42 Whether that results in an IAEA, it's some kind of international body that does that. But that's kind of what you're speaking to, yes? That's exactly right. I think we need an international body. And I think the better example is actually modeled on the financial stability board, which is an international group of central bank governors, which is very collaborative, has coordinated on many, many occasions in the last 15, 20 years in real time and overnight
Starting point is 00:25:09 to avoid catastrophe in financial markets. And so I think we need a geotechnology stability board, which is, proactively coordinating with all of the big corporations and companies like mine that are developing these kinds of models and all of the governments that are doing the same to do everything that you said audit transparency oversight licensing regimes etc etc and that that probably is a new organization and just just jump in there to say also agree that it is incredibly challenging and the hard work is going to be how do we know when we're getting it right Which is different than how do you stop the race.
Starting point is 00:25:49 You know, if exactly as you say, we are going to be in deep transformational relationships with the AI systems, many people are building. But given that distinction, how do we know that they are not transforming us in ways that we will later regret or do incredible amounts of damage? And we're going to need some kind of, I know people don't really like the FDA model, but like we need some kind of FDA to know that before we get into a transformative relationship, the things we're entering into relationships with, are good. And to me, it's almost like a new project for understanding, like, what is it to be humane? Because in every way that these systems do damage that we cannot tell that they're doing damage, in every domain from the effable to the ineffable, they will do. So I think that's true, but they're not the harms that I really worry about. So I think, Aza, maybe think about, just assume that the humane AI is eminently possible within the next
Starting point is 00:26:46 three to five years. Just assume that it is going to be infinitely smart, infinitely patient, infinitely respectful, really boundaryed, really healthy. Just assume that and then think about the potential consequences for those kinds of AIs in the world. What does it mean that you have access to infinite amounts of kind and respectful and patient interactions, right? So that's one sort of simulation. The other is, given that you can create such precise and and specific behaviours like the ones I've just outlined, many other people are going to use the same tools to create a slightly more aggressive AI,
Starting point is 00:27:26 or one that's really good at spreading fear and anxiety about the threat of vaccines, or one that is really good at spreading sticky conspiracy theories about this new politician. And that's the challenge. So the risk is not that an AI goes wrong. The risk is that it is working as intense, but that intention is to sow deceit and disruption.
Starting point is 00:27:51 And we do not have a governance framework given our kind of default free speech world. That's the challenge. It's the overlapping false positive moment. Like it's where the lines are really, really blurred. It's not that the AI won't be humane. It will be superhumane. But some people are going to use it to persuade
Starting point is 00:28:08 on agendas that are really antithetical to peace and prosperity. And it may be very difficult for us to prove that and to counter-argue that actually this is bad. So we have to get good at describing why a particular activity like that is bad for democracy, it's bad for people's health. Because it can't just be, well, these are our values
Starting point is 00:28:28 and we don't like your values, because that's what it might regress to. That's the problem from a kind of communication perspective. How do we describe what bad looks like? We need a framework for that. So given that we lack many of the things that you're describing that we need on the way to containing AI,
Starting point is 00:28:45 like an international body that's collaborating with big companies, developing these models or a way to describe what we don't want to happen in a way that everyone can agree on. Given all of that, is there any cause for optimism here that you can see? Yeah, I think my take is that when people have an opportunity to see the sort of harms and benefits over the next five to ten years, I think it will change the popular conversation entirely. I don't think it's possible to really move the needle
Starting point is 00:29:15 on the collective psyche today because everything is so theoretical, right? I mean, I may think that I can see a potential impact, good and bad, of the coming wave, but, I mean, it pales into insignificance when you actually have the opportunity to experience these models, right?
Starting point is 00:29:35 I mean, I was trying to sort of persuade you of the significance of LLMs two years ago, and I could barely get it on your radar, let alone everybody else, And so I'm very conscious that like from a communications perspective, until you deeply experience and really get a chance to play with this new putty, this new clay, you won't understand where it's crap and where it's amazing. And it's just really difficult to convey that until you get to experience it. So it's only now that everybody has seen chat GPT and GPT4 and all these other models that everyone's starting to kind of catch up essentially. on this AI governance debate, which is great. And so part of the reason why I feel optimistic about it is that I think that that's going to massively escalate as we start to see the harms unfold in practice, which there will
Starting point is 00:30:27 be some really horrible outcomes. But I think on the whole, there are going to be incredible benefits that basically drive people to say, yeah, I want to be able to make that too. I want to be able to learn from that. I want that to help me. I want an AI that is going to help me with my research. And that's why I sort of feel that there will be a reaction, a popular reaction in the next few years. And it may be surprising.
Starting point is 00:30:52 It may go different ways. A lot of people absolutely love talking to these models. A lot of people don't have the same taboo that we thought that they might have a year ago about having deep and intimate conversations with these models. And so I think we should just keep an open mind to how this might unfold. fault because I think it's going to be quite surprising. Mustafa, thank you so much for coming on your undivided attention. I know we're going to be in this journey together for a while.
Starting point is 00:31:24 Thank you, man. I appreciate that. Your undivided attention is produced by the Center for Humane Technology, a non-profit working to catalyze a humane future. Our senior producer is Julia Scott. Kirsten McMurray and Sarah McRae are our associate producers. Sasha Fegan is our managing edict. A very special thanks to our generous supporters who make this entire podcast possible, and if you would like to join them, you can visit humanetech.com slash donate. You can find show notes, transcripts, and much more at humanetech.com.
Starting point is 00:31:56 And if you made it all the way here, let me give one more thank you to you for giving us your undivided attention.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.