TED Talks Daily - Sunday Pick: The UN is speaking up about AI — here's what they're saying

Episode Date: September 29, 2024

Each Sunday, TED shares an episode of another podcast we think you'll love, handpicked for you… by us. AI is shaping every aspect of our lives — but only a handful of tech giants have a s...ay in what this technology can do. So what's going on with world governments? Bilawal Sidhu, host of The TED AI Show, sits down with geopolitical expert Ian Bremmer to unpack the UN's just-released plan for "Governing AI for Humanity," a report that focuses on the urgent need to guide AI towards helping everyone thrive, rather than just the powerful few. Together, they explore the complexities of AI's rapid growth on a worldwide scale and take a clear-eyed look at the pivotal decisions facing us in the very near future.

Transcript
Discussion (0)
Starting point is 00:00:00 TED Audio Collective report on global AI governance. On the TED AI Show, creative technologist Balavel Sidhu sits down with political scientist Ian Bremmer to discuss the findings. You might remember my conversations with Ian on TED Talks Daily before, but in this special episode, Ian and Balavel help us understand why this latest news matters and what we can do to make sure AI is on the path to doing good without unleashing its dark side. Stick around for the conversation and find more of the TED AI show wherever you get your podcasts. It's all coming up after a break. Support for this show comes from Airbnb. If you know me, you know I love staying in Airbnbs when I travel. They make my family feel most at home when we're away from home. As we settled down at our Airbnb during
Starting point is 00:01:09 a recent vacation to Palm Springs, I pictured my own home sitting empty. Wouldn't it be smart and better put to use welcoming a family like mine by hosting it on Airbnb? It feels like the practical thing to do, and with the extra income, I could save up for renovations to make the space even more inviting for ourselves and for future guests. Your home might be worth more than you think. Find out how much at Airbnb.ca slash host. Usually, the TED AI Show comes out every Tuesday, but we're dropping this episode a little off schedule because something of global significance has happened, and we wanted to be one of the first to share it with you. So without further ado, let's get right into it.
Starting point is 00:01:55 Imagine a world controlled not by nations, but by a handful of tech companies, driven primarily by profits, holding the reins of artificial intelligence. A technopolar world, as geopolitical expert Ian Bremmer calls it, where algorithms shape our lives, our economies, and even our wars. Now, this isn't some sci-fi movie. It's the future we're barreling towards right now. And the question is, who gets to write the rules?
Starting point is 00:02:23 Well, now the UN has stepped into the ring with a bold plan, global AI governance. They've released this plan on September 19th, 2024, and we're giving you the scoop right here. I had a chance to read an advanced copy of the report, and it gave me a lot to think about. So, this plan is a set of guardrails the UN wants to put in place for all sovereign nations to adhere to, just like what we do around nuclear power or aviation. Only this time, the stakes are even trickier
Starting point is 00:02:56 because it's AI, right? It's a domain tightly controlled by tech giants. It's an ever-evolving technology that even the most seasoned researchers are still trying to understand. It's an ever-evolving technology that even the most seasoned researchers are still trying to understand. It's a tool that has the potential to both solve humanity's biggest challenges and unleash worldwide chaos. In other words, do global rules of conduct for such a powerful, young technology, ruled by tech alliances, even apply? And is the UN, an almost 80-year-old bureaucracy consisting of sovereign nations,
Starting point is 00:03:28 the right institution for the job? I'm Bilal Volsadu, and this is the TED AI Show, where we figure out how to live and thrive in a world where AI is changing everything. Today, we're joined by Ian Bremmer, an American political scientist and one of the co-authors of this report. He's not just a policy wonk, but someone who's in the room with world leaders, feeling the pulse of this tech revolution. He's explored the very question of governance in the digital age and its shift in power from sovereignties to big tech. And he believes the UN is the only international body in a position to hammer out a credible
Starting point is 00:04:17 global consensus for this world-changing technology. And hey, y'all are going to want to buckle up for this one, because this is a ride into the future we're all building, whether we're ready or not. So, hey, Ian, welcome to the show. Thanks for the while. Looking forward to it. Cool. So this report comes at a pivotal time. AI is obviously no longer this futuristic concept that we hear about in movies. It is certainly in the zeitgeist. But I have to ask, what prompted the UN to take the step towards global AI governance at this specific moment in time? I've been talking with the Secretary General about this issue for more than five years now. He recognized when he became Secretary General, Antonio Guterres,
Starting point is 00:05:07 that there were two global issues where governance was wholly inadequate, that he needed to make utter priorities of his time. One being climate, where it's happening, which is happening way too slow and with far too limited resources for the majority of people on the planet. And second, disruptive technologies, notably artificial intelligence, that was coming way too fast. And there was a complete absence of governance. The reason it took so long for it to happen was in part because the UN is a bureaucratic creature and it takes a while to get things set up. But secondly, because they didn't have a senior envoy for geotechnology in place. But having said that, this has always been a top priority for the Secretary General. And the fact that you then had billions of dollars being
Starting point is 00:06:07 poured into this technology with not just enormous hype, but real excitement around real use cases, changing the way humans interact with each other, changing the way we innovate, changing the way society and the global economy works, meant that an absence of global governance was going to leave a lot on the table and was also likely to get us in real trouble. So the urgency has clearly stepped up in the past, I'm going to say, 18 months. And in that regard, I think that this report is landing at a really critical time. Yeah, one might even say it's coming at just the right time, because you're totally right, especially, let's call it the chat GPT moment or the large language model moment, where certainly ML has been on the steady trajectory. But really, you know, the fact that
Starting point is 00:07:01 the average person in the world can now start like experiencing this technology seems like a more recent phenomenon from machines sort of that can just understand to machines that can create. And so along with that sort of this bucket of technologies that requires governance, you know, the report points out that a shared understanding of the opportunities and risks is key. If you had to distill down for our audience, just a couple of the key points in terms of what the opportunities and risks are when it comes to this technology, what would they be? affects their daily lives. They will engage with chatbots. They will use creative AI tools. They'll create art with it. They'll create videos with it. They'll edit with it. They will brainstorm with it. Anyone that has access to a smartphone or a computer now has access to AI in at least a limited way that is transformative as a tool. It's a helpmate and it's about to become vastly more than that. In some cases, it's a friend, it's a counselor, it can be an advisor. Those are the opportunities for individual human beings. But then you have the industrial, the corporate, the enterprise use cases. You get innovation, you get new vaccines, you get new batteries,
Starting point is 00:08:23 you get more efficient ways of stacking traffic, you get new batteries, you get, you know, more efficient ways of, you know, stacking traffic. You get all of that kind of thing. who are talking to me today, this week, this month, about the billions of dollars that they are already saving from deploying AI in their industrial use cases. this is not a technology that is being overly hyped. In my view, these are technologies that are still radically underappreciated because most people don't yet know or see all the ways that they are about to be deployed and are already being deployed. So that's the first point. The second point is on the on the risk side is that all of the companies and they're mostly companies, mostly not governments that are driving the investment in AI are capitalists when it comes to any losses or costs that come from their products. And I don't know about you, but I would like them to be capitalists for both profits and losses. That's how we got ourselves in trouble on climate change, is we had large numbers of industry and governments aligned with that industry that was very happy to be capitalists in making lots of money from exploiting the natural wealth that exists in the planet and
Starting point is 00:10:13 were very happy to push the losses off to be everyone's problem, which means the global South, your children, the poor. And that's not capitalism. That is rapacious oligarchy. And so what we need is to ensure that the negative externalities that come from deploying AI, like any jobs that are lost, like skill sets that need to be redeployed, like disinformation and the use of AI
Starting point is 00:10:48 by malevolent actors or tinkerers in ways that are dangerous for our economy or national security. And also to make sure that the opportunities generated by AI are not simply taken advantage of by small numbers of comparatively wealthy people who have the market serve them most effectively because they have the deep pockets. And those risks will grow if they are not addressed. And they will not be addressed if the marketplace is allowed to capture the governance model. Remember, a functioning free market economy is one where corporations compete against each other in a level playing field with governments and regulators that act as independent and fair arbiters.
Starting point is 00:11:42 And that is not the way the American economic system runs. And that's not the way the global economic system runs. And that's particularly dangerous when you're deploying new technologies where there's an absence, a complete absence of regulation. That is, I think, the context where what the United Nations is presently trying to do is operating it. You know, I think you very nicely outlined why this is sort of a double-edged sword, right? You're making a brilliant point, which is that we're hearing about these large language models that are largely trained off of publicly available data, but all the latent value that
Starting point is 00:12:20 is sitting in all these enterprises and even governments across the world, I mean, we haven't yet begun to see the economic impact that's going to have. At the same time, yes, this technology could absolutely obliterate the world a bit, so to speak. And the climate analogy is a very interesting one, too. I want to do two things. I want to talk about the players here, and then I want to talk about incentives. You've stated in the past that we're headed towards this world that you call a technopolar world, right, where these large tech companies are effectively sovereign in this digital world that we're talking about. You know, yet it's important for incumbent sovereign powers to govern both of these digital and physical worlds that we spend so much time in and that are getting more
Starting point is 00:13:00 and more interconnected. Could you break down this shift in sort of power dynamics that is taking place and how AI plays into all of it? Sure. I mean, when you think about artificial intelligence, first of all, the bots and the models that are being created, that we're all using, they are being developed largely, overwhelmingly, by private sector corporations who have business models and respond to market forces. And they are primarily incented by profitability and to be able to out-compete other companies in this space. That will determine what they create, unless and until they are incented in other non-market ways by governments and by other actors.
Starting point is 00:13:51 And overwhelmingly in the first few years, couple years of the explosion of AI, the governments are playing catch up. There's no global governance model. It's overwhelmingly just a small number of advanced industrial economies. And further, a lot of it is catching up to the state of where AI is today, where you have no governance at all, as opposed to thinking ahead to where AI is likely to be in three or five years, which is a dramatically different and more capable place. And the fact that the models, the data, the priorities, the incentives, the actions are being determined by the companies as opposed to the
Starting point is 00:14:36 governments does create an environment which is technopolar. In other words, when you talk about the AI environment, at least today in 2024, you're really not talking about an environment where governments are sovereign. Now, I mean, ultimately, of course, governments have the ability to enforce laws and these companies don't exist in the virtual world. They exist in localities. They have bank accounts and they're not just all on crypto off the grid. And they live in houses and they have citizenship. And so they absolutely can be governed. But the de facto reality is that the decisions that are being made about what AI is and what is done with it are being made overwhelmingly by the technologists, by a small number of the technologists. So they're
Starting point is 00:15:34 the ones that are deciding. And that means a couple of things. It means, first of all, if you are a government actor or a multilateral organization comprised of government actors, and you want to create effective governance, you actually need to work with the private sector. You can't do this by yourself because you don't have the knowledge. You won't move fast enough. You don't understand what needs to be done. They're the ones doing it. And if you don't do that, they will de facto regulate themselves. And that is a challenging thing because their interests are in treating citizens at best as consumers and at worst as products. Not. It didn't work out too well for social media, did it? Not as citizens. And governments care at best in people as citizens. So, I mean, if the government's going to work for the people, the government has to work with the private sector on AI.
Starting point is 00:16:35 That is the reality of where we are today. certainly learned with a lot of humility at the United Nations is that if you want to start governing something effectively, you first need to define it. You need to understand it. You need a common understanding of what AI is, where it's going, both the opportunities and the concerns. And this is where the UN was so effective on climate change and it took a long time. But we're now in a situation where almost every single member state, no matter if they're poor or rich, small or big, a dictatorship or a democracy or something in between, they all agree that we've had 1.2 centigrade of climate change. They all agree on how much carbon and methane is in the atmosphere. They all agree on the implications of that for extreme weather conditions and loss of biodiversity, impact of plastics in the oceans, you name it, right? In other words, all of the negative externalities from globalization,
Starting point is 00:17:50 the world agrees on that. Now, they don't agree on what to do about it. They don't agree on the resources. But my God, if you can agree on the problem and the opportunity, then when you have resources, you'll deploy them more rationally. When you have conversations about global governance, it's much more likely to move in the right direction. It's so much easier to harvest cooperation as opposed to zero-sum conflict when you all agree on what the challenge is. So we are now at the point in the United Nations on global governance, not where we're calling for, we've got to create an authority that has the arm to compel behavior. It's not that. No, we're at the point where we want to commonly define the opportunities and the risks that come from the state of play of artificial intelligence right now and going forward.
Starting point is 00:18:42 Because doing that will give us the tools we need to govern. I think that is very important for technology as nebulously defined and many associated technologies that sort of need to be taken in concert to govern this effectively. And of course, a central point in what you're saying seems to be that we cannot leave this to market forces, right, to influence the development and deployment of this technology. We know exactly how that went with social media, not super well for the end user, maybe not even that well for society at whole. Oh, clearly, clearly horribly for society as a whole, right? I mean, social media as it stands right now is undermining the civic architecture of our societies. And that is not
Starting point is 00:19:27 the intention of social media. It is purely an unintended but logical consequence of the business model. Oh, 100 percent. If you're optimizing for engagements and that's the dial you're moving things towards, funny things will happen. Now, AI, again, presents its own challenges, and you're kind of alluding to this. So I have to ask, what should be the guiding principles for technology that, as the report states, is one, non-explainable by design? Number two, isn't still fully understood by experts? Number three, is borderless by nature? How on earth do we govern a technology like this at the global level? Well, I think that what you want from a technology like this, in addition to having the world come together to define the state of climate today and how it's been changed by humanity, the world has also come together to define. We have a universal declaration of human rights and we have sustainable development goals. The world agrees that people should have the opportunity to engage in productive labor. The world agrees that people should have adequate shelter. The world agrees that people should have adequate food. The world agrees that people should have adequate food. The world agrees that people should have adequate water. The world agrees people should have access to health care.
Starting point is 00:20:51 I mean, these are basic things, and yet the world agrees that there should be goals for the 8 billion plus people on the planet, and we should orient towards those goals. So the baseline for artificial intelligence and governance should be how do you help use and deploy these technologies in service of those goals? There's nothing really bigger than that, right? It's not more complicated than that. Right now, we are not actually on a path to achieving those goals. In fact, since the pandemic, we've actually backslid on things like hunger and on forced migration. AI is a singular tool available to humanity that could help, I think will help humanity achieve those goals if it is deployed in service of those goals. There is nothing more noble and nothing more sustainable for our planet than to accomplish that. And so I think those are your principles. And I think the goals of the UN high
Starting point is 00:22:06 level panel on artificial intelligence governance should be in service of that. It really is a question of objective functions. And what you're saying is, hey, there's already global consensus on these things that we want to achieve. You know, Agenda 2030, the sustainable development goals. Let's put this new technology to bear on, you know, making a huge dent in all of those goals as quickly as possible. And by the way, I think it's easier than climate in a fundamental way. And it makes me very excited, which is climate change. There were an awful lot of very rich, very powerful actors, governmental, individual, corporate, banking, that really saw that addressing the climate agenda, nevermind advancing it, just addressing it,
Starting point is 00:22:56 was existentially threatening to their wellbeing. That is absolutely not true of AI. You do not have people in industry out there that are saying, I need to pull the wool over collective humanity's eyes and not have AI used in service of humanity. It's not that. It's actually that if these people focus purely on the business models and the market agenda, they're just not going to get to all of the other opportunities for AI because it's not a priority. So they won't bother with the global South. They won't bother with a lot of the poor.
Starting point is 00:23:33 They won't bother as much with biodiversity because there's so much low-hanging fruit for them to make, you know, working with the Fortune 500 and working with, you know, the deep pocketed in the Gulf and Europe and the United States. And so the point is, if you just create that knowledge, if you create those opportunities, if you let everyone know, give you another point, Bill Wall,
Starting point is 00:23:58 if you talk to the owners of these AI companies, the technologists, the CEOs, the founders, the shareholders, you know, they are looking for good things to do. They want, I mean, it's not like they're not charitable people, right? Totally. Altruism is a huge part of the culture in many of these companies too. It really is, but they don't necessarily know what the right thing is to do. So if you don't have a global conversation with common standards and with a common global scientific and policy understanding of here are the ways that AI can have the greatest impact, then you are going to waste so much money. You're going to waste so much good effort. Instead of having AI for humanity, you're going to have AI washing. You know, you're going to have people just saying, hey, here's what I'm doing with AI and just make
Starting point is 00:24:50 sure that at least they have a good story to tell. I'm not interested in giving people good stories to tell. Right. So this is actually this is something that is completely doable and it's not a collective action problem. It is just a coordination issue. I love that. Yeah, it's like the saying in Silicon Valley, move metrics that matter. If you can't define the metrics to move, everyone's just going to be running
Starting point is 00:25:13 in conflicting directions. Exactly. And so by having sort of a common set of goals to move towards, I think that could be extremely, extremely impactful. You brought up the global South and the report emphasizes protecting the interests of the Global South, right?
Starting point is 00:25:29 And a couple of key ingredients are outlined. I'm paraphrasing here, but roughly I'd break it down into compute, data, and talent. What specific measures are being proposed to ensure this equitable access to AI technologies and sort of prevent this widening of the digital divide. There's so many languages that are already underrepresented in the modern large language models, but there's so much more than that too, right? Like how the heck do you even get access to the NVIDIA GPUs that you need to train or fine tune your models? So how is that going to work? Well, I mean, first of all, it's going to start slowly and then hopefully it picks up. So, you know, you've got two concrete recommendations that we hope to be taken up by the member states.
Starting point is 00:26:11 One is an AI capacity development network. And that idea is to have UN-affiliated development centers that make expertise compute and AI training data available. And, you know, that means, you know, you're going to have sandboxes to test AI solutions and learn by doing. It means you have educational opportunities and resources for university students in the global South and programs for researchers, social entrepreneurs, and training public sector officials, which they generally don't have right now, you know, a fellowship program, you know, for individuals to spend time from the global South into top
Starting point is 00:26:57 programs working on AI in the developed world. These things have to be created. There's a second recommendation, which is connected to that, which is a global fund for AI. In other words, that you end up raising the money from the private and the public sector that would support this development capacity. And when you have half of Africa that doesn't even have access to electricity, capacity development for their data isn't going to get them on AI, right? So there are big issues out there that have to be addressed, but this is a unique tool. And the idea is that every million dollars that goes into this global fund is going to be oriented in a way that the world believes is most effective to get AI to be used by the world for the development of humanity. It's inconceivable that an individual corporation would try to do this or government would try to do this by themselves.
Starting point is 00:28:01 It's well said. And I think you also make the point that the incentives are aligned, which is often not the case. You are basically by providing this baseline level of compute and abilities for these markets, you're basically unlocking markets for these companies as well. And so I think the incentives there are quite well aligned. I have to ask one question related to that, which is the sort of common pushback in Silicon Valley when it comes to all of this stuff, which is, you know, especially AI companies, particularly in Europe, sort of feel that the existing regulatory landscape, right, GDPR and so forth, DMA, are already so restrictive and it's the large companies that benefit from this. And hey, what's going to happen to all the smaller companies and the smaller players?
Starting point is 00:28:43 How do you sort of balance consumer protection while also enabling innovation? How do you think about that? I do fear that at the platform level, the amount of compute that is required, the energy to service that compute, the water to service that compute. I mean, you're talking about a small number of corporations and governments that are capable of doing that. And the data that they have available to them allows them to identify startup corporations very quickly
Starting point is 00:29:20 that they can buy, support, or kill. And we've seen that historically with Facebook slash meta. We've seen it with a lot of organizations. And we're seeing it play out right now. Some of the startups that raised hundreds of millions of dollars are like, we just can't keep paying this compute cost. They can't keep paying the compute cost. So, look, I mean, it's certainly there's an argument to be made that maybe if AI models go open source, that it's possible that you end up with a whole bunch of players that are using that AI in their own local applications, their own central applications that make a big difference. You wouldn't make, most people in AI would not make a strong bet that that's where AI is going, right? I mean, that Meta is making that model, making that bet, but most of
Starting point is 00:30:10 the money is not heading in that direction. Certainly that's not true in China, which we haven't talked about as a wholly different kettle of fish because it's largely self-contained. But on the industrial side, what China does is going to be extremely important. Look, I mean, I could answer this question in a more dystopian way, which is for most of my life, the reason that state capitalism was a failure is because you have kleptocracy, inefficiency, nepotism that makes those companies make very bad decisions. But it might well be that when AI is deployed in real time with access to massive data sets, that AI making decisions at the enterprise level to drive corporate growth might prove to be more efficient than what the private sector can do by itself. And if that is true, then you will have a very strong anti-competitive impulse because suddenly economies of scale together with efficiencies driven by AI means that the best possible outcome you can have
Starting point is 00:31:28 economically is non-competitive. But if that's where we're heading, then the role of effective regulation to ensure that individuals have rights that are safeguarded becomes far more important, right? Because people then consumers no longer have the ability to choose A as opposed to B, right? They're going to be in an environment, an AI environment that controls everything, that has, you know- It's like a fish in water. Right, exactly. I mean, it'll be like, I mean, you can, it'll get, I mean, I am of the view that consumers
Starting point is 00:32:06 within five years will probably use an AI ecosystem for an incredible amount of social, business, educational, health, and other relationships, and they'll probably all be the same AI because there'll be advantages in having that data set together in one place. And you'll have preferences depending on your economic well-being and what kind of service you want and everything else, also your geography. Okay, but if that's the case, there would be real costs and you switch it. You already feel this. Like, I mean, if you're on Twitter and you've spent like a bunch of years
Starting point is 00:32:46 building up your following and your connections and suddenly you want to go off, like you have no- You're locked in. You're locked in. You have no rights. If that's where we're heading on the consumer side, and if on the industrial side,
Starting point is 00:32:59 we end up with, again, a small number of players that have extraordinary amounts of data that are able to drive the innovation because they have the data to compute the energy, then the role of effective governance and regulation at the national and at the global level becomes far more urgent. So I am not answering your question well, because your question was like, well, what happens to the small companies? I'm like, I'm more concerned if the small companies end up dying or you can have a whole bunch of small companies. But as they become big enough to be noticed, they have to align. What happens to the small creators?
Starting point is 00:33:38 They can all create. But if you become big enough that you make a move, you've got to sign up with a record company. You've got to sign up with like you've got to be a Facebook influencer, Instagram, whatever it is. Right. I mean, I'm saying I could see that happening in every sector with every corporation and startup. But if that happens, then we've got to spend a lot more attention making sure that individuals still have these rights. You're bringing up this point where the current paradigm, at least that we're experiencing, it does seem to be that the companies that, you know, control large amounts of compute, large amounts of data are going to be at a huge advantage. They're going to be creating
Starting point is 00:34:13 these products and experiences that really mediate our interactions with the digital world and increasingly the physical world too. And so, you know, in this context, open source is brought up as this sort of counterbalance, if you will. And it's a very hotly debated topic, too, because there's a spectrum of sort of open source and closed source, right? Like, let's be honest, most of the open source models in the wild today are more like open weights versus fully open source models where you have access to the training data and the source code and all that good stuff. The report advocates for this stance of meaningful openness.
Starting point is 00:34:49 Can you just elaborate on that? It means that you are hoping that people, governments and organizations around the world aren't purely takers of AI models, values, preferences, and products that come from the biggest global platforms in the West and in China. That you hope that there is meaningful openness, enough openness that individuals and enterprises and governments will be able to actually deploy AI in meaningful ways for their benefit over time. That is the hope. That's what you're calling for. But there's no power behind that. There's no money. There's no authority behind that. That's just the collective view of 39 people around the world that happen to have various expertise on AI. And it's something that was fairly easy to agree on. Now, I will say, because I'm agnostic as to
Starting point is 00:35:53 which way this is going to go, as opposed to which way I would like to see it go. I mean, analysis is not preference here, right? And that's a very important thing to keep in mind when you're talking about these issues. There are different risks that come from these models. So if it turns out that overwhelmingly we have closed models and a very small number of companies maybe that need to be aligned with either the United States or China that end up dominating the AI space, then the critical governance framework you will need for national security and sustainability is a series of arms control agreements in AI between the Americans, the Chinese, and relevant private sector actors, like what you had between the Americans and the Soviets. But we only got that after the 1962 Cuban Missile Crisis, when we almost blew ourselves up.
Starting point is 00:36:51 That's a serious problem, right? Because in the first 15 years, the Americans and the Soviets are developing A-bombs and then H-bombs and then missiles to launch them and all of this capacity. And we're going as fast as we can to assert dominance. And the Soviets as fast as they can to undermine that dominance. And, you know, well, you don't want to talk to your enemy and let them know what you're doing because, I mean, that's just going to give them information. It's going to constrain you.
Starting point is 00:37:17 Well, no, it turns out you really need to have those agreements because otherwise you will blow yourselves up. And so that is essential, I would say, for the next US administration. Secondarily, if it turns out that actually the drivers of cutting edge AI turns out more like not just meaningful openness, but radical openness, where almost anybody in the world, good actor, bad actor,
Starting point is 00:37:48 in between, tinkerer, can take AI models and deploy it for their own purposes. That means that the future of AI technology is going to be more like the global financial marketplace. It'll be systemically important. Everyone will use it and everyone will need it for their own purposes, but also almost anyone could be a systemic threat to it, right? And could bring it down. Like kind of like, you know, you've got a couple of Yahoo's on Reddit talking about GameStop and before you know it, you've got like a challenge for the global markets, right? And so then what you need is a geotechnology stability board, something like the financial stability board that sees that, okay, we are actually not politicized. We're independent
Starting point is 00:38:40 technocrats with an intention of identifying potential threats to the AI marketplace, ensuring resilience, and responding to crises, ringing the bell and saying we've got a crisis, and then responding to them collectively as soon as humanly possible. That's a radically different kind of governance than the US-China arms control governance. And we presently need to start thinking about both because we're not actually sure which sort of methodology, if you will, is going to be most successful in how AI will be AI running the world, if you will, it could very well be something where traditional diplomatic channels might be far too low bandwidth or almost stifling to respond as quickly as possible. But I will. The book. Okay, so it's interesting. It's a book on climate. And it's about, you know, kind of medium term, somewhat dystopian, you know, massive wet bulb climate incident. Millions die in India. Lots of terrorism, eco-terrorism, and need to have global technocratic independent governance
Starting point is 00:40:24 over carbon emissions and related things because the world just has failed, failed, failed, right? Fascinating concept, but probably not necessary for climate, in part because the money is coming, the technology is there at scale. We are moving to a post-carbon environment that is, you know, I mean, it's later than you'd like it to be with a lot of costs, but nonetheless, it's going to happen in the next two generations. AI may well require a ministry of the future. You may need, like you have central bankers today, you may need to have ministers of AI play that kind of a role who are truly like in there, they're appointed
Starting point is 00:41:07 there, they're in parts of national governments, but they act in a politically independent way. They're not driven by ideology. They're not driven by political cycle. They're driven to ensure that the system globally continues to work. And when you have a major financial crisis, you know, the head of the People's Bank of China and the head of the Fed have the same tools. You've got fiscal tools, you've got monetary tools, and you've got the same playbook. You've got the same definitions
Starting point is 00:41:38 and you all want to get the markets working and you really want to avoid a global depression, right? I could easily see that being what the future of AI needs to be like, but it's much harder. It's much harder because the marketplace, you know, is dealing with, you know, exposures and movements of, you know, massive amounts of money. But AI, it's not just about, you know, sort of digital equivalent of money, right? It's like, it's everything. It's national security, it's an election, it's bio, it's, you know, it's drones, it's anything. So it is like the central bank issue, but it would be much more powerful. And, you know, for people that are thinking
Starting point is 00:42:25 about artificial general intelligence, I mean, the reality is that, you know, humanity could probably never handle artificial general intelligence unless you had vastly more powerful governance to deal with that and governance at a technocratic and global level. down at our Airbnb during a recent vacation to Palm Springs, I pictured my own home sitting empty. Wouldn't it be smart and better put to use welcoming a family like mine by hosting it on
Starting point is 00:43:11 Airbnb? It feels like the practical thing to do, and with the extra income, I could save up for renovations to make the space even more inviting for ourselves and for future guests. Your home might be worth more than you think. Find out how much at airbnb.ca slash host. I have to ask about, you know, you asked about the China question. Like there's this ambitious vision for global AI governance that's been outlined here. A common critique of the UN tends to be, how are you going to enforce these guidelines, especially with, you know, major tech companies that might play along right now, but at some point might want to assert their own independence or countries like China that might be reluctant to comply?
Starting point is 00:43:53 Well, that's why I use the climate change example is you did not start with compliance. You started with defining the challenge, the opportunity and the risk. So you had an intergovernmental panel on climate change. And that's what predated all of the committed, even if non-binding, carbon reductions and targets and everything else. First, you had to have a whole bunch of countries that all agreed these are what the opportunities and challenges are. And that's what we are starting with. We are starting with an international scientific panel on AI, just like the intergovernmental panel on climate change, which is meant to regularly convene and define the capabilities, the opportunities, the risks, the uncertainties, the gaps in scientific consensus on AI development and trends. And the point is, if you get the world to agree on that
Starting point is 00:44:53 and exchange standards and talk about opportunities and align with sustainable development goals, well, then clearly you'll have your annual summits on AI and you'll get governments to say, OK, let's apply some resources to this and let's make some commitments on this. And you have some countries that will want to be leaders on it and they will. And they'll, you know, they'll push, they'll prod, they'll try to get other countries to come on board. Now, the latter is much harder and much slower than the former. But you can't do the latter without the former. My view, which I think has has been one that has been shared really by consensus, not coming out there saying we want to have enforceable governance power over what technology companies if you don't actually know what you're trying to accomplish. So don't don't put the wagon before the horse. Get get the problem right first.
Starting point is 00:46:12 Define it. It makes a ton of sense, right? Like if you're not going to survey the landscape and build a common map, you're obviously going to have disagreements from the get go. But what about more challenging topics? Like, for instance, in light of the increasing use of AI in both the Ukraine-Russian conflict, right, and the Israel-Gaza war, what position does the report take on developing international norms or regulations specifically addressing the use of AI in warfare? Well, I mean, there's already been efforts at the United Nations to ban lethal autonomous drones and artificial intelligence in those areas where national
Starting point is 00:46:47 security is directly in the target. You know, you also have the beginnings of the U.S., China, and South Korea talking about the nuclear decision-making process, for example. My view is that governments, a small number of governments, are probably going to be necessary to drive that and not the U.N., even though the U.N. can have a say, because ultimately this will need to be policed. And the way you police is by having willingness to deter and some kind of actual enforcement, sanctioning enforcement. There are going to be costs. I mean, the Iranians, you know, are not standing up to the non-proliferation treaty. If they're not developing nukes, it's because they're concerned that the Israelis are going to blow them up with U.S. support if they actually decide to go full nuclear breakout. And that I think when it comes to using AI in these weapon
Starting point is 00:47:48 systems, when you are talking about life and death for the Russians and Ukrainians on the ground, and they have access to those tools to make into weapons, they will make them into weapons. So you're going to have to have the most powerful countries in the world be willing not just to opine, but to govern, to exert force, to ensure that AI is not used by itself in targeting decisions, is not used by itself in warfighting that takes decisions in life of death completely out of the loop. We've already seen what happens as war becomes more and more distant from humanity, is that it becomes much easier for people to pull the trigger and much easier for wars to escalate. And that's not where we want to be. Are you personally more worried about digital attacks or political interference, you know, via the internet or sort of these physical manifestations of AI systems when it comes to like drone warfare and things like that? I'm much more worried about the vulnerability of economic systems to AI manipulation.
Starting point is 00:48:55 I'd be much more worried about financial markets crash or critical infrastructure crash, especially when you already have antagonists that are deeply inside other systems and just sitting there. In other words, a crowd strike type situation, but instead of a mistake, a malevolent such hit. I mean, you saw what happened to, you know, sort of, I mean, cost billions of dollars immediately and shut down all sorts of systems, computers, you know, air travel, you name it, you know, around the West. And that was just a mistake, just a mistaken patch. What happens if that was done intentionally by an adversary, whether a rogue actor or by a government? I think that's a much bigger concern. But that's not my biggest concern. My biggest concern is something actually much more prosaic.
Starting point is 00:49:40 What is it? My biggest concern is that human beings are programmable. We are very susceptible to What is it? potentiality is shaped by human beings around you, individual human beings and collections of human beings. That's now shifting to algorithms. And, you know, social media is a part of this problem, but when it's AI and when human beings are developing principal relationships algorithmically, and they're no longer engaging principally with other human beings. So nurture is being replaced by algorithm and by AI. I worry that we will lose our humanity. I think that's a much more, for me, it's a much more profound, it's a much more philosophical issue, but it's also a really real near-term threat. We may develop AGI in 20 years,
Starting point is 00:50:48 but humans may become algorithmic in five. And I think we should resist that. I think that at the very least, we need a lot of testing to understand the implications before we want to run real-time experiments on humanity. And right now, what we're doing is running real-time experiments, right? We need a smaller sandbox. I think you're totally right. It goes back to technology mediating our interactions in the physical and digital world. And that gives us sort of the surface area for us to be impacted and influenced is greater than ever before. I'd argue this is already the case today. That's in part why online advertising is just so darn effective. That's why mimetic warfare is a thing, right? Yeah, but your relationships still are
Starting point is 00:51:35 principally with people, right? Oh yeah. Now that once we have intelligent entities that are sort of talking to us and, oh, good Lord. Right. I mean, we've already, you know, basically seen AI pass the Turing test in the last year. And now we have a bunch of people in AI that tell us that that's not that important. That's just moving the goalpost. See, it's moving the goalpost. I think actually it is essential for what matters, because what matters, for us at least, is humanity. And so if we can no longer tell the difference between what
Starting point is 00:52:07 is and is not a human being and we start engaging more principally with non-human actors in service of other goals that have nothing to do with human beings, then we have just unwillingly, unknowingly given up something essential. I'm deeply uncomfortable with that. Once this report is made public, you're obviously going to get a huge amount of feedback from various stakeholders, right? What would you like the audience that's listening to this to do? Should they engage with this report? And how should they give their feedback to what y'all are proposing? I think they should certainly talk about the fact that AI is a tool that needs to be used for all of us. You know, if there's any real message here, it's that we are turning much more inwards at a nation by nation level. And we're not thinking about common collective humanity. We are now inventing tools that allow us to improve ourselves together and to redress a lot of the damage that we've done
Starting point is 00:53:18 to the planet. I mean, if I look at young people today, we need to be honest with them. You know, we've not been effective stewards of the planet. And we need them to make better decisions than we have made. I would argue this report has been an effort over the last year to try to give young people an opportunity to make better decisions, right? And that's the way I want people to respond to this report. Awesome. Ian, thank you so much for joining us. My pleasure. Okay, Ian's insights are a wake-up call.
Starting point is 00:54:01 AI isn't just about code and algorithms. It's a global challenge that demands global cooperation, just like climate change or the interconnected web of financial markets. I mean, think about it. When a financial system teeters on the brink, central bankers don't care about borders or political alliances. They pick up the phone because their fates are intertwined. Their incentives align to prevent disaster. That's the kind of urgency and cooperation we need for AI. What gives me hope is that this isn't some impossible challenge. We have achieved this level of cooperation before.
Starting point is 00:54:38 Now, the question is, can we replicate it for AI before it's too late? Will we see a world dominated by a few tech giants hoarding the power of AI? Or can we foster a scenario where a thousand flowers bloom, where AI empowers individuals and nations across the globe in a truly interconnected, yet decentralized AI ecosystem? This conversation with Ian has been a reality check. AI isn't just about futuristic hype. It's about the choices we make today. This UN report is a vital blueprint to build a shared understanding on what exactly it is that we're trying to regulate. But the real work begins with us understanding the complexities, demanding transparency, and pushing for responsible AI development.
Starting point is 00:55:28 And doing that without stifling innovation and nimble startups. The future of AI is not predetermined. It's an interconnected tapestry and a global story we're all writing together. The TED AI Show is a part of the TED Audio Collective and is produced by TED with Cosmic Standard. Our producers are Ben Montoya and Alex Higgins. Our editors are Banban Chang and Alejandra Salazar. Our showrunner is Ivana Tucker and our engineer is Asia Pilar Simpson.
Starting point is 00:56:01 Our technical director is Jacob Winnick, and our executive producer is Eliza Smith. Our researcher and fact checker is Christian Aparta. And I'm your host, Bilal Velsidou. See y'all in the next one. Looking for a fun challenge to share with your friends and family? TED now has games designed to keep your mind sharp while having fun. Visit TED.com slash games to explore the joy and wonder of TED Games.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.