TED Talks Daily - Sunday Pick: The UN is speaking up about AI — here's what they're saying
Episode Date: September 29, 2024Each Sunday, TED shares an episode of another podcast we think you'll love, handpicked for you… by us. AI is shaping every aspect of our lives — but only a handful of tech giants have a s...ay in what this technology can do. So what's going on with world governments? Bilawal Sidhu, host of The TED AI Show, sits down with geopolitical expert Ian Bremmer to unpack the UN's just-released plan for "Governing AI for Humanity," a report that focuses on the urgent need to guide AI towards helping everyone thrive, rather than just the powerful few. Together, they explore the complexities of AI's rapid growth on a worldwide scale and take a clear-eyed look at the pivotal decisions facing us in the very near future.
Transcript
Discussion (0)
TED Audio Collective report on global AI governance. On the TED AI Show, creative technologist Balavel Sidhu sits
down with political scientist Ian Bremmer to discuss the findings. You might remember my
conversations with Ian on TED Talks Daily before, but in this special episode, Ian and Balavel help
us understand why this latest news matters and what we can do to make sure AI is on the path
to doing good without unleashing
its dark side. Stick around for the conversation and find more of the TED AI show wherever you get
your podcasts. It's all coming up after a break. Support for this show comes from Airbnb. If you
know me, you know I love staying in Airbnbs when I travel. They make my family feel most at home when we're away from home. As we settled down at our Airbnb during
a recent vacation to Palm Springs, I pictured my own home sitting empty. Wouldn't it be smart and
better put to use welcoming a family like mine by hosting it on Airbnb? It feels like the practical
thing to do, and with the extra income, I could save up for renovations to make the space even more inviting for ourselves and for future guests.
Your home might be worth more than you think.
Find out how much at Airbnb.ca slash host.
Usually, the TED AI Show comes out every Tuesday, but we're dropping this episode a little off schedule because something of global
significance has happened, and we wanted to be one of the first to share it with you.
So without further ado, let's get right into it.
Imagine a world controlled not by nations, but by a handful of tech companies, driven
primarily by profits, holding the reins of artificial intelligence. A technopolar world,
as geopolitical expert Ian Bremmer calls it,
where algorithms shape our lives,
our economies, and even our wars.
Now, this isn't some sci-fi movie.
It's the future we're barreling towards right now.
And the question is, who gets to write the rules?
Well, now the UN has stepped into the ring with a bold plan, global AI governance.
They've released this plan on September 19th, 2024, and we're giving you the scoop right here.
I had a chance to read an advanced copy of the report, and it gave me a lot to think about.
So, this plan is a set of guardrails
the UN wants to put in place
for all sovereign nations to adhere to,
just like what we do around nuclear power or aviation.
Only this time, the stakes are even trickier
because it's AI, right?
It's a domain tightly controlled by tech giants.
It's an ever-evolving technology
that even the most seasoned researchers are still trying to understand. It's an ever-evolving technology that even the most seasoned researchers are still
trying to understand. It's a tool that has the potential to both solve humanity's biggest
challenges and unleash worldwide chaos. In other words, do global rules of conduct for such a
powerful, young technology, ruled by tech alliances, even apply? And is the UN, an almost 80-year-old bureaucracy
consisting of sovereign nations,
the right institution for the job?
I'm Bilal Volsadu, and this is the TED AI Show,
where we figure out how to live and thrive in a world
where AI is changing everything. Today, we're joined by Ian Bremmer, an American political scientist
and one of the co-authors of this report. He's not just a policy wonk, but someone who's in the
room with world leaders, feeling the pulse of this tech revolution. He's explored the
very question of governance in the digital age and its shift in power from sovereignties to big tech.
And he believes the UN is the only international body in a position to hammer out a credible
global consensus for this world-changing technology. And hey, y'all are going to want
to buckle up for this one, because this is a ride
into the future we're all building, whether we're ready or not. So, hey, Ian, welcome to the show.
Thanks for the while. Looking forward to it. Cool. So this report comes at a pivotal time.
AI is obviously no longer this futuristic concept that we hear about in movies.
It is certainly in the zeitgeist.
But I have to ask, what prompted the UN to take the step towards global AI governance at this specific moment in time?
I've been talking with the Secretary General about this issue for more than five years now. He recognized when he became Secretary General, Antonio Guterres,
that there were two global issues where governance was wholly inadequate, that he needed to make
utter priorities of his time. One being climate, where it's happening, which is happening way too
slow and with far too limited resources for the
majority of people on the planet. And second, disruptive technologies, notably artificial
intelligence, that was coming way too fast. And there was a complete absence of governance.
The reason it took so long for it to happen was in part because the UN is a bureaucratic creature and it takes a while to get
things set up. But secondly, because they didn't have a senior envoy for geotechnology in place.
But having said that, this has always been a top priority for the Secretary General. And the fact that you then had billions of dollars being
poured into this technology with not just enormous hype, but real excitement around real use cases,
changing the way humans interact with each other, changing the way we innovate, changing the way society and
the global economy works, meant that an absence of global governance was going to leave a lot
on the table and was also likely to get us in real trouble. So the urgency has clearly stepped up
in the past, I'm going to say, 18 months. And in that regard, I think that this report is landing at a
really critical time. Yeah, one might even say it's coming at just the right time, because you're
totally right, especially, let's call it the chat GPT moment or the large language model moment,
where certainly ML has been on the steady trajectory. But really, you know, the fact that
the average person in the world can now start like experiencing this technology seems like a more recent phenomenon from machines sort of that can just understand to machines that can create.
And so along with that sort of this bucket of technologies that requires governance, you know, the report points out that a shared understanding of the opportunities and risks is key. If you had to distill down for our audience, just a couple of the key points in terms of what the opportunities and risks are when it comes to this technology, what would they be? affects their daily lives. They will engage with chatbots. They will use creative AI tools. They'll
create art with it. They'll create videos with it. They'll edit with it. They will brainstorm
with it. Anyone that has access to a smartphone or a computer now has access to AI in at least
a limited way that is transformative as a tool. It's a helpmate and it's about to become vastly more
than that. In some cases, it's a friend, it's a counselor, it can be an advisor. Those are the
opportunities for individual human beings. But then you have the industrial, the corporate,
the enterprise use cases. You get innovation, you get new vaccines, you get new batteries,
you get more efficient ways of stacking traffic, you get new batteries, you get, you know, more efficient ways of, you know, stacking traffic.
You get all of that kind of thing. who are talking to me today, this week, this month, about the billions of dollars that they are already saving from deploying AI in their industrial use cases. this is not a technology that is being overly hyped. In my view, these are technologies that
are still radically underappreciated because most people don't yet know or see all the ways that
they are about to be deployed and are already being deployed. So that's the first point. The second point is on the on the risk side is that all of the companies and they're mostly companies, mostly not governments that are driving the investment in AI are capitalists when it comes to any losses or costs that come from their products.
And I don't know about you, but I would like them to be capitalists for both profits and losses.
That's how we got ourselves in trouble on climate change,
is we had large numbers of industry and governments aligned with that industry that was very happy to be
capitalists in making lots of money from exploiting the natural wealth that exists in the planet and
were very happy to push the losses off to be everyone's problem, which means the global South, your children, the poor. And that's not capitalism.
That is rapacious oligarchy.
And so what we need is to ensure
that the negative externalities
that come from deploying AI,
like any jobs that are lost,
like skill sets that need to be redeployed,
like disinformation and the use of AI
by malevolent actors or tinkerers in ways that are dangerous for our economy or national security.
And also to make sure that the opportunities generated by AI are not simply taken advantage of by small numbers of comparatively wealthy people who have the market serve them most effectively because they have the deep pockets.
And those risks will grow if they are not addressed.
And they will not be addressed if the marketplace is allowed to capture the governance model.
Remember, a functioning free market economy
is one where corporations compete against each other
in a level playing field with governments and regulators
that act as independent and fair arbiters.
And that is not the way the American economic system runs.
And that's not the way the global economic system runs. And that's particularly dangerous when
you're deploying new technologies where there's an absence, a complete absence of regulation.
That is, I think, the context where what the United Nations is presently trying to do is operating it.
You know, I think you very nicely outlined why this is sort of a double-edged sword,
right?
You're making a brilliant point, which is that we're hearing about these large language
models that are largely trained off of publicly available data, but all the latent value that
is sitting in all these enterprises and even governments across the world, I mean, we haven't
yet begun to see the economic impact that's going to have. At the same time, yes,
this technology could absolutely obliterate the world a bit, so to speak. And the climate analogy
is a very interesting one, too. I want to do two things. I want to talk about the players here,
and then I want to talk about incentives. You've stated in the past that we're headed towards this
world that you call a technopolar world, right, where these large tech companies are effectively sovereign in this digital
world that we're talking about. You know, yet it's important for incumbent sovereign powers to govern
both of these digital and physical worlds that we spend so much time in and that are getting more
and more interconnected. Could you break down this shift in sort of power dynamics that is taking
place and how AI plays into all of it? Sure. I mean, when you think about artificial intelligence,
first of all, the bots and the models that are being created, that we're all using,
they are being developed largely, overwhelmingly, by private sector corporations
who have business models and respond to market forces. And they are primarily incented by
profitability and to be able to out-compete other companies in this space. That will determine what
they create, unless and until they are incented in other non-market
ways by governments and by other actors.
And overwhelmingly in the first few years, couple years of the explosion of AI, the governments
are playing catch up.
There's no global governance model.
It's overwhelmingly just a small number of advanced industrial
economies. And further, a lot of it is catching up to the state of where AI is today, where you
have no governance at all, as opposed to thinking ahead to where AI is likely to be in three or five
years, which is a dramatically different and more capable place. And the fact that the models, the data, the
priorities, the incentives, the actions are being determined by the companies as opposed to the
governments does create an environment which is technopolar. In other words, when you talk about
the AI environment, at least today in 2024,
you're really not talking about an environment where governments are sovereign. Now, I mean,
ultimately, of course, governments have the ability to enforce laws and these companies
don't exist in the virtual world. They exist in localities. They have bank accounts
and they're not just all on crypto off the grid. And they live in houses and they have citizenship.
And so they absolutely can be governed. But the de facto reality is that the decisions that are being made about what AI is and what is done with it
are being made overwhelmingly by the technologists, by a small number of the technologists. So they're
the ones that are deciding. And that means a couple of things. It means, first of all,
if you are a government actor or a multilateral organization comprised of government actors, and you want to
create effective governance, you actually need to work with the private sector. You can't do this
by yourself because you don't have the knowledge. You won't move fast enough. You don't understand
what needs to be done. They're the ones doing it. And if you don't do that, they will de facto regulate themselves. And that is a
challenging thing because their interests are in treating citizens at best as consumers and at
worst as products. Not. It didn't work out too well for social media, did it? Not as citizens. And governments care at best in people as citizens.
So, I mean, if the government's going to work for the people, the government has to work with the private sector on AI.
That is the reality of where we are today. certainly learned with a lot of humility at the United Nations is that if you want to start
governing something effectively, you first need to define it. You need to understand it. You need
a common understanding of what AI is, where it's going, both the opportunities and the concerns.
And this is where the UN was so effective on climate change and it took a long time.
But we're now in a situation where almost every single member state, no matter if they're poor or rich, small or big, a dictatorship or a democracy or something in between, they all agree that we've had 1.2 centigrade of climate change.
They all agree on how much carbon and methane is in the atmosphere. They all agree on the
implications of that for extreme weather conditions and loss of biodiversity, impact of plastics in
the oceans, you name it, right? In other words, all of the negative externalities from globalization,
the world agrees on that. Now, they don't agree on what to do about it. They don't agree on the resources. But my God, if you can agree on the problem and the opportunity, then when you have
resources, you'll deploy them more rationally. When you have conversations about global governance,
it's much more likely to move in the right direction. It's so much easier to harvest cooperation as opposed to zero-sum conflict
when you all agree on what the challenge is. So we are now at the point in the United Nations
on global governance, not where we're calling for, we've got to create an authority that has the arm to compel behavior.
It's not that.
No, we're at the point where we want to commonly define the opportunities and the risks that
come from the state of play of artificial intelligence right now and going forward.
Because doing that will give us the tools we need to
govern. I think that is very important for technology as nebulously defined and many
associated technologies that sort of need to be taken in concert to govern this effectively.
And of course, a central point in what you're saying seems to be that we cannot leave this
to market forces, right, to influence the development
and deployment of this technology. We know exactly how that went with social media, not super well
for the end user, maybe not even that well for society at whole. Oh, clearly, clearly horribly
for society as a whole, right? I mean, social media as it stands right now is undermining the civic architecture of our societies. And that is not
the intention of social media. It is purely an unintended but logical consequence of the business
model. Oh, 100 percent. If you're optimizing for engagements and that's the dial you're moving
things towards, funny things will happen. Now, AI, again, presents its own challenges, and you're
kind of alluding to this. So I have to ask, what should be the guiding principles for technology that, as the report states, is one, non-explainable by design?
Number two, isn't still fully understood by experts?
Number three, is borderless by nature?
How on earth do we govern a technology like this at the global level? Well, I think that what you want from a technology like this, in addition to having the world come together to define the state of climate today and how it's been changed by humanity, the world has also come together to define. We have a universal declaration of human rights and we have sustainable development goals. The world agrees that people should have the opportunity to engage in productive labor. The world agrees that people should have adequate shelter. The world agrees that people should have adequate food. The world agrees that people should have adequate food. The world agrees
that people should have adequate water. The world agrees people should have access to health care.
I mean, these are basic things, and yet the world agrees that there should be goals for the 8
billion plus people on the planet, and we should orient towards those goals. So the baseline for artificial intelligence and governance should
be how do you help use and deploy these technologies in service of those goals? There's nothing really
bigger than that, right? It's not more complicated than that. Right now, we are not actually on a path to achieving those goals. In fact,
since the pandemic, we've actually backslid on things like hunger and on forced migration.
AI is a singular tool available to humanity that could help, I think will help humanity achieve those goals if it is deployed in service
of those goals. There is nothing more noble and nothing more sustainable for our planet
than to accomplish that. And so I think those are your principles. And I think the goals of the UN high
level panel on artificial intelligence governance should be in service of that. It really is a
question of objective functions. And what you're saying is, hey, there's already global consensus
on these things that we want to achieve. You know, Agenda 2030, the sustainable development goals.
Let's put this new technology to bear on, you know, making a huge
dent in all of those goals as quickly as possible. And by the way, I think it's easier than climate
in a fundamental way. And it makes me very excited, which is climate change. There were an
awful lot of very rich, very powerful actors, governmental, individual, corporate, banking,
that really saw that addressing the climate agenda, nevermind advancing it, just addressing it,
was existentially threatening to their wellbeing. That is absolutely not true of AI. You do not have people in industry out there that are saying, I need to pull the wool over
collective humanity's eyes and not have AI used in service of humanity.
It's not that.
It's actually that if these people focus purely on the business models and the market agenda,
they're just not going to get to all of the other opportunities for AI
because it's not a priority.
So they won't bother with the global South.
They won't bother with a lot of the poor.
They won't bother as much with biodiversity
because there's so much low-hanging fruit
for them to make, you know,
working with the Fortune 500
and working with, you know,
the deep pocketed
in the Gulf and Europe and the United States. And so the point is, if you just create that knowledge,
if you create those opportunities, if you let everyone know, give you another point, Bill Wall,
if you talk to the owners of these AI companies, the technologists, the CEOs, the founders,
the shareholders, you know, they are looking for good things to do.
They want, I mean, it's not like they're not charitable people, right?
Totally. Altruism is a huge part of the culture in many of these companies too.
It really is, but they don't necessarily know what the right thing is to do. So if you don't have a global conversation with common standards and with a common global scientific and policy understanding of here are the ways that AI can have the greatest impact, then you are going to waste so much money.
You're going to waste so much good effort.
Instead of having AI for humanity, you're going to have AI washing.
You know, you're going to have people just saying, hey, here's what I'm doing with AI and just make
sure that at least they have a good story to tell. I'm not interested in giving people good stories
to tell. Right. So this is actually this is something that is completely doable and it's
not a collective action problem. It is just a coordination issue.
I love that.
Yeah, it's like the saying in Silicon Valley,
move metrics that matter.
If you can't define the metrics to move,
everyone's just going to be running
in conflicting directions.
Exactly.
And so by having sort of a common set of goals
to move towards,
I think that could be extremely, extremely impactful.
You brought up the global South
and the report emphasizes protecting the interests of the
Global South, right?
And a couple of key ingredients are outlined.
I'm paraphrasing here, but roughly I'd break it down into compute, data, and talent.
What specific measures are being proposed to ensure this equitable access to AI technologies
and sort of prevent this widening of the digital divide. There's so many languages that are already underrepresented in the modern large language
models, but there's so much more than that too, right? Like how the heck do you even get access
to the NVIDIA GPUs that you need to train or fine tune your models? So how is that going to work?
Well, I mean, first of all, it's going to start slowly and then hopefully it picks up.
So, you know, you've got two concrete recommendations that we hope to be taken up by the member states.
One is an AI capacity development network.
And that idea is to have UN-affiliated development centers that make expertise compute and AI training data available.
And, you know, that means, you know, you're going to have sandboxes to test AI solutions
and learn by doing.
It means you have educational opportunities and resources for university students in the
global South and programs for researchers, social entrepreneurs,
and training public sector officials, which they generally don't have right now, you know,
a fellowship program, you know, for individuals to spend time from the global South into top
programs working on AI in the developed world. These things have to be created. There's a second
recommendation, which is connected to that, which is a global fund for AI. In other words, that you end up raising the money from the private and the public sector that would support this development capacity. And when you have half of Africa that doesn't even have access to electricity, capacity
development for their data isn't going to get them on AI, right?
So there are big issues out there that have to be addressed, but this is a unique tool.
And the idea is that every million dollars that goes into this global fund is going to be oriented in a way that the world believes
is most effective to get AI to be used by the world for the development of humanity.
It's inconceivable that an individual corporation would try to do this or government would try
to do this by themselves.
It's well said.
And I think you also make the point that the incentives are aligned, which is often not the case. You are basically by providing this baseline level of compute and
abilities for these markets, you're basically unlocking markets for these companies as well.
And so I think the incentives there are quite well aligned. I have to ask one question related
to that, which is the sort of common pushback in Silicon Valley when it comes to all of this stuff, which is, you know, especially AI companies, particularly in Europe, sort
of feel that the existing regulatory landscape, right, GDPR and so forth, DMA, are already
so restrictive and it's the large companies that benefit from this.
And hey, what's going to happen to all the smaller companies and the smaller players?
How do you sort of balance consumer protection while also enabling innovation?
How do you think about that?
I do fear that at the platform level, the amount of compute that is required,
the energy to service that compute, the water to service that compute.
I mean, you're talking about a small number of corporations
and governments that are capable of doing that.
And the data that they have available to them
allows them to identify startup corporations very quickly
that they can buy, support, or kill. And we've seen that historically with Facebook
slash meta. We've seen it with a lot of organizations. And we're seeing it play out
right now. Some of the startups that raised hundreds of millions of dollars are like,
we just can't keep paying this compute cost. They can't keep paying the compute cost. So,
look, I mean, it's certainly there's an argument to be made that maybe if AI models go open source, that it's possible that you end up with a whole bunch of
players that are using that AI in their own local applications, their own central applications that
make a big difference. You wouldn't make, most people in AI would not make a strong bet that that's
where AI is going, right? I mean, that Meta is making that model, making that bet, but most of
the money is not heading in that direction. Certainly that's not true in China, which we
haven't talked about as a wholly different kettle of fish because it's largely self-contained. But
on the industrial side, what China does is going to be extremely important. Look, I mean, I could answer this question in a more dystopian way, which is for most of my life, the reason that state capitalism
was a failure is because you have kleptocracy, inefficiency, nepotism that makes those companies make very bad decisions.
But it might well be that when AI is deployed in real time with access to massive data sets,
that AI making decisions at the enterprise level to drive corporate growth might prove to be more efficient
than what the private sector can do by itself. And if that is true, then you will have a very
strong anti-competitive impulse because suddenly economies of scale together with efficiencies driven by AI means that the best possible outcome you can have
economically is non-competitive. But if that's where we're heading, then the role of effective
regulation to ensure that individuals have rights that are safeguarded becomes far more important,
right? Because people then consumers no longer have the ability to choose A as opposed to B, right?
They're going to be in an environment, an AI environment that controls everything, that
has, you know-
It's like a fish in water.
Right, exactly.
I mean, it'll be like, I mean, you can, it'll get, I mean, I am of the view that consumers
within five years will probably use an AI ecosystem for an incredible amount of social,
business, educational, health, and other relationships, and they'll probably all be
the same AI because there'll be advantages in having that data set together in one place.
And you'll have preferences depending on your economic well-being and what kind of service
you want and everything else, also your geography.
Okay, but if that's the case, there would be real costs and you switch it.
You already feel this.
Like, I mean, if you're on Twitter and you've spent like a bunch of years
building up your following and your connections
and suddenly you want to go off,
like you have no-
You're locked in.
You're locked in.
You have no rights.
If that's where we're heading on the consumer side,
and if on the industrial side,
we end up with, again, a small number of players
that have extraordinary amounts of data that are
able to drive the innovation because they have the data to compute the energy, then the role
of effective governance and regulation at the national and at the global level becomes far
more urgent. So I am not answering your question well, because your question was like, well, what happens to the small companies?
I'm like, I'm more concerned if the small companies end up dying or you can have a whole bunch of small companies.
But as they become big enough to be noticed, they have to align.
What happens to the small creators?
They can all create.
But if you become big enough that you make a move, you've got to sign up with a record company.
You've got to sign up with like you've got to be a Facebook influencer, Instagram, whatever it is.
Right. I mean, I'm saying I could see that happening in every sector with every corporation and startup.
But if that happens, then we've got to spend a lot more attention making sure that individuals still have these rights.
You're bringing up this point where the current paradigm, at least that we're experiencing,
it does seem to be that the companies that, you know, control large amounts of compute,
large amounts of data are going to be at a huge advantage. They're going to be creating
these products and experiences that really mediate our interactions
with the digital world and increasingly the physical world too. And so, you know,
in this context, open source is brought up as this sort of
counterbalance, if you will. And it's a very hotly debated topic, too, because there's a spectrum of
sort of open source and closed source, right? Like, let's be honest, most of the open source
models in the wild today are more like open weights versus fully open source models where
you have access to the training data and the source code and all that good stuff.
The report advocates for this stance of meaningful openness.
Can you just elaborate on that?
It means that you are hoping that people, governments and organizations around the world aren't purely takers of AI models, values, preferences, and products that come from the
biggest global platforms in the West and in China. That you hope that there is meaningful openness,
enough openness that individuals and enterprises and governments will be able to actually deploy AI in meaningful ways for their
benefit over time. That is the hope. That's what you're calling for. But there's no power behind
that. There's no money. There's no authority behind that. That's just the collective view
of 39 people around the world that happen to have various expertise on AI.
And it's something that was fairly easy to agree on. Now, I will say, because I'm agnostic as to
which way this is going to go, as opposed to which way I would like to see it go. I mean,
analysis is not preference here, right? And that's a very important thing to keep in mind when you're talking about these issues. There are different risks that come from these models. So if it turns
out that overwhelmingly we have closed models and a very small number of companies maybe that need
to be aligned with either the United States or China that end up dominating the AI space,
then the critical governance framework you will need for national security and sustainability
is a series of arms control agreements in AI between the Americans, the Chinese,
and relevant private sector actors, like what you had between the Americans and the Soviets. But we only got that after the 1962 Cuban Missile Crisis, when we almost blew ourselves
up.
That's a serious problem, right?
Because in the first 15 years, the Americans and the Soviets are developing A-bombs and
then H-bombs and then missiles to launch them and all of this capacity.
And we're going as fast as we can to assert dominance.
And the Soviets as fast as they can to undermine that dominance.
And, you know, well, you don't want to talk to your enemy and let them know what you're
doing because, I mean, that's just going to give them information.
It's going to constrain you.
Well, no, it turns out you really need to have those agreements because otherwise you
will blow yourselves up.
And so that is essential,
I would say, for the next US administration.
Secondarily, if it turns out that actually
the drivers of cutting edge AI turns out more like
not just meaningful openness, but radical openness,
where almost anybody in the world, good actor, bad actor,
in between, tinkerer, can take AI models and deploy it for their own purposes.
That means that the future of AI technology is going to be more like the global financial
marketplace.
It'll be systemically important. Everyone will use it and
everyone will need it for their own purposes, but also almost anyone could be a systemic threat to
it, right? And could bring it down. Like kind of like, you know, you've got a couple of Yahoo's
on Reddit talking about GameStop and before you know it, you've got like a challenge for the global markets, right? And so then what you need is a geotechnology stability board, something like the financial
stability board that sees that, okay, we are actually not politicized. We're independent
technocrats with an intention of identifying potential threats to the AI
marketplace, ensuring resilience, and responding to crises, ringing the bell and saying we've got
a crisis, and then responding to them collectively as soon as humanly possible. That's a radically different kind of governance
than the US-China arms control governance. And we presently need to start thinking about both
because we're not actually sure which sort of methodology, if you will, is going to be most successful in how AI will be AI running the world, if you will, it could very well be something where traditional diplomatic channels might be far too low bandwidth or almost stifling to respond as quickly as possible. But I will. The book. Okay, so it's interesting. It's a book on climate.
And it's about, you know, kind of medium term, somewhat dystopian, you know, massive wet bulb climate incident.
Millions die in India.
Lots of terrorism, eco-terrorism, and need to have global technocratic independent governance
over carbon emissions and related
things because the world just has failed, failed, failed, right? Fascinating concept,
but probably not necessary for climate, in part because the money is coming, the technology is
there at scale. We are moving to a post-carbon environment that is, you know, I mean, it's
later than you'd like it to be with a lot of costs, but nonetheless, it's going to happen
in the next two generations.
AI may well require a ministry of the future.
You may need, like you have central bankers today, you may need to have ministers of AI play that kind of a role who are truly like in there, they're appointed
there, they're in parts of national governments, but they act in a politically independent way.
They're not driven by ideology. They're not driven by political cycle. They're driven
to ensure that the system globally continues to work. And when you have a major financial crisis,
you know, the head of the People's Bank of China
and the head of the Fed have the same tools.
You've got fiscal tools, you've got monetary tools,
and you've got the same playbook.
You've got the same definitions
and you all want to get the markets working
and you really want to avoid a global depression, right?
I could easily see that being what the future of AI needs to be like, but it's much harder. It's much harder
because the marketplace, you know, is dealing with, you know, exposures and movements of,
you know, massive amounts of money. But AI, it's not just about, you know, sort of digital
equivalent of money, right? It's like, it's everything. It's national security,
it's an election, it's bio, it's, you know, it's drones, it's anything. So it is like
the central bank issue, but it would be much more powerful. And, you know, for people that are thinking
about artificial general intelligence,
I mean, the reality is that, you know,
humanity could probably never handle
artificial general intelligence
unless you had vastly more powerful governance
to deal with that
and governance at a technocratic and global level. down at our Airbnb during a recent vacation to Palm Springs, I pictured my own home sitting empty.
Wouldn't it be smart and better put to use welcoming a family like mine by hosting it on
Airbnb? It feels like the practical thing to do, and with the extra income, I could save up for
renovations to make the space even more inviting for ourselves and for future guests. Your home
might be worth more than you think. Find out how much at airbnb.ca slash host.
I have to ask about, you know, you asked about the China question. Like there's this
ambitious vision for global AI governance that's been outlined here.
A common critique of the UN tends to be, how are you going to enforce these guidelines,
especially with, you know, major tech companies that might play along right now, but at some point might want to
assert their own independence or countries like China that might be reluctant to comply?
Well, that's why I use the climate change example is you did not start with compliance.
You started with defining the challenge, the opportunity and the risk. So you had an intergovernmental panel on climate change.
And that's what predated all of the committed,
even if non-binding, carbon reductions and targets
and everything else.
First, you had to have a whole bunch of countries
that all agreed these are what the opportunities and challenges are. And that's what we are starting with. We are starting with an international scientific panel on AI, just like the intergovernmental panel on climate change, which is meant to regularly convene and define the capabilities, the opportunities, the risks, the uncertainties, the gaps in scientific consensus
on AI development and trends. And the point is, if you get the world to agree on that
and exchange standards and talk about opportunities and align with sustainable development goals,
well, then clearly you'll have your annual summits on AI and you'll get governments to say, OK, let's apply some resources to this and let's make some commitments on this.
And you have some countries that will want to be leaders on it and they will.
And they'll, you know, they'll push, they'll prod, they'll try to get other countries to come on board.
Now, the latter is much harder and much slower than the former.
But you can't do the latter without the former. My view, which I think has has been one that has been shared really by consensus, not coming out there saying we want to have enforceable governance power over what technology companies if you don't actually know what you're trying to accomplish.
So don't don't put the wagon before the horse.
Get get the problem right first.
Define it.
It makes a ton of sense, right?
Like if you're not going to survey the landscape and build a common map, you're obviously going to have disagreements from the get go.
But what about more challenging topics?
Like, for instance, in light of the increasing use of AI in both the
Ukraine-Russian conflict, right, and the Israel-Gaza war, what position does the report take on
developing international norms or regulations specifically addressing the use of AI in warfare?
Well, I mean, there's already been efforts at the United Nations to ban lethal autonomous drones and artificial intelligence in those areas where national
security is directly in the target. You know, you also have the beginnings of the U.S., China,
and South Korea talking about the nuclear decision-making process, for example.
My view is that governments, a small number of governments, are probably going to be necessary to drive that and not the U.N., even though the U.N. can have a say, because ultimately this will need to be policed.
And the way you police is by having willingness to deter and some kind of actual enforcement, sanctioning enforcement.
There are going to be costs.
I mean, the Iranians, you know, are not standing up to the non-proliferation treaty.
If they're not developing nukes, it's because they're concerned that the Israelis are going
to blow them up with U.S. support if they actually decide to go full nuclear breakout. And that I think when it comes to using AI in these weapon
systems, when you are talking about life and death for the Russians and Ukrainians on the ground,
and they have access to those tools to make into weapons, they will make them into weapons.
So you're going to have to have the most powerful countries in the world be willing not just to opine, but to govern, to exert force, to ensure that AI is not used by itself in targeting decisions, is not used by itself in warfighting that takes decisions in life of death completely out of the loop. We've already
seen what happens as war becomes more and more distant from humanity, is that it becomes much
easier for people to pull the trigger and much easier for wars to escalate. And that's not where
we want to be. Are you personally more worried about digital attacks or political interference,
you know, via the internet or sort of these physical manifestations of AI systems when it comes to like drone warfare and things like that?
I'm much more worried about the vulnerability of economic systems to AI manipulation.
I'd be much more worried about financial markets crash or critical infrastructure crash, especially when you already have antagonists that are deeply inside other
systems and just sitting there. In other words, a crowd strike type situation, but instead of a
mistake, a malevolent such hit. I mean, you saw what happened to, you know, sort of, I mean,
cost billions of dollars immediately and shut down all sorts of systems, computers, you know,
air travel, you name it, you know, around the West. And that was just a
mistake, just a mistaken patch. What happens if that was done intentionally by an adversary,
whether a rogue actor or by a government? I think that's a much bigger concern. But that's not my
biggest concern. My biggest concern is something actually much more prosaic.
What is it?
My biggest concern is that human beings are programmable. We are very susceptible to What is it? potentiality is shaped by human beings around you, individual human beings and collections
of human beings. That's now shifting to algorithms. And, you know, social media is a part of this
problem, but when it's AI and when human beings are developing principal relationships
algorithmically, and they're no longer engaging principally with other human
beings. So nurture is being replaced by algorithm and by AI. I worry that we will lose our humanity.
I think that's a much more, for me, it's a much more profound, it's a much more philosophical
issue, but it's also a really real near-term threat. We may develop AGI in 20 years,
but humans may become algorithmic in five. And I think we should resist that. I think that at the
very least, we need a lot of testing to understand the implications before we want to run real-time experiments
on humanity. And right now, what we're doing is running real-time experiments, right?
We need a smaller sandbox. I think you're totally right. It goes back to technology
mediating our interactions in the physical and digital world. And that gives us sort of the
surface area for us to be impacted and influenced is greater than ever before. I'd
argue this is already the case today. That's in part why online advertising is just so darn
effective. That's why mimetic warfare is a thing, right? Yeah, but your relationships still are
principally with people, right? Oh yeah. Now that once we have intelligent entities that are sort
of talking to us and, oh, good Lord. Right. I mean, we've already, you know, basically seen AI pass the Turing test in the last year.
And now we have a bunch of people in AI that tell us that that's not that important.
That's just moving the goalpost.
See, it's moving the goalpost.
I think actually it is essential for what matters, because what matters, for us at least,
is humanity.
And so if we can no longer tell the difference between what
is and is not a human being and we start engaging more principally with non-human actors in service
of other goals that have nothing to do with human beings, then we have just unwillingly, unknowingly
given up something essential. I'm deeply uncomfortable with that.
Once this report is made public, you're obviously going to get a huge amount of feedback from
various stakeholders, right? What would you like the audience that's listening to this to do?
Should they engage with this report? And how should they give their feedback to what y'all
are proposing? I think they should certainly talk about the fact that AI is a tool that needs to be used for all of us. You know, if there's any real message here, it's that we are turning much more inwards at a nation by nation level. And we're not thinking about common collective humanity. We are now inventing tools
that allow us to improve ourselves together and to redress a lot of the damage that we've done
to the planet. I mean, if I look at young people today, we need to be honest with them.
You know, we've not been effective stewards of the planet. And we need them to make better decisions than we have made.
I would argue this report has been an effort over the last year to try to give young people
an opportunity to make better decisions, right?
And that's the way I want people to respond to this report.
Awesome. Ian, thank you so much for joining us.
My pleasure.
Okay, Ian's insights are a wake-up call.
AI isn't just about code and algorithms.
It's a global challenge that demands
global cooperation, just like climate change or the interconnected web of financial markets.
I mean, think about it. When a financial system teeters on the brink, central bankers don't care
about borders or political alliances. They pick up the phone because their fates are intertwined. Their incentives align to prevent disaster.
That's the kind of urgency and cooperation we need for AI.
What gives me hope is that this isn't some impossible challenge.
We have achieved this level of cooperation before.
Now, the question is, can we replicate it for AI before it's too late?
Will we see a world dominated by a few tech giants
hoarding the power of AI? Or can we foster a scenario where a thousand flowers bloom,
where AI empowers individuals and nations across the globe in a truly interconnected,
yet decentralized AI ecosystem? This conversation with Ian has been a reality check. AI isn't just about futuristic hype.
It's about the choices we make today. This UN report is a vital blueprint to build a shared
understanding on what exactly it is that we're trying to regulate. But the real work begins with
us understanding the complexities, demanding transparency, and pushing for responsible AI development.
And doing that without stifling innovation
and nimble startups.
The future of AI is not predetermined.
It's an interconnected tapestry
and a global story we're all writing together.
The TED AI Show is a part of the TED Audio Collective and is produced by TED
with Cosmic Standard. Our producers are Ben Montoya and Alex Higgins. Our editors are Banban
Chang and Alejandra Salazar. Our showrunner is Ivana Tucker and our engineer is Asia Pilar Simpson.
Our technical director is Jacob Winnick, and our executive producer is Eliza
Smith. Our researcher and fact checker is Christian Aparta. And I'm your host,
Bilal Velsidou. See y'all in the next one.
Looking for a fun challenge to share with your friends and family? TED now has games designed
to keep your mind sharp while having fun. Visit TED.com slash games to explore the joy and wonder of TED Games.