The Diary Of A CEO with Steven Bartlett - AI AGENTS EMERGENCY DEBATE: These Jobs Won't Exist In 24 Months! Containment Has Failed, We Must Prepare For What's Coming!
Episode Date: May 12, 2025Will AI replace God, steal your job, and change your future? Amjad Masad, Bret Weinstein, and Daniel Priestley debate the terrifying warning signs, and why you need to understand them now. Amjad ...Masad is the founder and CEO of Replit, the world's leading online programming environment and community. Daniel Priestley is a serial entrepreneur and advises fast-growing companies on innovation, marketing, and the future of work in the age of AI. Bret Weinstein is an evolutionary biologist, expert in complex systems and host of the DarkHorse Podcast. In this debate, they explain: Why AI threatens 50% of the global workforce. How AI agents are already replacing millions of jobs and how to use them to your advantage. How AI will disrupt creative industries and hijack human consciousness. The critical skills that will matter most in the AI-powered future. What parents must teach their kids now to survive the AI age. How to harness AI’s power ethically. 00:00 Intro 07:14 What Is an AI Agent? 09:11 Who Is Bret and What Are His Views on AI? 12:58 Who Is Dan? 14:42 Where Are the Boundaries? 15:56 What Could AI Potentially Do? 17:01 Bret's Concerns: AI and a New Species 19:33 The Disruptive Potential of AI in Its Current Form 20:33 Is AI Just a Tool? 21:38 Those Who Leverage AI Will Be the Winners 25:15 What Abuse Are We Currently Seeing? 30:57 The Collateral Damage of AI 39:08 What Will Happen to Humans? 42:15 Which Jobs Will Be Replaced by AI? 45:34 Could AI Development Affect Western Economies? 49:00 Is AI Removing Our Agency? 57:39 Will Authenticity Be More Valued in the AI Era? 59:16 Will Markets Become Fairer or More Unbalanced? 1:03:45 The Economic Displacement 1:05:45 Worldcoin and the Case for Universal Basic Income 1:11:47 Are We Losing Meaning and Purpose? 1:14:44 AI's Impact on Loneliness, Relationships, and Connection 1:18:57 Can Education Adapt to the AI Era? 1:26:07 What Should AI Teach Our Children? 1:31:19 Ads 1:32:20 Is This Inevitable? 1:38:18 Will We Start Living Like House Cats? 1:44:44 Hyper-Changing World: Are We Designed for It? 1:50:25 The 5 Key Threats of AI 1:51:51 Deepfakes and AI Scams 1:59:43 An Optimistic Take on the AI Era 2:03:37 AI for Business Opportunities 2:08:34 Ads 2:10:32 AI Autonomous Weapons 2:17:59 Do We Live Among Aliens or in a Simulation? 2:21:31 How to Live a Good Life in the AI Era 📣 100 CEOs: Ready to think like a CEO? Gain access to the 100 CEOs newsletter here: https://bit.ly/100-ceos-megaphone Follow Amjad: X - https://bit.ly/3SsT92D Website - https://bit.ly/44y7YIO Follow Daniel: X - https://bit.ly/4jZRa1S Instagram - https://bit.ly/3F2Q8mK Website - https://bit.ly/3Ssg6mG Follow Bret: X - https://bit.ly/3EVAHwH You can purchase Daniel’s book, ‘Scorecard Marketing: The four-step playbook for getting better leads and bigger profits’, here: https://amzn.to/45aByEa You can purchase Bret’s book, ‘A Hunter-Gatherer's Guide to the 21st Century: Evolution and the Challenges of Modern Life’, here: https://amzn.to/4dcJoPU The 1% Diary is back - limited time only: https://bit.ly/3YFbJbt The Diary Of A CEO Conversation Cards (Second Edition): https://g2ul0.app.link/f31dsUttKKb Get email updates: https://bit.ly/diary-of-a-ceo-yt Follow Steven: https://g2ul0.app.link/gnGqL4IsKKb Sponsors: Vivobarefoot - https://vivobarefoot.com/DOAC with code DIARY20 for 20% offShopify - https://shopify.com/bartlett Learn more about your ad choices. Visit megaphone.fm/adchoices
Transcript
Discussion (0)
I think a lot of people don't realize how massive the positive impact AI is going to have on their life.
Well, I would argue that the idea that this AI disruption doesn't lead us to human catastrophe is optimistic.
For example, people are going to be unemployed in huge numbers.
You agree with that, don't you?
Yes. If your job is as routine as it comes, it's gone in the next couple years.
But it's going to create new opportunities for wealth creation.
Let me put it to you this way.
We've created a new species
and nobody on Earth can predict what's going to happen.
We are joined by three leading voices
to debate the most disruptive shift in human history,
the rise of AI.
And they're answering the questions you're most scared about.
This technology is going to get so much more powerful.
And yes, we're going to go through a period of disruption,
but at the other end, we're going to create a fairer world.
It's enabling people to run their businesses, make a lot of money.
And you can solve meaningful problems, such as the breakthroughs in global health care and education will be phenomenal.
And you can live an incredibly fulfilling existence.
Well, I would just say on that front, this has always been the fantasy of technologists,
to do marvelous things with our spare time, but we end up doom-scrolling.
Loneliness epidemic.
Right, following birth rates. So the potential for good here is infinite, and the potential
for bad is ten times. For example, there's war, undetectable beat fakes and scams, so
people don't understand how many different ways they are going to be robbed.
Look, I don't think blaming technology for all of it is the right thing.
All these issues, they're already here.
We're all fathers here. So what are you saying to your children?
Well, first of all...
I find it incredibly fascinating that when we look at the back end of Spotify and Apple and our audio channels,
the majority of people that watch this podcast haven't yet hit the follow button or the subscribe button.
Wherever you're listening to this, I would like to make a deal with you. If you could do me a huge favor and hit that subscribe button
I will work tirelessly from now until forever to make the show better and better and better and better
I can't tell you how much it helps when you hit that subscribe button. The show gets bigger
Which means we can expand the production
Bring in all the guests you want to see and continue to doing this thing we love. If you could do me that small favor
And hit the follow button wherever you're listening to this,
that would mean the world to me.
That is the only favor I will ever ask you.
Thank you so much for your time.
The reason why I wanted to have this conversation
with all of you is because the subject matter of AI,
but more specifically AI agents, has occupied my free time for several weeks in a row.
And actually Amjad, when I started using Replit for me, it was a paradigm shift.
There was two paradigm shifts in a row that happened about a week apart.
ChatGPT released that image generation model where you could create any image.
It was incredibly detailed with text and all those things.
That was a huge paradigm shift.
And then in the same week, I finally gave in to try and figure out what this term
AI agent was that I was hearing all over the internet.
I had viber coding, I had AI agent.
I was like, I will give it a shot.
And when I used Replit 20 minutes in to using Replit, my mind was blown.
And I think that night I stayed up till three or four AM in the morning.
For anyone that doesn't know, Replet is a piece of software that allows you to
create software and pretty much any software you want.
So someone like me with absolutely no coding skills was able to build a website,
build in Stripe, take payment, integrate AI into my website,
add Google login to the front of my website and do it within minutes.
I then got the piece of software that I had built with no coding skills, sent it to my
friends and one of my friends put his credit card in and paid.
Amazing.
So I just launched a SaaS company with no coding skills.
To demonstrate an AI agent in a very simple way, I used an online AI agent called Operator
to order us all some water from a CVS around the corner.
The AI agent did everything end to end and people will be watching on the screen.
They put my credit card details in, it picked the water for me, it gave the person a tip,
it put some delivery notes in.
At some point, a guy is going to walk in.
He has not interacted with a human.
He's interacted with my AI agent.
And I just, the reason I use this as an example is, again,
it was a paradigm shift moment for me when I heard about agents
about a month ago, and I went on and I ordered a bottle of electrolytes.
And when my doorbell rang, I freaked out.
I freaked out.
Amjad, who are you?
And what are you doing?
So, I started programming at a very young age.
I built my first business when I was a teenager.
I used to go to internet cafes and program there.
And I realized that they don't have software
to manage the business.
I was like, oh, why didn't you create accounts?
Why don't you have a server?
It took me two years to build that pace of software.
And that sort of embedded in my mind, this idea that, hey,
like, you know, there's a lot of people in the world with
really amazing ideas, especially in the context where
they live in, that allows them to build businesses.
However, the main source of friction between an idea and
software, or call it an idea and wealth creation, is
infrastructure, is physical infrastructure, is meaning a computer in front of you, it
is an internet connection, it is the set of tools and skills that you need to build that.
If we make it so that anyone who has ideas, who wants to solve problems will be able to do it.
I mean, imagine the kind of world that we could live in
where no one can be, anyone who has merit,
anyone who can think clearly,
anyone who can generate a lot of ideas can generate wealth.
I mean, that's an amazing world to live in, right?
Anywhere in the world.
So with Replit, the company that I started in 2016,
the idea was like, okay, coding is difficult.
How do we solve coding?
And we built every part of the process,
the hosting, the code editor.
The only missing thing was the AI agent.
And so over the past two years,
we've been working on this AI agent that you can just similar,
since ChatGPT, this revolution with GenAI,
and you can just speak your ideas into existence.
I mean, this starts sounding religious.
This is like the gods, the myths that humans have created.
They used to imagine a world where you can
be everywhere and anywhere at once.
That's sort of the internet.
And you can also speak your ideas into existence.
And you know, it's still early.
I think Replet Agent is a fantastic tool.
And I think this technology is going to get so much more powerful.
Specifically, what is an AI agent?
I've got this graph, actually, here,
which I don't need to pass to any of you
to be able to see the growth of AI agents.
But this graph is Google search trend data.
This also resembles our revenue, too.
Oh, OK, right.
The water has arrived.
Hello.
Thank you.
You can come on in.
Can I have a go, please? Yes, it's 3951. Thank you. You can come on in. Can I have a go please?
Yes, it's 3951.
Thank you so much.
Thank you.
This is like a supernatural kind of power.
You conjured water.
I conjured water from my mind and it's shown up here with us.
And it clearly thinks we need a lot.
But just to define the term AI agent for someone
that's never heard the term before.
Yeah.
Yeah, so I assume most of the audience now
are familiar with ChatGPT, right?
You can go in and you can talk to an AI.
It can search the web for you.
It has a limited amount of tools.
Maybe it can call a calculator to do some additional
subtraction for you.
But that's about it.
It's a request response style.
Agents are when you give it a request
and they can work indefinitely until they achieve a goal
or they run into an error and they need your help.
It's an AI bot that has access to tools.
Those tools are access to a web browser like Operator,
access to a programming environment, say like Replet,
access to credit cards.
The more tools you give the agent,
the more powerful it is.
Of course, there's all these considerations
around security and safety and all of that stuff.
But the most important thing is that AI agent
will determine when it finished executing.
Today, AI agents can run for anywhere between 30 seconds to 30 minutes.
There's a recent paper that came out that's showing that every seven months,
the number of minutes that the agent can run for is doubling.
So we're at like 30 minutes now.
In seven months, we're going to be at an hour, then two hours.
Pretty soon, we're going to be at days.
And at that point, AI agent is doing labor,
is doing kind of human-like labor.
And actually, OpenAI's new model, 03, beat the expectation.
So it sort of doubled coherence over long horizon tasks in just three or four months.
So we're in this massive, and I mean, this looks, this exponential graph, you know,
that shows you the massive trend we're on.
Brett, give us a little bit of your background, but also I saw you writing some notes there.
There was a couple of words used there that I thought were quite interesting, especially
considering what I know about you.
The word God was used a few times.
Well, let me just say I'm an evolutionary biologist and probably for the purposes of
this conversation, it would be best to think of me as a complex systems theorist. One of the things that I believe is true about AI is that this is the first time that we
have built machines that have crossed the threshold from the highly complicated into
the truly complex.
And I will say I'm listening to this conversation with a mixture of profound hope and dread,
because it seems to me that it is obvious that the potential good that can come from
this technology is effectively infinite.
But I would say that the harm is probably 10 times.
It's a bigger infinity. And the question of how we are going to get to a place where we can leverage the obvious
power that is here to do good and dodge the worst harms, I have no idea.
I know we're not prepared.
So I hear you talking about agents and I think that's marvelous.
We can all use such a thing right away and the more powerful it is, the better the idea I hear you talking about agents and I think that's marvelous.
We can all use such a thing right away and the more powerful it is, the better.
The idea of something that can solve problems on your behalf while you're doing something
else is marvelous.
But of course, that is the precondition for absolute devastation to arise out of a miscommunication,
right? for absolute devastation to arise out of a miscommunication.
To have something acting autonomously to accomplish a goal, you damn well better understand what
the goal really is and how to pull back the reins in the event that it starts accomplishing
something that wasn't the goal.
The potential for abuse is also utterly profound.
You can imagine, just pick your dark mirror fantasy dystopia where something has been
told to hunt you down until you're dead and it sees that as a technical challenge.
So I don't know quite how to balance a discussion about all of the
things that can clearly come from this that are utterly transcendent. I mean, I do think
it is not inappropriate to be invoking God or biblical metaphors here. You know, you're
producing water seemingly from thin air. I believe that does have an exact biblical parallel.
So any case, the power is here.
But so too is the need for cautionary tales, which we don't have.
That's the problem is that there's no body of myth that will warn us properly of this
tool because we've just crossed a threshold that is similar in
its capacity to alter the world as the invention of writing.
I really think that's where we are.
We're talking about something that is going to fundamentally alter what humans are with
no plan.
Writing alters the world slowly because the number of people who can do it is tiny at first and remain so for thousands of years.
This is changing things weekly.
And that's an awful lot of power to just simply have dumped on a system that wasn't well regulated
to begin with.
Dan.
Yeah, so I'm an entrepreneur.
I've been building businesses for the last 20 plus years.
I'm completely well positioned between the two of you here, the excitement of
the opportunity and the terror of what could go on.
There's this image that I saw of New York City in 1900, and every single vehicle
on the street is a horse and cart.
And then 13 years later, the same photo from the same vantage point and every single vehicle on the street is a horse and cart. And then 13 years later, the same photo from
the same vantage point and every single vehicle on the street is a car. And in 13 years, all
the horses had been removed and cars had been put in place. And if you had interviewed the
horses in 1900 and said, how do you feel about your level of confidence in the world? The
horses would have said, well, we've been part of humanity for, you know, horse and hoof, hand and hoof for many, many years,
for thousands of years. There's one horse for every three humans. Like, how bad could it be?
You know, we'll always have a special place. We'll always be part of society. And little did the
horses realize that that was not the case, that the horses were going
to be put out of business very, very rapidly.
And to reason through analogy, there's a lot of us who are now sitting there going, hey,
wait a second, does this make me a horse in 1900?
I think a lot of people don't realize how massive these kind of technologies are going
to have as an impact.
One minute we're ordering water and that's cute.
And the next minute it can run for days.
And in your words, uh, it doesn't stop until it achieves its goal.
And it comes up with as many different ways as it could possibly come up
with to achieve its goal.
And in your words, it better know what that goal is.
I'm thinking a lot, Amjad, as Daniel's speaking about the vast application of AI
agents and where are the bounds?
Because if this thing is going to get incrementally smarter, well, incrementally
might be an understatement, it's going to get incredibly smart, incredibly quick.
And we're seeing this AI race where all of these large language models are competing
for intelligence with one another.
And if it's able to traverse the internet and click things and order things and write things
and create things and all of our lives run off the internet today, what can't it do?
It's going to be smarter than me. No doubt it already is. And it's going to be able to take
actions across the internet, which is pretty much where most of my professional life operates.
It's like how I build my businesses. Even this podcast is an internet product at the
end of the day. Because you can create, we've done experiments now and I can show you the
graphs on my phone to make AI podcasts. And they have, we've just managed to get it to
have the same retention as the driver CEO. Now with the image generation model, retention
as in viewers tension, the percentage of people that get to one hour is the same now.
So we can make the video, we can publish it, we can script it.
You can synthesize my voice.
Sounds like me.
So what, what is it going to be able to do?
And can you give me the variety of use cases that the average person might
not have intuitively conceived?
Yeah.
So I tend to be an optimist.
And part of the reason is because I
try to understand the limits of the technology.
What can it do is anything that we can,
any sort of set of human data that we can train it on.
What can it not do is anything that humans
don't know what to do because we don't have the training data.
Of course, it's super smart because it integrates massive amount of knowledge that you wouldn't
be able to read, right?
It also much faster.
It can run through massive amount of computation that your brain can't even comprehend.
Because all of that, they're smart, they can take actions, but we know the limits of what they can do
because we train them.
They're able to simulate what a human can do.
So the reason you were able to order the water there is because it was trained NSF data that
includes clicking on DoorDash and ordering water.
I applaud your optimism and I like the way you think about these puzzles, but I think
I see you making a mistake that we are about to discover is very commonplace.
So we have several different categories of systems.
We have simple systems, we have complicated systems, we have complex systems, and then
we have complex adaptive systems.
And to most of us, a highly complicated system appears like a complex system.
We don't understand the distinction.
Technologists often master highly complicated systems and they know, for example, a computer.
It's a perfectly predictable system inside.
It's deterministic.
But to most of us, it functions, it is mysterious enough that it feels like a complex system.
And if you're in the position of having mastered highly complicated systems, and you look at
complex systems and you think it's a natural extension, you fail to anticipate just how
unpredictable they are.
So even if it is true that today there are limits to what these machines can do based
on their training data, I think the problem is to see what's going to happen, you really
want to start thinking of this as the evolution of a new species that will continue to evolve.
It will partially be shaped by what we ask it to do, the direction we lead it, and it
will partially be shaped by things we don't understand.
So how does this computer that we have work?
Well, one of the things that it does is we plug them into each other.
Using language, it's almost as if you've plugged an ethernet cable in between human minds.
And that means that the cognitive potential
exceeds the sum of the individual minds in question.
Your AIs are going to do that.
And that means that our ability to say what they are capable of
does not come down to, well, we didn't train it on that data.
As they begin to interact, that feedback is going to take them to capabilities we don't
anticipate and may not even recognize once they become present.
That's one of my fears.
This is an evolving creature and it's not even an animal.
If it were an animal, you could say something about what the limits of that capability are. But this is a new type of biological creature, and it will become capable of things that
we don't even have names for.
Even if it didn't do that, even if it just stayed within the boundaries that you're
talking about, you mentioned about it having median level intelligence.
Well that by definition means 50% of the people on the planet are less intelligent
than AI. To a degree, it's almost as if we've just invented a new continent of remote workers.
There's billions of them. They've all got a master's or a PhD. They all speak all the
languages. Anything that you could call someone or ask someone over the internet to do their 24-7 and their 25 cents an hour, if that.
So if that really happened, if we really did just discover that there were a billion extra
people on the planet who all had PhDs and were happy to work for almost free, that would have a
massive disruptive impact on society. Society would have to rethink how everyone lives and works and gets meaning.
So like, and that's, that's if it just stays at a median level of intelligence.
Like it's, it's pretty profound.
I still think it's, it's a tool.
This is power that is there to be harnessed by entrepreneurs.
You know, I think that the world is going to get disrupted, right?
You know, I think that the world is going to get disrupted, right? And this post-war world that we created, where you go through life, you go through 12 years
of education, you get to college, and you just check the boxes, you get a job.
We can already see the fractures of that.
This American dream is perhaps no longer there.
And so I think the world has already changed.
But like, what are the opportunities?
Obviously there are downsides.
The opportunities is for the first time, access to opportunity is equal.
And I do think there's going to be more inequality.
And the reason for this inequality is because actually Steve Jobs made this analogy. It's like the best taxi driver in New York
is like 20% better than the average taxi driver.
The best programmer can be 10x better.
We say the 10x engineer.
Now the variance will be in the 1,000x.
The best entrepreneur that can leverage those agents could be better.
It could be a thousand times better than someone who doesn't have the grit,
doesn't have the skill, doesn't have the ambition, right?
So, so that, that will create a world.
Yes, there's massive access to opportunity, but there are people who will take.
Seize it.
And then there'll be people who don't.
I imagine it almost like a, um, a marathon race and AI has two superpowers.
One superpower is to distract people.
Um, such as tick tock algorithm.
That's right.
And the other superpower is to make you hyper creative.
So you become a hyper consumer, a hyper creator.
And in this marathon race, the vast majority of people have got their shoes tied together
because AI is distracting them.
Some people are running traditional race.
Some people have got a bicycle and some people have got a formula one vehicle.
And it's going to be very confronting when the results go on the scoreboard and you
see, oh, wait a second, there's a
few people who finished this marathon in about 30 minutes and there's a lot of us who finished
in like 18 hours because we had our shoes tied together.
And I can't understand if we've got equal opportunity why there's so much disparity
between how fast it, you know, and I'm using an analogy, but this idea that, you know,
someone like a lot of people are going to
start earning a million dollars a month, and a lot of people are going to say, hey, I can't
even get a job for $15 an hour.
There's going to be this kind of interesting wedge.
Well, but I hear in what both of you are saying, a kind of assumption that this will all be
done on the up and up.
And I do want to just, I am not a doomer.
I agree that the doomers are likely incorrect, that their fears are misplaced.
But I do think we have a question of a related rates problem.
You know, I said the potential for good here is infinite and the potential for bad is 10
times, right?
What I mean is there are lots of ways in which this obviously empowers people to do things
that they were going to be otherwise stuck in the mundane process of learning to code
and then figuring out how to make the code work and bring it to market
and all of that.
And this solves a lot of those problems.
And that's obviously a good thing.
Really what we should want is the wealth creation object as quickly as we can get there.
But the problem is, as much as that hyper creative individual is empowered to make wealth,
the person who is interested in stealing
may be even more empowered.
And I'm concerned about that at a pretty high level.
The abuse cases may outnumber the use cases.
And we don't have a plan for what to do about that.
Can I give you a quick introduction here, like the optimistic view?
OpenAI invented the first version of GPT came out in 2019.
2020 was GPT-2.
And so OpenAI, now they get a lot of criticism and lawsuit from Elon Musk that they're no
longer open source, right?
They used to be.
Right.
The reason is in GPT-2, they said, we are no longer going to open source this technology
because it's going to create opportunities for abuse, such as influencing elections,
stealing grandma's credit card, and so on and so forth.
Wouldn't you say, Brett, that it is kind of surprising how little abuse we've seen so
far?
I don't know how much abuse we've seen so far.
I don't know how any of us do.
And I also, even the example that you suggest, where chat GPT is no longer open source to
prevent abuse, I'm taking their word for it that that's the motivation, where as a systems theorist, I
would say, well, if you had a technology that was excellent at enhancing your capacity to
wield power, then open sourcing it is a failure to capitalize on that and that the most remunerative use is to keep it private
and then either sell the ability to manipulate elections to people who want to do so or sell
the ability to have it kept off the table for people who don't.
And I would expect that that's probably what's going on.
There's no, if you have a technology as transformative as this, giving it away for free is counterintuitive,
which leaves those of us in the public more or less at the mercy of the people who have
it.
So I don't see the reason for comfort.
We are at the dawn of this radical transformation of humans that by its very nature as a truly complex and
emergent innovation, nobody on earth can predict what's going to happen.
We're on the event horizon of something.
And the problem is, you know, we can talk about the obvious disruptions, the job disruption,
and that's going to be massive.
And does that lead some group of elites to decide, oh, well, suddenly we have a lot of
useless eaters, and what are we going to do about that?
Because that conversation tends to lead somewhere very dark very quickly.
But I think that's just the beginning of the various ways in which this could go wrong.
Without the Doomer scenarios coming into play. This is an uncontrolled experiment
in which all of humanity is downstream.
Yeah, so I was trying to make the point
that OPENAI has been sort of wrong
about the sort of how big of a potential for harm it is.
Like, you know, I think we would have heard about it
in the news, like the sort of how much harm it's
done and maybe some of it is working in the shadows, but like the few incidents that we've
heard about where the cause of LLMs, large language models, the technology that's powering
chat chippity, has been huge headliners.
Like New York Times talked about this kid that was perhaps goaded by some kind of chat
software that helps teenagers to be less lonely into suicide, which is tragic.
And obviously, these are the kind of safety and abuse issues that we want to worry about.
But these are kind of these isolated incidents.
And we do have open source large language models.
Obviously the thing that everyone talks about is DeepSeek.
DeepSeek is coming from China.
So what is DeepSeek's incentive?
Perhaps the incentive is to destroy the AI industry in the US.
When they released DeepSeek, the market tanked, the market for Nvidia, the market for AI,
and all of that.
But there is an incentive to open source. Meta is open sourcing Llama.
Llama is another AI similar to ChatGPT.
The reason they're open sourcing Llama,
and Zuckerberg just says that out loud,
is basically they don't want to be beholden to open AI.
They don't sell AI as a service.
They use it to build products.
And there's this concept in business
called commoditize your compliment.
Because you need AI's technology to run your service,
the best strategy to do is to open source it.
So these market forces are gonna create conditions
that I think are actually beneficial.
So I'll give you a few examples.
One is, first of all, the AI companies are motivated
to create AI that is safe so that they can sell it.
Second, there are security companies investing in AIs
that allows them to protect from the sort of malicious
acting of AI.
And so you have the free market, and we've always had that.
But generally, as humanity, we've been able to leverage the same technology to protect
against the abuse.
So I don't really understand this.
And maybe this is actually the exact discussion that you would expect between somebody at
the frontier of the highly complicated staring at a complex system and a biologist who comes from the land of the
complex and is looking back at highly complicated systems.
In game theory, we have something called a collective action problem.
And in the market that you're describing, an individual company has no capacity to hold
back the abuses of AI.
The most you can do is not participate in them.
You can't stop other people from programming LLMs in some dangerous way.
And you can limit your own ability to earn based on your own limitations of what you're
willing to do.
And then effectively what happens is the technology gets invented anyway, it's just that the dollars end up in somebody else's pocket.
So the incentive is not to restrain yourself so that you can at least compete and participate
in the market that's going to be opened.
And so the number of ways in which you can abuse this technology. Let's take a couple. What is to stop somebody from training LLMs on an individual's creative output and then
creating an LLM that can out-compete that individual, can effectively not only produce
what they would naturally produce over the course of a lifetime, but can extrapolate from it and can even hybridize it with the insights of other people so that effectively
those who have the LLM can train it on the creativity of others, not cut them in on the
use of that insight.
You can effectively end up putting yourself out of business by putting your creative ideas
in the world where they get sucked up as training data for future LLMs.
That is unscrupulous, but it's effectively guaranteed.
In fact, it's already happened.
So that's a problem.
And likewise, what would stop somebody from interacting with an individual and training
an LLM to become like a personalized con artist, something that would play exactly to your blind spots.
That does happen. That is starting to happen.
People get phone calls and it sounds like their daughter,
and I've lost my phone and I'm borrowing a friend's phone
and all of that sort of stuff.
What's interesting is that I think you make a really good point.
I worry about the impact on society,
and yet when I look at every single individual
who uses AI regularly, it almost has nothing but profoundly positive impact on their life.
I look at people like, I was just spending some time with my parents-in-law, who are
in their 70s and early 80s, and they use AI regularly for all sorts of things that they find incredibly
valuable and that improves the quality of their life.
I personally did an M&A, a mergers and acquisitions deal where I bought a company last year.
The AI was so powerful at helping that process.
The conversations were transcribed and they were turned into letters of intent and then
press releases and
legal documents. And we probably shaved $100,000 worth of costs and we sped up the whole process
and it was pretty magical to see how it could happen. With that said, you know, there's all
of these like, well, $100,000 worth of lawyers didn't get paid, right? So.
Well, what I want to know. That's okay. Yeah. And I'll agree. There's a lot of, well, $100,000 worth of lawyers didn't get paid, right? Well, what I want to know, yeah.
There's a lot of people upset about that.
But if we look back at the invention of the cell phone or the invention of the social
media platforms, there would be every reason to have exactly the same perspective.
I remember the beginning of Facebook and I remember the idea that suddenly the process that
used to afflict people where you would just lose touch
with most of the people who had been important to you,
that was not something that needed to happen anymore.
You could just retain them permanently
as a part of a diffuse social grouping that just simply grew
and value was added.
There's no end to how much good that did.
But what it did to us was profound and not evident
in the first chapter.
Say the same thing about the cell phones
and the dopamine traps and the way this has disconnected us
from each other, the way it has disconnected us from nature,
the way it has altered the very
patterns with which we think it has altered every classroom.
And those things I think are going to turn out to have been sort of minor foreshadowings
of the disruption that AI will produce.
So I agree with you today, the amount you can do with AI is a tremendous
amount of good, there's a little bit of harm, maybe that's something we need to worry about.
But as this develops, as we get to peer over the edge of this cliff that we're headed to,
I think we're going to discover that we can't yet detect the nature of the alteration that's
coming.
I just wanted to add some context to that because Amjad, I saw the interview you did in a newsletter in 2023 where you said, I wouldn't prepare for AGI in the alteration that's come. I just wanted to add some context to that, because Amjad, I saw the interview you did
in a newsletter in 2023 where you said, I wouldn't prepare for AGI in the same way that
I wouldn't prepare for the end of days.
It's effectively the end of days if the vision of AGI that some of these companies have comes
to bear, because it's called the singularity moment, because you can't really predict what
happens after that.
And so how would you even prepare for that?
And you want to prepare for the more likely world. And that world that you can actually predict
is a world where yes, there's like a massive improvements of technology. And
there's like insane compounding effects of technology. And it's pretty hard to
keep up. From that it appeared that in 2023, you were saying a similar thing to
Brett in terms of, we can't see around the corner here, because it is a
singularity.
Sorry, you also used AGI, artificial general intelligence.
It'd be interesting to know what your definition of AGI is.
So that-
Yeah.
So what I was saying there is even if I'm wrong, that you can actually
create a unbounded, seemingly conscious artificial intelligence that can entirely replace humans
and can act autonomously in a way that even humans can't act and can coordinate across
different AIs, different data centers to take over the world. So the definition of AGI is
artificial general intelligence, meaning that AI can acquire
new skills efficiently in the same way that humans can acquire skills.
Right now, AIs don't acquire skills efficiently.
They acquire a massive amount of energy and compute, entire data set of compute to acquire
these skills.
And I think there's, again, a limit on how general intelligence can get.
I think for most of the time, it will be lagging in terms of what humans are capable of doing.
The singularity is based on this concept of intelligence explosion.
So once you create an AGI, once you create an artificial general intelligence, that intelligence
will be able to modify its own source code and create the next version that is much more
intelligent and the next version creates the next version and the next, you know, for infinity,
right?
Within a week.
Within a week, perhaps within milliseconds at some point, right?
Because it might invent new computing substrate and all of that.
Perhaps it will use quantum computing.
And so then you have this intelligence explosion in a way that is impossible to predict how
the world is going to be.
And what I'm saying is, this is sort of like an end of time story.
Like how would you even prepare for that?
So if that's coming, like why would I spend my time preparing for, I think
it's unlikely to happen.
Can't see around the corner.
Yeah.
But I'd rather prepare what I was saying there.
I'd rather prepare for the more likely world in which we have access to tremendous power,
but the world's not ending and humans are still important.
I don't know why you say more likely.
I mean, I think the structure of your argument is sound.
You would prepare for the world that might happen for which you can prepare.
There's literally no point in trying to prepare for a world you can't predict at all.
The only thing you can do is just sort of upgrade your own skills and pay attention.
But if I have one message for the technologists, it's that your confidence about what this
can and cannot do is misplaced because you have, without
noticing, stepped into the realm of the truly complex.
In the truly complex, your confidence should drop to near zero that you know what's going
on.
Are these things conscious?
I don't know.
But will they be?
Highly likely they will become conscious and that we will not have a test to tell us whether that has happened.
Elon Musk predicts that by 2029, we will have AI with us, AGI, that surpasses the combined intelligence of all humans.
And Sam Altman actually wrote a blog three months ago that I read where he said, we are confident now, Sam Altman being the founder of OpenAero, which created ChatGPT,
we are confident now that we know how to build AGI as we have traditionally understood it.
When I put these things together, I go back to the central question of what role do humans have
in this sort of professional output in GDP creation if it's smarter than all humans combined,
if Elon Musk is correct there, and it's able to take actions across the internet and continue
to learn.
This is like a central question that I'm hoping I can answer today, which is like, where do
we go?
Yeah.
I mean, in my vision of the world, we're in the creative seat. We're sitting there where we are controlling swarms of intelligence being to do our job.
The way you run your business, for example, you're sitting at a computer, you have an
hour to work, and you're going to launch like a thousand SDR, sales representative, to go
grab as many leads as possible.
And you're generating a new update on Replet for your website here.
And then on this side, you're actually, you have an AI that's crunching data about your
existing business to figure out how to improve it.
And these AIs are kind of somehow all coordinating together. And
I am trying to privilege the human. Like this is my mission, is to build tools for people. I'm not
building tools for agents. And agents are tool. And so ultimately, not only do I think that humans
have a privileged position in the world and in the universe.
We don't know where consciousness is coming from.
We don't really have the science to explain it.
I think humans are special.
That's one side is my belief that humans are special in the world.
And another side, which I understand that the technology today, and I think for
the foreseeable future,
is going to be a function of us training data.
So there was this whole idea of like, what if chat GPT generates pathogens?
Well have you trained it on pathogens?
They were doing that kind of stuff in Wuhan.
I mean, a lot of the biotech companies are essentially using artificial intelligence. Like I can think of Abcellera, I think it's Abcellera in Canada.
Their whole business is using AI to create new vaccines using artificial intelligence
and bigger data sets than we've never had before.
And I know because I was very close to one of the founders of people involved in Abcellera.
So that work is going on anyway.
And if you think about Wuhan, that's, I mean, it's quite probably well known now that it came out of a lab and people working in a lab and in that scenario,
that had a huge impact and shut down the world.
What I'm, the central question I'd love to answer before I threw it back
open to the room is what jobs, because I know that you have this perspective,
what jobs are going to be made redundant in a world where I am sat here
as a CEO with a thousand AI agents.
I was thinking of all the names of the people in my company who are currently
doing those jobs.
I was thinking about my CFO when you talked about processing business data, my
graphic designers, my video editors, et cetera.
So what jobs are going to be impacted?
Yeah, all of those.
So I think-
And what do they do?
Maybe this is useful for the audience.
And what do they do? Maybe this is useful for the audience.
I think if your job is as routine as it comes, your job is gone in the next couple of years.
Meaning, in those jobs, for example, quality assurance jobs, data entry jobs, you're sitting
in front of a computer and you're supposed to click and type things in a certain order, operator and those technologies are coming on the market really quickly and
those are going to displace a lot of labor.
Accountants?
Accountants.
Lawyers?
Yes.
I mean, I've just pulled a ligament in my foot and they did an MRI scan and I had to
wait a couple of days for someone to look at the MRI scan and tell me what it meant.
I'm guessing that that's gone.
Yeah, I think the health care ecosystem is hard to predict
because of regulation.
And again, there's so many limiting factors
on how this technology can permeate the economy because
of regulations and people's willingness to take it.
But things, unregulated jobs that are purely text in,
text out.
If your job, you get a message and you produce some kind
of artifact that's probably text or images, that job is at risk.
So just to give you some stats here as well,
about 50% of Americans who have a college degree currently
use AI.
The stats are significantly lower for Americans
without a college degree.
So you can see how a splinter might emerge there
and that crack will widen because people like us
at this table are all messing around with it.
But my mom and dad in Plymouth in the Southwest,
rural England, haven't got, they just figured out iPhones.
So like I got them an iPhone and now they're like
texting me back, AI is a million miles away.
And if I start running off with my AGI agents, that gap is going to widen.
Women are disproportionately affected by automation, which is what you're
talking about there, with about 80% of working women in an at-risk job compared
to just over 50% of men, according to the Harvard Business Review, and jobs requiring
only a high school diploma have an automation risk of 80%, while those requiring
a bachelor's degree have an automation risk of 80%, while those requiring a bachelor's degree have an
automation risk of just 20%.
So we can see again how this will cause a sort of a...
It's also a huge risk with business processor outsourcing, which is essentially Western
countries sending jobs to India, to the Philippines.
Like at the moment, millions of people have been lifted out of poverty through the ability to do those kind of business process outsourcing jobs, and those are all going
to go.
But these, they're going to have a thousand employees.
But also, these people are actually already transitioning to training AIs.
Mm-hmm.
You know?
So there's going to be a massive industry around training AIs.
Until they're trained.
Well, no, you have to continuously acquire new skills.
And this is what I'm talking about.
I mean, this is again, if AI is a function of its data, then you need increasingly more
data. And by the way, we ran out of internet data.
I was actually thinking, interestingly, that this might not be great for the United States
or the UK, the Western world, because it is going to be a leveler when our kid in India
doesn't need a Silicon Valley office and $7 million in investment to throw up a software company basically.
Yeah.
My belief is that, so I have a more broad definition of AGI and the
singularity and for me, AGI is do we have artificial general intelligence in
terms of generally speaking, can AI just do stuff that humans used to be able to
do?
And we've already crossed that point.
We have this general intelligence that we can now all access and 800 million
people a week are now using chat GBT.
It's, it's exploded in the last three months.
And then to me, a singularity, uh, when the first tractor went out onto a farm,
for me, that was a singularity moment.
Uh, because everyone who worked in farming, it used to take 100 people to plough a field and now a tractor comes along and two guys
with the tractor can now plough the field in just as much time and now 98 people out of 100
completely out of a job. We also always underestimate a technology. If it does go
on to change history, when you look back through cars, horses, planes, the Wright brothers just
thought of a plane as being something that the army
could use, we had no idea of the application.
So someone said to me recently, they said, when it does change the world, we
underestimate the impact that it will change the world.
And I see people now with their estimations of AI and AI agents,
already incredibly optimistic.
And so if history holds here, we are undershooting the impact it's going to have.
And I think this is the first time in my life where the industrial revolution and allergies
seem to fall a little bit short.
Yeah.
Because we've never seen intelligence.
That's like, you could, I could think of this as a, I'm not an intelligent person on this,
but I could see that as like the disruption of muscles.
Whereas this is the disruption of intelligence.
Intelligence.
That's exactly the thing, is that what makes human beings special is our cognitive capacity
and very specifically our ability to plug our minds into each other so that the whole
is greater than the sum of the parts.
That's what makes human cognition special.
And what we are doing is we are creating something that can technologically surpass it without
any of the preconditions that make that a safe process.
So yes, we've revolutionized the world how many different times.
It's innumerable, but we've made farming vastly more efficient.
That's different than taking our core competency as a species and surpassing ourselves with
the product of our labor.
I think your question is a good one.
Then what does become of it?
Well, we only have one thing left.
We have our muscles, which we got rid of in the Industrial Revolution.
And then we have our intellect, which is this digital revolution.
Now we're left with emotions and agency.
So essentially, the agency idea, I think we used to judge people on IQ, and now IQ is
the big leveler.
And now going forward for the next 10 years, we're going to look at, are you a high agency
person or a low agency person?
Do you have the ability to get things done and coordinate agents? Do you have the ability to start businesses or give orders to digital
armies, you know, and essentially these high agency people are going to
thrive in this new world because they have this other thing that's been
bubbling under the surface, which is really interesting.
When you said agency is going to remain as an important thing, we're
sat here talking about AI
agents. And the crazy thing in a world of AI agents that have super intelligence is I can just
tell my agent, listen, I'm going on holiday, please build me a SaaS company that spots a market
opportunity, throw up the website, post it on my social media channel, I'll be in Hawaii. And this
new agentic world is stealing that too, because now it can take action.
In the same way that I can browse the internet,
I can call Domino's Pizza, speak to their agentic agent,
organize my pizza to be there before I even wake up.
And in fact, predictability, you know,
OpenAI now learns, and Sam Altman said that
they've expanded the memory feature.
So it knows every... It's knowing more and more and more about me.
It'll almost be able to predict what I want when I want it.
It'll know Steve's calendar, he's arriving at the studio,
make sure his cadence is on the side,
make sure his iPad has the brief on it, do the brief,
do the research for me, and everything else.
So it'll remember Brett's birthday,
so when I arrive there'll be something.
In fact, it's removing my need for any agency.
Yes, and again, I don't know how to make this point so that it occurs to people what I'm really
suggesting.
But today, maybe it's not conscious.
But let me put it to you this way.
If you're conscious, you started out as a child that wasn't.
And although this may not fully encapsulate it, you are effectively an LLM.
You go from an unconscious infant to a highly conscious adult.
And the process by which you do that has a lot to do with being trained effectively on
words and other things in an environment in exactly the way that we now train these AIs.
So the idea that we can take consciousness off the table, it won't be there till we figure
out how to program it in and we're safe because we don't know how consciousness works, I take
the opposite lesson.
We've created the exact thing that will produce that phenomenon and then we can have philosophers
debate whether it's real consciousness or it just behaves exactly as if it were and
the answer is those aren't different.
And the same thing is true for agency, especially if you've created an environment in which
these AIs are de facto competitors.
What you're effectively doing is creating an evolutionary environment in which they
will evolve to fill whatever niches are there.
And we didn't spell out the niches. So I have the sense we have, um, we have invited, we have, we have created
something that truly is going to function like a new kind of life.
And it's especially troubling because it speaks our language.
So that leads us to believe it's more like us than it is.
And it's actually potentially quite different.
So, by the way, he's the optimist here.
potentially quite different. So by the way, he's the optimist here.
He's so optimistic about all of them
and how they're gonna evolve.
Yes, it's amazing.
It's amazing technology.
I think it raised global IQ, right?
Like 800 million people.
Like 800 million people are that much more intelligent
and emotionally intelligent as well.
Like I know people who previously were very coarse and they kind of
rob people the wrong way.
They would say things in not so polite way.
And then suddenly they started putting their, you know, what they're saying
through chat GPT in order to kind of make it kinder and nicer and they're more
liked now.
And so not only is it making us more intelligent,
but also it allows us to be the best version of ourselves.
And the scenario that you're talking about,
I don't know what's wrong with that.
Like, you know, I would want less agency in certain places.
Like I would want something to help me not, you know, open up a peanut
butter jar at night. Right? You know, there are places in my life where I need more control
and I would rather ceded to some kind of entity that could help me make better choices.
I mean, unfortunately, even if there is some small group of elites that are able to go
to Hawaii while something else does the mundane details of their business building, we are
rather soon going to be faced with a world that has billions of people who do not have
the skills to leverage AI.
Some of them will be necessary for a time.
You're going to need plumbers.
But this is also not a long-term solution because not only are there not enough of those
jobs, but of course we have humanoid robots that once imbued with AI capacity, will also be
able to take, they'll be able to crawl under your house into the crawl space and fix your
plumbing.
So what typically happens when you have a massive economic contraction that arises from
the fact that a huge number of people are out of work is that the elites start looking at those people and thinking, well, we don't really need them
anyway.
And so the idea that this AI disruption doesn't lead us to some very human catastrophe, I
think, is overly optimistic and that we need to start preparing right now.
What are the rights of a person who has had whatever it is that they've invested in completely
erased from the list of needs?
Is that person responsible for not having anticipated AI coming?
And is it their problem that they are now starving and they're being eyed by others
as a useless eater?
I don't think so.
How is it different than when the, what's it called, the looming machine came and the
textile workers, the result of the Luddite revolution.
How is it different than any time in history when technology automated a lot of people
out of jobs?
I would say scale and speed.
That's how it's different.
And the scale and speed is going to result in an unprecedented catastrophe because the
rate at which people are going to be simultaneously sidelined, not just in one industry, but across
every industry is just simply new.
And it also did actually happen.
There was a, for the first 50 years of industrialization from like late 1700s to early 1800s, you
actually, the Charles Dickens novels are essentially
people coming from the farms who were displaced, arriving in cities, kids living on the streets.
The British decided to pick everyone up and send them over to Australia, which is where
I came from.
And you know, there was this massive issue of displacement.
I think we're going to go into a high velocity economy where rather than this long arc of
career that lasts 45 years, we're going to have these very fast careers that last 10
months to 36 months.
And you invent something, you take it to market, you put together a team of five to 10 people
who work together, you then get disrupted.
You come up.
Can I mention a story here?
There's an entrepreneur that uses Replet in a similar way.
His name is Billy Howell.
You can find him on YouTube, on the internet.
He would go to Upwork and he would find what people are asking for, different requests for
certain apps, technologies, then he would take the one they're asking for, put it
into Repload, make it an application, call them, tell them I already have your
application, would you pay $10,000 for it?
And so that's sort of an arbitrage opportunity that's there right now.
That's not arbitrage, it's theft.
How is it?
How is it theft?
It's theft.
You have somebody who has an idea that can be brought to market, and somebody else is
cryptically detecting it and then selling back their own idea to them.
Well they're paying them to do that.
They're saying, I will give you $500 if someone makes this for me.
Right.
But this is what I more or less think is going to happen across the whole economy is that,
yes, from this perspective, we can see that everybody is suddenly empowered to
build a great business.
Well, what do we think about the folks who are going to be displaced from the top?
What are they going to think about all these people building all of these highly competitive
businesses and are they going to find a way to do what venture capital has done or what
record producers have done? What they're going to do is they're going to or what record producers have done.
What they're going to do is they're going to take their superior position at the top
and they are going to take most of the wealth that is produced by all of these people who
have these ideas that in a proper market would actually create businesses for them and they're
going to parasitize them.
I think that with this introduction of AI and AI agents, old value has moved.
And now it's not going to be the case that the idea itself is the moat.
And it's not going to be the case that resources are the moat.
So in such a scenario, you still have to figure out distribution.
You still have to have, for example, like an audience.
So if you're a podcast, and now you have a million followers on Twitter,
you're in a prime position because you now have something that the guy with a
great idea with no audience has, you have inbuilt distribution. So I now think
actually much of the game might be moving to like, yeah, still about taste and
idea, but also the motives distribution.
Yeah. And speaking of adaptive systems, one of the adaptations that will happen is people
will seek humans and will seek proof of humanity.
Oh, I agree that authenticity is going to become the coin of the realm and anything
that can be faked or cheated is going to be devalued and things, you know, spontaneous
jazz or comedy that is interactive enough that
it couldn't possibly have been generated with the aid of AI.
Those things are going to become prioritized, spontaneous oratory rather than speeches.
That answers some of your questions.
It answers my question for the tiny number of people who are in a position to do those
things. Steven, you used the word moat, which I think is a really important word for entrepreneurs.
We have to have a moat.
We think a lot about moats and it's an industrial age.
A lot of people don't even know what a moat, what you mean by moat.
It's just the thing that's-
I often think about this idea of what are the moats that are left.
So to define how I define a moat, you've got a castle and it's got a small circle of water
around it.
And once upon a time, that circle of water defended the castle from attack and you can
pull up the drawbridge so nobody can attack you very easily.
It's a defense from something.
So it's your shield, it's your defense.
And once upon a time as an entrepreneur, I've got a software company in San Francisco called
Third Web and we raised almost $30 million.
We have a team of 50 great developers.
And much of our moat was, you can't compete with us
because you don't have the 50 developers
and the $30 million in the bank.
How much of that $30 million went to coding?
The vast majority of it.
I mean, what else are we going to do?
This is a good thing.
I think moats are a bad thing.
Let me make the argument there.
So everyone is looking for modes. You
know, for example, like one of the more significant modes is network effects. You know, so you
can't compete with Facebook or Twitter because it to move people from Facebook or Twitter,
you need to, it's the collective action problem. You need to move them all at once because
if one of them moves,
then the network is not valuable, they'll go back.
So you have this chicken and egg problem.
Let's say that we have a more decentralized way
of doing social networks.
That will remove the power of Twitter to kind of censor.
And I think you're at the other end of censorship, right?
And so part of my optimism about humanity
is that generally there's self-correction.
Democracy is a self-correcting system.
Free markets are largely self-correcting systems.
There are obvious problems with free markets
that we can discuss.
But take health, you know?
There is obesity epidemic, this period of time when companies ran loose making this
sugary, salty, fatty kind of snacks and everyone gorged on them and everyone got very unhealthy.
And now you have Whole Foods everywhere.
Today, people in Silicon Valley,
they don't go to bars at all.
They go to running clubs.
That's how you meet, that's how you go find a date.
You go to running clubs.
And so there was a shift that happened
because there was a reaction.
Obviously cigarettes is another example.
You were talking about phones and our addiction to phones
and I see a shift right now.
Like in my friend circle,
like people who are constantly kind of on their phones
is already kind of frowned upon.
And they don't wanna hang out with you
because you're constantly staring on your phone.
So there's always these reactions.
But the problem is you reference self-correction.
And I agree that there's actually an automatic feature of the universe in which the self-correction
happens.
You can't have a positive feedback that isn't reined in by some outer negative feedback.
But the corrections, the list of corrections involves things like YouPoint2, where people
become enlightened and they realize that they're doing themselves harm with either the sugar
that they're consuming or the dopamine traps on their phone, and they get better.
But also on the list of corrective patterns are genocide and war and parasitism.
The problem is these things are destructive of wealth. And so you allude to the superior fact of an open market without moats.
Presumably the benefit of that is that more wealth gets created because people aren't
kept from doing things that are productive.
I see that.
But then what is the product of all of this new wealth that is going to be generated by
a world empowered by AI?
Does it end up so highly concentrated that you have a tiny number of ultra elites and a huge number of
people who are utterly dependent on them? What becomes of those people? The learning
process, the self-correction process goes through harm in order to get to that more
enlightened solution. There's nothing that protects us from the harm phase
being so apocalyptically terrible that we get to the other side of it and we say, well,
that was a hell of a correction. Or maybe there's nobody there to even say that. Those
are also on the table.
It reminds me of a mousetrap where you see the cheese and we're going, oh my god, my
grandmother is going to be able to do some research and oh my god, my life is going to
get easier. So you head closer and closer to the cheese. And then snap.
And historically, if we look at all of the last 10,000 years, it's a very small number
of elites who own absolutely everything and a very large number of serfs and peasants
who have a subsistence living.
If the elites are too greedy and they freeze out the peasants at too high a level and they
try to use brutality-
Pitchforks. Yeah. greedy and they freeze out the peasants at too high a level and they try to use brutality in order to –
Pitchforks.
Yeah.
Eventually it comes back to haunt them.
And so what you get is a recognition that you need a system that does balance these things
and the West has the best system that we've ever seen.
It's one in which we agree on a level playing field.
We never achieve it, but we agree that it's a desirable thing.
And the closer we get to it, the more wealth we create.
But again, if AI empowers those with ill intent at a higher rate than it empowers those who
are wealth creating and pro-social, we may be in for a massive regression in how fair
the market that the West is.
Is that your top concern versus economic displacement?
I think they're the same thing.
How are they the same thing?
Because the economic displacement is going to start.
I don't know how many million people are going to be displaced from their jobs in the US.
Suddenly, we're going to have a question about whether or not we have obligations to them.
You agree with that, don't you?
Yes, but again, it's no pain, no gain.
We're going to go through a period of disruption and I think at the other end, the old oppressive
systems will be broken and we're going to create perhaps a fair world, but it's going
to have its own problems.
And what's the scale of that disruption in your estimation?
It's hard to say because there's this concept of limiting factors, like there is regulation,
there's the appetite of people today.
For example, the healthcare system is very resistant to innovation because of regulation.
And that's a bad thing.
On the regulation point, it's worth saying that when Trump
came into power, he signed in a new law, which
is called removing barriers to American leadership in AI,
which revokes previous AI policies that
were deemed to be restrictive.
And obviously, when you think about where
the funding is going in AI, it's going to two places.
It's going to America, and it's basically going to China.
That's the vast majority of investments.
So with those two in competition, any regulation that restricts AI in any way is actually self-sabotage.
And this is, you know, I live in Europe some of the time.
And it's already annoying to me that when Sam Altman and OpenAI released the 03 model,
this new incredible model, it's not in Europe because Europe has a regulation which prevents
it from coming to Europe.
So we're now at a competitive disadvantage, which Sam Altman's spoken about.
And more broadly on this point of disruption, I was quite unnerved when I heard that Sam
Altman's other startup was called WorldCoin.
And WorldCoin was conceived with the goal of facilitating universal basic income,
i.e. helping to create a system where people who don't have money are given money by the government just for being alive to help them cover their basic food and
housing needs, which suggests to me that the guy that has bought the biggest
air company in the world can see something that a lot of us can't see, which is there's
going to be- Poison and the cure.
Yeah.
There's going to need to be a system to just hand out money to people because they're not
going to be able to survive otherwise.
I disagree with that.
Which part do you disagree with?
I disagree that, first of all, that humans would be happy with UBI.
I think that, you know, core value of humans and be curious about the evolutionary reasons
is we want to be useful.
It's really important to know that a lot of the jobs that are at risk are the most high
status, highly paid jobs in the world.
Let's take the highest paid job in America, which is an anesthesiologist.
This is the highest paid job.
Highest paid salaried job. Sal is the highest paid job. Highest paid salary job.
Salary job, yeah. And the majority of that job is observing a patient,
knowing which type of medication would work best with their body,
giving them the exact right amount,
monitoring the impact of that on the body,
and then making slight adjustments.
The right technology and any nurse will be able to do that job.
And you might have one anesthesiologist on, on site supervising 10, 20, 30, 40 wards.
And the technology is, you know, doing the job, but that one person is there just to
kind of supervise if something went wrong or if there was an ethical dilemma.
What's wrong with that?
I mean, if, if the precision is better where there's nothing wrong with that,
except for the fact that a lot of people, hundreds of thousands of people have
spent their entire life training to be that they get an enormous amount of
purpose and satisfaction about the fact that that's their career, that's their job.
They have mortgages, they have houses, they have status and that's about to go away.
It's high spade jobs, maybe you should start saving.
Well, I mean, but you're talking about people who have done vital work, highly specialized
work and are therefore not in a great position to pivot based on
the invention of a technology that they didn't see coming because frankly, I mean, in the
abstract, maybe we all saw AI coming somewhere down the road, but we did not know that it
was going to suddenly dawn.
And we do have to figure out what to do with those people.
It's not their fault that they've suddenly become obsolete. And it's inconceivable that people will accept this.
It is fundamentally incompatible with our nature.
We have to have things to strive for.
And, you know, you can sustain life that way, but you cannot sustain a
meaningful existence.
And so it's a short-term plan at best.
Let's talk about meaning. On that point of job displacement, this is already happening.
Klarna's CEO, who has been on this podcast before, a great guy, said on a blog post that they published on Klarna's website
saying that they now have AI customer service agents handling 2.3 million chats per month,
which is equal to having to hire 700 full-time people to
do that.
So they've already been able to save on 700 customer service people by having air agents
to do that.
And they actually got rid of those 700 jobs, right?
I don't have that information in front of me, but I'll have a look.
I'll throw it up on screen for anyone that wants context on that.
But that's already happening.
This isn't some anesthet necrotologist or something.
And these aren't high paid people in every case.
We've done something similar by the way.
We've, we've internally, we've replaced that function for 70%.
I mean, our company, we're 65 people and you know, we, you know, make millions
per head, you know, we make millions per head.
You know, so it's a.
Are you going to need to hire more people to get up to?
I think so, but we're hiring slowly.
Like, you know, we're using customer support AI
and that meant that we need less customer support.
And we're trying to leverage AI as much as possible.
The person in HR at Replit writes software using Replit.
So I'll give you an example.
She needed OrgChart software.
And she looked at a bunch of them, got a lot of demos,
and they're all very expensive,
and they're missing the kind of features that she wants.
For example, she wanted like version control.
She wanted to know when something changed
and to go back in history.
So she went into Repl.it and in three days she got exactly the kind of software that she wants.
And what was the cost?
You know, perhaps $20, you know, something like that.
Once.
$30.
Once, right?
And how many employees in HR do we need?
Right now we have two.
If they're highly levered like that, maybe we do
not need a 20 HR team.
On this point of meaning, I've heard so many billionaires in AI
describe this as the age of abundance. And I'm not
necessarily sure if abundance is always a great thing. Because,
you know, when we look at mental health, and we look at why, how
people derive their meaning and their
purpose in life, much of it is having something to strive towards and some struggle in a meaningful
direction to you.
This is maybe adjacent, but when there was a study done, I think it was in Australia,
where they looked at suicide letters, and in the suicide letters, the sentiment of men
in those suicide letters was they didn't feel worthy.
They didn't feel like they were worth it.
They didn't feel like they were needed by their families. And this is much of what caused their
psychological state. And I wonder in a world of abundance where we, you know, a lot of these AI
billionaires are telling us that we're going to have so much free time and we're not going to need
to work if there is at all going to be a crisis of meaning, a mental health problem.
I mean, there already is, and it doesn't require AI and it's going to get worse.
I don't know what to do about it because essentially as human beings, we are built, like all organisms, to find opportunity and figure out how to exploit it.
That's what we do.
And the world you're describing is really the opposite of that. It's one where you're effectively having your biological needs
at the physiological level satisfied and there isn't an obvious place for your spare time,
if that's what you end up with, to be utilized in something that there's no place to strive.
to be utilized in something that there's no place to strive. And I do imagine almost at best what would happen is you have people who are being sustained
by a universal basic income and then parasitized whatever currency they have to spend, somebody
will be targeting it and they will be targeting it with an AI augmented
system that spots their defects of character.
I mean, again, we're already living in this world, but it will be that much worse when
the AI is figuring out what kind of porn to target you with specifically.
It's a nightmare scenario.
And I do think it would be worth our time as a species to start considering if we are
about to find ourselves in this situation and we find some way of dealing with the basic
needs of the large number of people who are going to be sidelined, what would a world
have to look like in order for them to have real meaning, not pseudo meaning, not something
that superficially a video game is not meaning,
even if it feels very meaningful in the moment. I think that would be a worthy investment for us
to figure out how to produce it. But frankly, I'm not expecting us to either have that conversation
or get very far down that road. I think it's much more likely that we will squander the wealth
dividend that will be produced by AI.
Interestingly, you also see in Western countries that when we get more abundance, we start having less kids.
And we're already seeing this sort of population decline in the Western
world, which is kind of scary.
And I think it's often associated with affluence.
The more money someone makes, the less likely they are to want to have
children, the more they try and protect their freedoms.
But also on this point of AI, relationships are hard. My girlfriend is happy sometimes
and not happy other times. And I have to go through that struggle with her of working
on the relationship. Children are hard. And if we are optimizing ourselves, and much of
the reason that I sustain the struggle with my girlfriend is, I'm sure, from some evolutionary reason, because I want to reproduce and I want to have kin.
But if I didn't have to deal with the struggle that comes with human relationships, romantic or platonic, there's going to be a proportion of people that actually choose that outcome.
And I wonder what's going to happen to birth rates in such a scenario, because we're already struggling. We're already in a situation where we used to be having five children per woman
in the 1950s to about two in 2021. And we're seeing a decline. If you look at South Korea,
their fertility rate has fallen to 0.72, the lowest recorded globally. And if this trend continues, the country's population could half by 2,100.
So yeah, relationships, connections.
And also, I guess we've got to overlay that with the loneliness epidemic, which is
they promised us social connection when social media came about.
When we got Wi-Fi connections, the promise was that we would become more connected.
But it's so clear that because we spend so long alone,
isolated, having our needs met by UberEats drivers
and social media and TikTok and the internet,
that we're investing less in the very difficult thing
of like going and making a friend
and like going and finding a girlfriend.
Young people are having sex less than ever before.
Everything that is associated with the difficult job of making
in real life connections seems to be falling away.
I'll make the case that everything that we've discussed here, all the
negative things around loneliness, around meaning, they're already here.
And I don't think blaming technology for all of it
is the right thing.
Like, I think there are a lot of things that happened
because of existing human impulses and motivations.
Well, I wanted to go back to where you started,
because I do think that this maybe
is the fundamental
question.
Why is it that we are already living in a world that is not making us happy?
And is that the responsibility of technology?
And I don't think it's exactly technology.
Human beings among our gifts are fundamentally technological, whether we're talking about
quantum computing or flint-napping an arrowhead.
What has happened to us that has created the growing, spreading, morphing dystopia is a
process that Heather and I in our book, A Hunter Gatherer's Guide to the 21st Century,
call hyper novelty. Hyper novelty is the fact of the rate of change outpacing our capacity to adapt to change.
And we are already well past the threshold here where the world that we are young in
is not the world that we are adults in.
And that mismatch is making us sick across multiple different domains.
So the question that I ask is, is the change that you're talking about going to reduce
the rate of change, in which case we could build a world that would start meeting human
needs better, open opportunities for pursuing meaningful work?
Or is it going to accelerate the rate of change, which is, in my opinion, guaranteed to make
us worse off? So if it was a one-time shift, AI is going to accelerate the rate of change, which is, in my opinion, guaranteed to make us worse off?
So if it was a one-time shift, AI is going to dawn, it's going to open all sorts of new
opportunities, there's going to be a tremendous amount of disruption, but from that we'll
be able to build a world.
Is that world going to be stable or is it going to be just one event horizon after the
next?
If it's the latter, then it effectively says what
it does to the humans, which is it's going to dismantle us.
When I look out at society, I go, okay, it's having a negative impact. When I look at individual
use cases, it's having a profoundly positive impact, including for me, it's having a very
positive impact. So it's one of these things where I wonder, what is it that we need to teach people at
school so that they understand the world that we're going into?
Because one of the biggest issues that we're having is that we're sending kids to school
with this blueprint, this template that they're going to have this long arc career that no
longer exists, that essentially we're treating them like learning
LLMs.
And we're saying, okay, we're going to prompt you, you're going to give us the right answer,
you're going to hallucinate it if possible.
And then we go, okay, now go off into the world and they go, oh, but wait a second,
I don't know how money works, I don't know how society works, I don't know how my brain
works, I don't know how I'm meant to handle this novelty problem.
I'm not sure how to approach someone in a, in a social situation and ask
if they want to go on a date.
Um, so all the important things that actually are the important milestones
that people want to be able to hit and that technology can actually have an
impact on, we get no user manual.
So I think one of the biggest things that has to happen is we have to equip,
uh, young people all through school to actually prepare them for the world that's coming or the world that's
here.
Well, on the one hand, I think you outlined the problem very well.
Effectively, we have a model of what school is supposed to do that, you know, at best
was sort of a match for the fifties or something like that.
And it woefully misses the mark with respect to preparing people for
the world they actually face.
If we were going to prepare them, I would argue that the only toolkit
worth having at the moment is a highly general toolkit.
The capacity to think on your feet and pivot as things change is the only game in town
with respect to our ability to prepare you in advance.
Maybe the other auxiliary component to that would be teaching you what we know, which
is frankly not enough, about how to live a healthy life.
If we could induce people into the kinds of habits of behavior and consumption
of food and then train them to think on their feet, they might have a chance in the world
that's coming. But the fly in the ointment is we don't have the teachers to do it. We
don't have people who know and that is the question is could the AI actually be utilized
in this manner to actually induce the right habits of mind for people to live in that
world.
I spent a lot of time in education technology.
One thing that is, as we say on the internet, a black pill about education in general, education
intervention, is there's a lot of data that shows that there are very little
interventions you can make in education to generate better outcomes.
And so, you know, there's been a lot of experiment around pedagogy,
around, you know, how to configure the classroom
that have resulted in very marginal improvements.
There's only one intervention, and this has been reproduced many times, that creates two
sigma two standard deviation positive outcomes in education, meaning you're better than 99%
of everyone else.
And that is one-on-one tutoring.
I was going to say smaller classrooms and personalization. One-on-one tutoring. I was going to say smaller classrooms and personalization.
One-on-one tutoring.
Yeah.
And by the way, if you look, someone also did a survey of all the geniuses, the understanders
of the world and found that they all had one-on-one tutoring.
They all had someone in their lives that took interest in them and tutored them.
So what can create one-on-one tutoring optionary for every child in the world.
AI.
My kids use it and it's incredible.
As in like, um, they're interacting and it's adapting to their speed.
And, um, it's giving them different analogies to work with.
So like, you know, my son was learning about division and it's asking him to smash glass and how many pieces he smashes it into with this hammer.
And, you know, and it's saying things like, no, Xander, go for it.
Really smash it.
And, um, and he's loving it.
Right.
Is that synthesis?
Yeah.
Yeah.
I am an investor in this company.
Oh, it was, it was great.
It was great to watch that simulated one-on-one tutoring because it's talking to him.
It's asking him questions.
Right.
You're an educator.
You, uh, spent much of your life teaching
people in universities. How do you receive all of this?
Well, on the one hand, I agree that the closer to one to one,
you get the better. But I also personally believe that zero to
one is best. And what I mean by that is part of what's gone wrong with our educational system is
that it is done through abstraction.
And effectively, the arbiter of whether you have succeeded or failed in learning the lesson
is the person at the front of the room.
And that's okay if the person at the front of the room is truly insightful.
And it's terrible if the person at the front of the room is lackluster, which happens a
lot.
So what doesn't work that way is interaction with the physical world in which nobody has
to tell you whether you've succeeded or failed.
If you're faced with an engine that doesn't start, you can't argue it into starting.
You have to figure out what the thing is that has caused it to fail and then there is a
great reward when you alter that thing and suddenly it fires up.
So I'm a big fan of being as light-handed as possible and as concrete as possible in
teaching.
In other words, when I've done it and and not just with students, but with my own children,
I like to say as little as possible, and I like to let physical systems tell the person
when they've succeeded or failed.
That creates an understanding you can extrapolate from one system to the next, and you know
that you're not just extrapolating from one person's misunderstanding, you're extrapolating from the way things actually work.
So I don't know if AI can be leveraged in that context.
My sense is there's probably a way to do it, but one would have to be deliberate about
it.
Especially with robotics and humanoid robots.
Actually that is the place where you can do this is with robotics. That it seems to me, yeah.
Oh, well, robotics will teach you the physical computing part of it.
And then the question is, how do you infuse this with AI so that it is,
that it provokes you out of some eddy where you're caught and
moves you into the ability to solve some next level problem that you wouldn't have found on your own.
What do you think should be taught in the classroom with everything that you now know?
Well, you're all fathers here.
You will have your own children.
So it's a good question.
Yeah.
How old are your kids?
How old are your kids?
Three and five.
19 and 21.
And six, seven and 10.
My children are very young, but we already do use AI and I sit down with them in front
of Rappler and we generate ideas and make games.
And I would say, you know what Brad said about generality is very important, the ability
to pivot and kind of learn skills quickly.
Being generative is very, very important.
Having a, you know, a fast pace of generating ideas
and iterating of those ideas.
We sit down in front of chat GPT
and my kid imagines scenario.
What if there's a cat on the moon?
And then, you know, what if the moon is made of cheese?
And what if there's a mouse inside it?
And so we keep generating these variations of these different ideas.
And I find that makes them more imaginative and creative.
Rule number one that I tell my kids is stay away from porn at all costs.
I'd rather you have a drug problem than a porn problem.
And I actually mean that.
I think porn is more dangerous to the human being as bad as a drug problem is.
But when we get to the question of how to confront the world and the things that you're
going to be expected to do in the workplace and all of that, my point to them is you are
facing the dawning of the age of complex systems that you are going to have to interact with.
And in the age of complex systems, you have to understand that you cannot blueprint a
solution and you have to approach these systems with a upgraded toolkit of humility because
the ability of the system to do something you don't predict
is much greater than a highly complicated system.
So you have to anticipate that and be very sensitive to the fact that what you intended
to happen is not what's going to happen.
So you have to monitor the unintended consequences of whatever your action is and that there
are really two tools which work, one of which you just mentioned, which is the prototyping. You prototype things.
You don't imagine that I know the solution to this and I'm going to build it. You imagine
I think there's a solution down there. I'm going to make a proof of concept and then
I'm going to discover what I don't know and I'm going to make the next version, discover
what I don't know. And eventually you may get to something that actually truly accomplishes
the goal. So prototyping is one thing.
And also, instead of using the blueprint as the metaphor in your mind, navigate.
You can navigate somewhere.
And you know that the way I think of it is a surfer is in some ways mastering a complex
system, but they're not doing it by planning their days surfing down the waves.
You can't do that.
What you can do is you can be expert at absorbing feedback and navigating your way down the
wave and that that's the right approach for a complex system.
Nothing else is going to work.
And so I guess the final piece is general tools, always no specialization. This is the age of generalists and invest in those tools and they will pay.
So the guiding philosophy for me is to produce high agency generalists.
So ultimately I want them to be motivated self starters and have a wide general toolkit.
I imagine them very much what you imagine, which is instructing robots,
instructing agents, coming up with ideas.
And I imagine them having a very high velocity life where they may be writing a book, organizing
a festival, having a podcast, starting a business and being part of somebody else's business
all at once.
As they are of the ADHD.
Yeah.
So the high agency generalist is the kind of guiding philosophy.
Some of the things that we do is like, we do chess, we do Brazilian
jiu-jitsu, we do dancing, we do acting classes, playing in nature,
entrepreneurship, understanding that you can start a lemonade.
We just did lemonade stands, which was amazing.
We sold lots of lemonade on the street.
So those kind of things
and jumping from one thing to the next thing, but also trying to avoid too many screens and
forcing them into making stuff from what's going on around the house. Some distinctions that we try
and give them is the difference between creating and consuming, because I think AI has this
superpower of making you a hyper consumer or a hyper creator.
And if you don't understand the distinction between creation and consumption,
you end up falling into the consumption trap, whether it be porn or just news or things that feel like you're productive, but you're actually just consuming stuff.
Wouldn't that be the most successful AI? The one that plays with my dopamine the most.
successful AI, the one that plays with my dope and mean the most. Yeah, and makes you think that you're achieving something when you're actually just consuming
something. So trying to give them the understanding that there is this difference in their life
between creation and consumption and to be on the creation side.
I started my first business at 12 years old and I started more businesses at 14, 15, 16, 17 and 18 and at that
time what I didn't realize is that being a founder with no money meant that I also had to be the
marketeer, the sales rep, the finance team, customer service and the recruiter. But if you're
starting a business today thankfully there's a tool that wears all of those hats for you,
our sponsor today which is Shopify.
Because of all of its AI integrations using Shopify feels a bit like you've hired an entire growth team from day one, taking care of writing product descriptions, your website design and
enhancing your product's images. Not to mention the bits you'd expect Shopify to handle like the
shipping, like the taxes, like the inventory. And if you're looking to get your business started, go to Shopify.com slash Bartlett and sign up for a $1 per month
trial. That's Shopify.com slash Bartlett. The thing that we, I think, all agree on is
that this is inevitable. Do you agree with that, Brad?
I think it's sad that it is inevitable, but at this point it is.
What part of it do you find sad?
We have squandered a long period of productivity and peace in which we could have prepared for this moment. that. And our narrow focus on competition has created a fragile world that I'm afraid
is not going to survive the disruption that's coming. And it didn't have to be that way.
This was foreseeable. I mean, frankly, the movie 2001, which came out the year before I was born, anticipates some of these problems.
And we treated it too much like education, I mean, like entertainment and not enough
like education.
So we are now, we've had the AI era opened without a discussion about its implications
for humanity. There is now, for game theoretic reasons, no way to slow that pace because as you point
out, if we restrain ourselves, we simply put the AI in the hands of our competitors.
That's not a solution, so I don't advocate it.
But there's a lot more preparation we could have done.
We could have recognized that there were a lot of people in jobs that were about to be obliterated. And we could have thought deeply about what the moral implications
were and what the solutions at our disposal might have been. And having not prepared,
it's going to be a lot more carnage than it needed to be.
Amjad, I heard you say a second ago that what we should be talking about is how we deal
with job displacement. Do you have any theories?
If you were prime prime minister or president of the world and your job was to deal with job displacement, let's just say in the United States, how would you go about that?
The first thing I would do is teach people about these systems, whether it's programs on the TV or outreach or what
have you, just trying to get people to understand how chat GPT works, how these algorithms work.
And as the new jobs arrive, I think there's going to be an opportunity for people to be
able to detect that this job know, the, this job
requires this set of skills.
And I, I have this, this kind of experience.
And although my experience are potentially outdated, I can repurpose
that experience to do that job.
I'll give you an example.
A teacher, his name is Adeel Khan.
You know, he started using at the time GPT-3 and felt like it does amazing work as a tools
for teachers or even potentially a teacher itself.
So he learned a little bit of coding and he went to a replica and he built this company.
And just two years later, they're worth hundreds of millions of dollars.
Obviously, not everyone will be able to create businesses of that scale.
But because you have an experience in a certain domain, you'll be able to build the next
iteration of that user technology.
So even if your job was displaced, you'll be able to figure out what potentially comes
after that.
So I think people's expertise that they built,
I don't think they're all for waste,
even if your job went away.
You can never really predict what jobs are coming.
I mean, I think of this crazy situation
where I tell my grandfather,
what is a personal fitness trainer?
And his mind would be blown by this idea that,
well, okay, I don't really want to go to the gym. So I have to make an appointment and pay someone
to go to the gym and meet with me there. And then he stands there and tells me to lift heavy things
that I don't really want to lift. And then he counts them and tells me that I've done a good job.
And then I put the heavy things down. And then at the end of that, I feel really good. And I pay him
a bunch of money.
My grandfather would be like, what on earth have you been scammed?
So we can never predict what this future of jobs would look like.
Even just 20, 30, 40 years apart, the jobs rapidly and convincingly just can morph into
something else.
I think it's very dangerous the idea that we need to focus on skills.
I think the future is not in skills.
Skills are being replaced.
It's this idea that the education system has to stop being compartmentalized and has to
be a lifelong learning approach.
The Department of Education needs to be seeing people as lifelong learners who are constantly
disrupted need re-education.
Interesting.
That that's going to be a thing. The Department of Education needs to start as a kid and go right through to maybe 70.
Does the Department of Education have a role anymore at all?
Depends on your definition of education.
I think if you're trying to teach kids to remember facts and figures from a history
book, then no.
But if it's about coaching, mentoring, being displaced,
finding the next thing, and maybe if it's AI driven
and all of those kind of things,
then it's a different paradigm shift
around what education is and what its purpose is.
And if we see it as a fluid thing
where we wave into an opportunity
and then wave back into education,
spotting a new opportunity and then back here.
If we're learning rather than skills, but we're learning tools, so it's a tools-based
education as opposed to a skills-based education.
The purpose of education for most of human history was about virtue, about becoming a
great person who had good judgment and who had good values, and we don't really do much
of that anymore.
But I think if we essentially said, if we get back to what is, if we ask the
question, what is the purpose of education and where does it fit in our lives?
And what timeframe does it go for?
And then we just trust that people are going to come up with weird and wonderful jobs.
You know, this is sounds crazy, but also, and this is a weird analogy.
My cat is incredibly happy.
How'd you know?
Well, it demonstrates all the characteristics of being a happy cat.
And it lives in a world of super intelligence as far as it's concerned.
So there's this house and food just magically happens.
It has no idea that there's this Google calendar that runs a lot of things that happen around it, the food gets delivered,
money is magically made by something
that is inconceivably more intelligent than the cat.
And yet the cat has evolved to be living this life
of purpose and meaning inside the house.
And as far as it's aware, it's got a great life.
But you have the power at any moment, if you're having a bad day, to do something
not so pleasant to that cat, and it can't really reciprocate that.
Exactly.
But what's in it for me to hurt the cat?
Because in this analogy-
You might want to move house and the landlord doesn't allow cats.
So you've got a decision to make.
Okay.
Yeah.
There are things that the cat is highly disrupted by due to no fault of the cat.
I get it.
But as far as cat existence goes and the history of cats, if you were to ask that cat, do you
want to trade places with any of the other cats that came before you?
It would probably say, I don't want to take the risk because all the other cats had to
fend for themselves in a way that I don't have to.
It's very possible that we end up living in a life a lot like the house cat in the sense
that from our perspective, we're extremely like we're having very interesting lives and
purpose meaning and just there's this massive higher intelligence that's just running stuff.
And we don't know how it works, but it doesn't really matter how it works.
We are the beneficiaries of it,
and it's doing important things,
and we're enjoying being house cats in its life.
I have a few things to say about this.
One, I'm pretty sure your cat's not as impressed
with your capacity as you are, or as you think he is.
I just know cats well enough to be pretty sure of that.
It looks down on me.
Yeah, you're right.
I think it's a fair point that there is an existence.
And actually, pets really do have it.
If they have loving owners, they really
do have it pretty great.
And I would also point out that there's a way in which we
already are this way.
Most of us do not understand the process that
results in electricity coming out of the walls of
our house or the water that comes out of the tap.
We're pretty much okay with the fact that somebody takes care of that and we can busy
ourselves with whatever it might be.
But the place that I find something troubling in your description is that you say that the
nature of what we do is to deal with the fact that jobs are always being
upended.
That's a very new process.
That is the hyper novelty process.
It used to be that it was only very rarely that a population had a circumstance where
you didn't effectively do exactly what your immediate ancestors did, right?
In general, you took what the jobs were, you picked something that was suited to you, and
you did that thing.
Intergenerationally.
Intergenerationally.
And the point is, we've now gotten to the point where even within your lifetime, what
is possible to get paid for is going to shift radically in ways that nobody can predict.
And that is a dangerous situation.
Like probably every two years, like two or three years.
Right.
And so maybe there's some model by which we can surf that wave and you can learn a generalist
toolkit and that your survival doesn't depend on your being able to switch up every two
years and never miss a beat,
or maybe we can't.
But I do think it is worth asking the question, if the rate of technological change has taken
us out of the normal human circumstance of being able to deduce what you might do for
a living based on what your ancestors did and put us in a situation where what your
ancestors did is going to be perfectly irrelevant no matter what.
That is effectively a choice that has been made for us.
We could choose to slow the rate of change so that we would live in some kind of harmony
where our developmental environment and our adult environment were a match.
As a biologist, I would argue if we don't do something like that, this is a matter of
time.
How would we slow the rate of change?
Well, I mean, you can be the Amish, right?
You can be the Amish and live in your own communities.
And I would assume some people would want that.
Well, you know, when Heather and I wrote our book, I wanted the first chapter to be, are
the Amish right?
And the answer is, they can't be exactly right because they picked an arbitrary moment
to step off the escalator.
But are they right that there's something dangerous about this continuing pattern of
technological change?
Clearly they are.
What do the Amish do for anyone that doesn't know?
The Amish live as if it was, what, 1880?
1950 or something.
They live in a, they don't use cars.
They also have phones, but they do not have electricity.
Basically they voluntarily accept a technology.
They're basically a Luddite community and they have turned
out to fare surprisingly well against many of the things that have upended
modern folks.
COVID one of them, right?
Yeah, COVID they did beautifully.
Quite happy people.
Very low autism rates.
They have all sorts of advantages.
So anyway, I'm not arguing that we should live like the Amish.
I don't see that, but I do think the idea that they had an insight, which was you need to step off that escalator
because you're just going to keep making yourself sicker is probably right.
Now maybe this is a one-time shift.
We've stepped over the event horizon.
We are going to be living in the AI world.
And maybe if we're careful about it, we can figure out how to turn that landscape of infinite possibility that you're describing into a place that doesn't change, that you always
have the opportunity to decide what needs to be done, but that living over that event
horizon is not an ever-changing process.
It's just the next frontier.
I do want to also propose or ask the question when we talk about our hyper-changing world,
isn't it harder for older people to learn because of the way that the brain works in
terms of processing speed and memory flexibility?
So I was wondering if you're going to get a situation where like my father, because
of his brain and the reduced memory flexibility and processing speed that happens when you're older, is going to struggle significantly more than
my niece who can seem to learn, I mean, my niece knows five languages and she's
seven or something crazy like that.
Five languages.
But I mean, the brain is much more plastic, isn't it?
So it's, and that goes back to our evolutionary psychology, which you know,
much evolutionary history, which you know much more than I do about of we're meant
to learn our lessons when we're young, use that information for a lifetime.
But if that information is changing quickly. Well, that's, I mean, this is exactly what I'm
pointing to. It is not normal for your developmental environment to fail to prepare you for your adult
environment. The normal thing is as a young person, you take on ever more of the responsibilities
of the adult environment and then at some point, you know, in a properly functioning
culture, there's a rite of passage. You go into the bush for 10 days, you come back with,
you know, a large, you know, game animal and now you're an adult and you take that program
that you've been building and you activate it and that is normal and you're a lot happier person, you're a lot
more fulfilled if your life has that kind of continuity to it.
And I'm not against the idea that we have enabled ourselves to do things that can't
be done if that's the limit, but we have also harmed ourselves gravely.
I would like to somehow pry apart our ability to improve our well-being from our self-inflicted
wounds that come from this never-ending pace of change.
I don't know if it's possible, but I think it's a worthy goal.
Something amusing, I don't know if it's exactly
a counterpoint, but during COVID, especially,
and through the recent technological change,
some people have started living closer
to the more ancestral environment.
So people whose jobs are online,
some of my friends like went and built communities
like collectives where they live
and they create farms and they have like an email job.
They do their email jobs for five hours and go out
and they all have children and it's a fascinating life.
And there was so much rethinking in Silicon Valley about how we live.
And there's a bunch of startups that are trying to create cities where they're
like, okay, we know that we've, we're suffering because our cities are not
really walkable and there's so many reasons why we're suffering.
First, we're not getting the movement.
Second, there's a social aspect of walkable city, where you're able to interact with people, you'll
make friends by just happening to be in the same place as others. Let's actually build walkable
cities. And if we want to transport faster, we'll have these self-driving cars on the perimeter of
the city that are going around.
And I think there are ways in which technology can afford us to live in a way that reverses, I guess, in a more local way.
I like that vision, but I also am aware that there's a different vision, right? You see people in Palo Alto, for example, actually exerting
very strong controls on how much their children are exposed to phones and…
I live in Palo Alto.
Yeah. So you see that. On the other hand, what I am worried about is that the elites
of Palo Alto don't realize that what they're doing is they're figuring out
how to reduce the harm to their own families as they're exporting the harm to the world
of these technologies that for everybody else are unregulated.
And so the question is, can we bring everybody along?
If the AI revolution is going to alter our relationship to work and everything else,
can we bring everybody along so that at the end
of this process, instead of saying, well, it's a shame that 3 billion people were sacrificed to
this transition, but progress is progress, we can really say, well, we figured it out. And everybody
now is living in a style that is closer to their programming and closer to the expectations of
their physical bodies. If that were true, then I and closer to the expectations of their physical bodies.
If that were true, then I would love to be wrong in my fears about what's coming.
But unfortunately, the market is not going to solve this problem without our being deliberate
about forcing it to.
What's your biggest fear?
When you say my fears about what's coming, what's the picture that comes in your mind?
It's a whole different topic actually.
My fear coming, stemming from technology and AI is that this is a runaway process and that
that runaway process is going to interface very badly with some latent human programs. That in effect, the need for workers largely disappears and the people who are
at the head of the processes that result in that elimination for the need for workers
start talking about useless eaters. Maybe they come up with a new term this time.
Thin the herd.
Yep, or they allow it to be thinned or something.
Right.
I've heard you talk about the five key concerns you have or the five key threats you have
before.
Could you name those five?
So the first one is the one I worry least about.
I don't worry zero about it, but I worry least about it, which is the malevolent AI that
the do-mers are so focused on. The second one is the idea that an AI can be misaligned not because it has divergent
interests but because it just misunderstands what you've asked of these autonomous agents.
The famous example is you ask them to produce as many paperclips as possible and they start
liquidating the universe to make paperclips and it's a sorcerer's apprentice kind of
issue. The third one I would say is actually all of the remainder of them I would say are guaranteed.
And the third of them is the derangement of human intellect that we are already living
in a world where it's very difficult to know what
the facts even mean. The facts are so filtered and we are so persuaded by algorithms that
our ability to be confident even in the basic facts, even within our own discipline sometimes
is at an all-time low and it's getting worse. And that problem takes a giant leap forward at the
point that you have the ability to generate undetectable deep fakes, right? That's going to
alter the world very radically when the fact that you're looking at videotape of somebody
robbing a bank doesn't mean that they robbed a bank or that a bank was even robbed. So anyway, I call this.
We deal with this a lot, by the way. I think every single week, every single week I send
to my, I have a chat with people that just are now basically spending, I'd say 30% of
their time dealing with deep fakes of me doing crypto scams, inviting people to telegram
groups and then asking them for credit card details. We had one on X.
I think you probably saw it down, didn't you, of me?
But someone was running ads, deepfake ads on X of me.
And it wasn't just one ad.
It was like, it was like swatting flies.
There was 10 of them and I messaged them to X and there was 10 more.
Then the day after there was 10 more, then the day after there was 10 more.
Then it started happening on Meta.
So it's a video of me basically asking you to come to a Telegram group
where people are being scammed and audience members of mine are being scammed.
And when I send them to Meta, they thankfully remove them,
but then there's five more.
So I don't want a LinkedIn yesterday, and my DMs are,
Steve, by the way, there's this new scam.
And actually at this point, I need someone full time
just sending this over to Meta.
I'm the same, but on a small scale.
Every week, it's, did you really message me on Facebook asking me for my crypto wallet
and blah, blah, blah?
My least favorite one's when the single mother messages me saying that she just paid £500
of her money and how devastated she is.
And I feel this moral obligation to give her her money back because she's fallen for some
kind of scam.
That was me.
It was my voice. It was a video of me telling her something.
Yeah.
And I don't know, I don't know how you deal with that, but sorry, do you continue?
Well, I mean, that's actually on the list here.
The massive disruption to the way things function, both because people are going
to be unemployed in huge numbers and because those who are not abiding by our
social contract are going to find themselves empowered more than the people who do.
So in this case, not only is this poor woman now out 500 bucks for whatever the scam was,
but you've also been robbed, whether or not you pay her back for the thing that she thought
she purchased, your credibility is being stolen by somebody and you have no capacity to prevent
it.
This has happened to me also and it is profoundly disturbing and it is only one of a dozen different
ways that AI enables those who are absolutely willing to shrink the pie from which we all
derive in order to enlarge their slice.
There are innumerable ways that this can happen.
And I think people do not see it coming.
They don't understand how many different ways they are going to be robbed every bit
as surely as if somebody was printing money.
And then the last one is that this just simply accelerates demographic processes that do
potentially result in the unleashing of technologies that pre-existed AI. This can easily
result in an escalation into wars that turn nuclear. So anyway, I think that list could probably be augmented at this point now that we've
spent a little time in the AI era.
We can begin to put a little more flesh on the bones, both of what is possible in this
era and what we should fear.
One of those, you mentioned the truth, the problem of truth.
Would you say, just a thought experiment,
someone today, like an average person,
college-educated person,
are they more propagandized or led astray
than someone in Soviet Russia?
Well, I don't know because I didn't live in Soviet Russia,
but my understanding from
people who did was that there was a wide awareness that the propaganda wasn't true.
It doesn't mean they knew what to believe, but there was a cynicism, which is one of
my fears here is that you're really stuck choosing between two bad options.
In a world where you can't tell what is true.
You can either be overly credulous and be a sucker all the time, or you can become a
cynic and you can be paralyzed by the fact that you just don't believe anything.
But neither of those is a recipe for- Do you think Google search first and maybe now
chat GPT has helped people more or less to find truth. I think that's not chat GPT exactly, but all of the various AI engines have briefly enhanced
our capacity to know what's true because in fact, they allow us to see through the algorithmic
manipulation because the AI is not well policed.
You can get it to recognize patterns that people will swear are not true.
And so anyway, a lot of us have found it useful in just simply unhooking the gas lighting.
So that's been very positive.
But I also remember the early days of search.
And search used to be a matter of there are some pages out there.
I don't know where they
are, here's a mechanized something that's looked through this stuff and just point me
at the direction of things that contain these words, right?
Before the algorithmic manipulation started steering us into believing pure nonsense because
somebody who controlled these things decided it was useful for us to believe those things.
So my guess is at the moment, AI is enhancing our ability
to see more clearly, but that really
depends on some kind of agreement
to protect that capacity that I'm not aware of us having.
Anjali, you're implying there that AI will protect us
from AI, i.e. the woman that got scammed in my audience.
The platforms would have a tool built in,
which would be able to identify shortly that that is not me.
And the ad has been launched by someone in another country potentially.
And then also when she starts being asked for her credit card details in such a way
on Telegram, 10 minutes later, the system will able to understand there that this is
probably a scam at that touch point too.
And it will also be the defense, not just the offense.
First thing, question, is Meta incentivized to solve this problem?
Yes.
Yes. And so Meta is probably actively working on AIs. And again, it's going to be a cat and mouse
game, like every abuse that happens out there. So I think that the market will naturally respond
to things like that.
And the same way that, you know, we installed antiviruses
as annoying as they are,
I think we'll install AIs on our computers
that will allow us to at least help us kind of sort
the fake from the truth.
Well, but let's take the example you say.
Is Metta incentivized to solve this problem?
Superficially, it seems that it should be.
But how many times in recent history have we watched a corporation cannibalize its own
business over what at best is the bizarre desires of its shareholders?
Why was X throwing off people with large accounts or Facebook or Google?
It would seem that you would expect based on the market choosing search engines or social
media sites, you would expect these companies to be absolutely mercenary and say, if Alex
Jones has a big audience, who are we to say?
That's what I would have expected.
Instead, you had these companies policing the morality of thought, even though it reduced
the size of the population using the platforms.
I have a hard time explaining why that happened, but I have every reason to expect the same
thing will happen with AI.
What are you excited about with AI? What's your optimistic take? Because at the start
of this conversation, you said that there's infinite ways that it could improve our lives
and there's 10 times more ways that it could hurt our lives. But let's investigate some
of those ways that it could drastically improve our lives.
There's a couple of different ways. One, we have, as we mentioned before, a dearth of competent teachers and
professors. And that is a problem that will take three generations at least to solve if
what we're going to do is start tomorrow and start educating people in the right way that
would make them competent to stand at the front of a room and educate. But if we can
augment that process, if we can leverage a tool like AI so that a small
number of competent teachers can maybe reach a larger number of pupils, that's plausible,
I think.
Second thing is we have a tremendous number of problems that are obstacles to us living
well on this planet that AI might be able to manage that human intellect alone
cannot, right?
Just in the same way that, you know, compute power can calculate things at a rate that
human beings can't keep up and there are certain things you want calculated very well.
There are also some reasoning problems.
You could imagine that instead of having static laws that govern behavior poorly because they
get gained, that you could have a dynamic interaction.
You could specify an objective of something like a law and then you could monitor whether
or not a particular intervention successfully moved you in the direction that
you were hoping to go or did something paradoxical, which happens all the time.
And you could have, you could basically have governance that is targeted to navigation
and prototyping rather than to specifying a blueprint for how we are to live.
So we wouldn't need politicians.
At the moment, we're stuck with, protections that are as good as has been constructed
and still inadequate to modern realities.
Dan, what are you excited about with AI?
From an individual level, but also from a societal level?
Yeah, well, the big ones are healthcare and education.
It's ridiculous that you are sitting there in pain having had an MRI and there just hasn't
been someone to look at that MRI yet and tell you what to do.
That could easily be solved.
There's all sorts of healthcare issues where, and also not only that, throughout the entire
world there are places that just don't have general practitioners and they don't have
medical advisors and the breakthroughs in global healthcare will be phenomenal and the breakthroughs in
global education could be transformational on the planet.
I'm excited at an individual level that I think the industrial age created a bunch of
jobs that are very dehumanizing and we've just kind of gotten used to them and put up
with them.
The idea that work should be repetitive
and you just repeat the same loop over and over and over
again and over a 10 year period of time,
you might get graduated up one gear
and all that kind of stuff.
I don't think that's very human.
The idea that you could be simultaneously writing a book,
launching a business, running a team, launching a festival,
having an event that you could actually be doing this kind of like mini kingdom work
where you've got this little ecosystem around you of fun things that you're involved in.
That is actually made possible for a vast majority of people if they embrace these kind
of tools. You can live an incredibly fulfilling and amazing and impactful existence.
Oh, I know that I do as a result of having these tools in my life.
Like I'm doing things that I could have only dreamed about as a kid.
And what would you say to entrepreneurs?
I know you work with thousands of entrepreneurs.
What are you telling them in terms of their current businesses or business
opportunities that you're foreseeing?
So I think that small teams have infinite leverage now and that when you have a
team of say, five to 10 people who share an incredible passion for a meaningful
problem in the world and they want to see that meaningful problem solved and they
come together in the spirit of entrepreneurship to solve that problem,
that little five to 10 person team armed with the technology that we now have available,
you can have a big impact.
You can make a lot of money, you can have a lot of fun, you can solve meaningful problems
in the world, you can scale solutions.
You can probably do more in a three year window than most people did in a 30 year career.
And then that little band of five to 10 people could either go together onto a new meaningful
problem, or they could disband and work on other meaningful problems with different teams.
In such a world where you have this sort of infinite leverage, but everyone else has access
to the same infinite leverage, what becomes the USP, going back to the side of the moat,
what is the thing of value when we've all got access to $ same infinite leverage. What becomes the USP, going back to the side of the moat, what is the thing of value when we've all got access
to $20 infinite leverage?
Well, first of all, the first thing you need to do,
you need to understand is that this moment of time
is the least competitive moment.
If you understand how to use these tools,
you can start making money tomorrow.
I see countless examples of people
making thousands of dollars with these hustles
that I talked about, or building businesses
that generates millions of dollars
in the first couple months of existence.
So I would say start moving now, start building things.
So it's an unprecedented time of wealth creation.
Clearly, at some point,
as the market gets more efficient,
as more and more people understand how to use these tools,
there's less potential for creating
these massive businesses quickly.
And we've seen this, like the dawn of the internet
or dawn of the web, you know,
it was a lot easier to create Facebook than it is now than we had mobile.
And for three, four, five years, it was very easy to create massive businesses and then
it became harder being just at the edge of what's possible.
It's going to be very, very important over the next couple of years.
And that gets me really excited because the entrepreneurs who are paying attention are
going to be having the most amount of fun, but they're also going to be able to make a lot of money.
How many applications have been built on Replet to date?
I can talk about the millions of things that have been built since we started the company,
but just since September when we launched Replet Agent, there's been about three million applications built purely in
natural language with no coding at all, purely natural language.
Of those, I think 300, 400,000 of them were deployed in a real, the site was deployed
and it is having, people are using it, some kind of business, some
kind of internal tool.
I built one last night, by the way.
An internal tool or?
I built an application to track how my kids earn pocket money.
So I just told it that I wanted to track the tasks that are happening around the house
and put a sign of value to them.
And I want to be able to, at the end of the week, push a button and get a
summary of how much to pay each child, um, for their pocket money.
We are so screwed.
And within 15 minutes, it had created this application and it was amazing.
Like you could toggle between like, here's the place where you have the kids
and here's the weekly reports and here's the, um, how much per task and you can tick off the tasks or remove tasks or
add tasks.
So then I now have this application for, which took 15 minutes of just talking
about what I wanted and now I have an application to run the pocket money
situation in the, in the house.
And this is by the way, it's something having run an IT agency years ago, that's something that we would have charged five to 10,000 pounds to
create, five to 10,000 US dollars to, to create.
And how much time?
Probably talking something that would have been a three, four week project.
At the start of the S curve now that you're describing, and it's already, if
you, if it's a $20 replicates roughly $20 a month.
25.
For the base case, you did one day of usage.
Let's say it's a dollar.
Yeah.
It costs you and it costs you minutes and a dollar now.
And we're at the start of the S curve.
And you talk to it like you're chatting to a developer.
So one of the things that slows down the development process is you have to send
the information to a developer and they need to understand it and then they need to create something and then come back
to you. This just happens in front of your eyes while you're watching it and it's actually
showing you what's being built. It's really wild.
This one change has transformed how my team and I move, train and think about our bodies.
When Dr. Daniel Lieberman came on the Diary of a CEO, he explained how modern shoes,
with their cushioning and support,
are making our feet weaker and less capable
of doing what nature intended them to do.
We've lost the natural strength and mobility in our feet,
and this is leading to issues like back pain and knee pain.
I'd already purchased a pair of Viva Barefoot shoes,
so I showed them to Daniel Lieberman,
and he told me that they were exactly the type of shoe that would help me restore natural foot movement and rebuild
my strength. But I think it was plantarhystitis that I had where suddenly my feet started
hurting all the time. And after that, I decided to start strengthening my own foot by using
the Vivo Barefoot. And research from Liverpool University has backed this up. They've shown
that wearing Vivo Barefoot shoes for six months can increase foot strength by up to 60%. Visit vivobarefoot.com
slash DOAC and use code DIARY20 from my sponsor for 20% off. A strong body starts with strong
feet. This has never been done before. A newsletter that is ran by 100 of the world's top CEOs.
All the time people say to me they say can you mentor me?
Can you get this person to mentor me? How do I find a mentor?
So here is what we're gonna do you're gonna send me a question and the most popular question you send me
I'm gonna text it to 100 CEOs
some of which are the top CEOs in the world running a hundred billion dollar companies and
Then I'm gonna reply to you via email
with how they answered that question.
You might say, how do you hold onto a relationship
when you're building a startup?
What is the most important thing
if I've got an idea and don't know where to start?
We email it to the CEOs, they email back.
We take the five, six top best answers.
We email it to you.
I was nervous because I thought
the marketing might not match the reality,
but then I saw what the founders were replying with and their willingness to reply. And I thought, actually might not match the reality. But then I saw what the founders were replying with
and their willingness to reply.
And I thought, actually, this is really good.
And all you've got to do is sign up completely free.
I don't think we've spent a lot of time
talking about autonomous weapons.
This is the thing that really worries me.
And the thing that worries people about AI
is this idea is that it is this emergent system and there's
no one thing behind it.
It can act in a way that's unpredictable and not really guided by humans.
Also think it's true of corporations, of governments.
And so I think individual people can often have the best intentions, but the collective can land on doing things
in a way that's harmful or morally repugnant.
And I think we talked about China versus the US,
and that creates a certain race dynamics
where they're both incentivized to cut corners
and potentially do harmful things.
And in the world of geopolitics and wars, They're both incentivized to cut corners and potentially do harmful things.
And in the world of geopolitics and wars, what really scares me is autonomous weapons.
What does it scare you?
Because you can imagine autonomous drones being trained on someone's face, and you can send a swarm
of drones and they can be this sort of autonomous killing assassination machine.
And it can sort of function as a country versus country technology in the world of war, which is still crazy,
but it can also become a tool for governments
to subjugate the citizens.
And people think we're safe in the West,
but I think the experience with COVID
showed that even the systems in the West can very quickly become draconian.
Yeah.
Apparently I've heard in Iran that they have facial recognition cameras that detect
whether women are wearing hijabs in their own cars and it automatically detains the
car if you're driving and you're not wearing a hijab.
And certainly if you're walking down the street, it just picks that up and immediately you're
in trouble.
It acts as a police officer and a judge and a lawmaker.
It's the judge, jury, and executioner essentially, and it just happens instantaneously.
What happened in Canada with the truckers' sort of protests, where they froze their bank
account by virtue of just being there.
Just by being in that location.
Dan, just to confirm that, Iran has implemented a comprehensive surveillance system to enforce
its mandatory hijab laws utilizing various technologies, one of which is cameras and facial recognition. So they've put cameras in public spaces to
identify women who are not adhering to the hijab dress code.
And just on that, London has just put those facial recognition systems into London, and
also all throughout Wales, and they're being rolled out at speed. And like all you would need is a
change of government that wanted to implement something similar and all the base layer
technologies already in there. It gets a little bit worse in Iran because they have this new app
called the Nizar app, where the government has introduced the Nizar mobile application, which
allows you as a citizen to report another citizen who is not wearing their hijab, and it
logs their location, the time when they weren't wearing it, and the vehicle license plate. With
the crowdsourced data, it can then go after that individual. I would also just point out that I
think we're not being imaginative enough. I agree with you, I have the same concern about these
autonomous weapons, but I also think this doesn't have to occur in the context of war or even
governmental oppression, that it is perfectly conceivable that effectively this drops the price
of an undetectable or an unprosecutable crime. And maybe economic moats return in the form of people
taking out their competitors or anybody who attempts to compete with them using an autonomous drone that can't be traced back to them, you know,
that follows facial recognition and, you know, you don't have to kill very many people for
others to get the message that this is a zone that you shouldn't mess around in.
So I could imagine, you know, effectively a new high- tech organized crime that protects rackets and makes tons of money
and subjugates people who haven't done anything wrong.
I had Mustafa Salman on the podcast in 2023 when all of this stuff started kicking off
and he is the CEO of Microsoft AI.
You're familiar with Mustafa?
Of course.
Yeah.
And one of the things he said to me at the time was, one of my fears is a tiny group of people who wish to cause harm
are going to have access to tools that can instantly destabilise our world.
That's the challenge.
How to stop something that can cause harm or potentially kill.
That's where we need containment.
And it sounds a little bit like what you're saying, Amjad,
that we will now have these tools.
You were talking in the context of the military, but
as Brett said, even smaller groups of people, that might have been cartels or gangs, can
do similar harm. At the moment, in terms of autonomous weapons, both the US and China
are investing heavily in AI-powered weapons, autonomous drones, and cyber warfare because
they're scared of the other one getting it first. And we, we talked about how much of our lives run on the internet, but cyber
weapons and cyber AI agents that could be deployed to take down China's X, Y,
or Z or vice versa are of real concern.
Yeah.
I think all of that is, is, is a real concern, you know, unlike Mustafa, I, I
don't think containment is possible.
Part of the reason why this game theoretic system
of competition between the US, China, corporations,
individuals, makes it so that this technology is already,
is already out, really hard to put it back in the bag.
I did ask him this question and I remember the answer because it was such a stark moment
for me.
I said to Mustafa, do you think it's possible to contain it?
And he replied, we must.
So I asked him again, I said, do you think it's possible to contain it?
And he replied, we must.
And I asked him again, I said, do you think it's possible?
He said, we must.
So the problem with that chain of thinking is that it might lead to an oppressive system.
There is one of the, say, do-mers or philosophers of AI, which I respect his work, his name
is Nick Bostrom.
And he was trying to think of ways in which we can contain AI.
And the thing that he came up with
is perhaps more oppressive than something
that the AI would come up with.
It's like total surveillance state.
You need total surveillance on compute,
on people's computers, on people's ideas
to not invent AI or AGI.
It's like taking the guns or something.
Right, exactly.
I mean, there's always this problem of containing any sort of technology is that you do need
oppression and draconian policies to do that.
Are you scared of anything else or concerned about anything else as that relates to AI
outside of autonomous weapons?
You know, we talked about the birthright crisis, and I think a more generalized problem there is creating
virtualized environments via VR where everyone is living in their own created universe and it's so
enticing and even create simulates work and simulate struggle such that you don't really need
to leave this this world. And so every one of us will be solipsistic, you know, similar to the
Matrix. Ready Player One. Ready Player One. We're all kind of plug, even worse than a Ready Player
One, at least that's a massively networked environment. I'm talking about AI simulating everything for us.
And therefore you're literally in the matrix, you know?
Maybe this is...
I had that same thought.
I've enjoyed this great simulation.
Yes.
And so, I mean, are you familiar with the Fermi's paradox?
No, no.
So Fermi's paradox is the question of, you know, professor, his name is Fermi.
He asked the question, if the universe is that vast, then where are the aliens?
The fact that humans exist, you can deduce that other civilizations exist.
And if they do exist, then why don't we see them?
And then that spurred a bunch of Fermi solutions. So there's, I don't know, you can find hundreds
of solutions on the internet. One of them is the sort of house cat thought experiments,
where actually aliens exist, but they kind of put us in an environment like the Amish in a certain time and do not expose us to
what's going on out there. So we're pets, maybe they're watching us and kind of enjoying what we're
doing. Stopping us from hurting ourselves. Stopping us from hurting ourselves. There are so many
things, but one of the things that I think is potentially a solution of the Fermi's paradox. And one of the saddest outcomes is that civilizations
progress until they invent technology that will lock us into infinite pleasure and infinite
simulation such that we don't have the motivation to go into space to seek out the exploration
of potentially other alien civilizations.
And perhaps that is a determined outcome of humanity,
or like a highly likely outcome
of any species like humanity.
We like pleasure, pleasure and pain is the main motivators.
And so if you create an infinite pleasure machine,
does that mean that we're just at home in our VR environment
with everything taking care for us
and literally like the matrix?
And the world, the real world would suck in such a scenario.
Yes, it'd be terrible.
I mean, the other simpler explanation
of the Fermi paradox is that you generate
sufficient technology that you can end your species and it's only
a matter of time from that point, which we can have that discussion about nuclear weapons,
we can have it about AI, but does some technology, if we stay on that escalator, does some technology
that we generate ultimately?
Whatever allows you to get off the planet allows you to blow up the planet.
There you go.
I want to get everyone's closing thoughts and closing remarks.
And hopefully in your closing remarks, you can capture something actionable
for the individual that's listening to this now on their commute to work, or
the single mother, the average person who maybe isn't as technologically
advanced as many of us at this table, but is trying to navigate through this
to figure out how to live a good life over the next 10, 20, 30 years.
Yeah.
I think as long as you need.
I think we live in the most interesting time in human history.
So for the single mother that's listening, for someone who wouldn't be the stereotype of a tech bro, don't assume that you can't do this stuff.
It's never been more accessible today within your work.
You can be an entrepreneur.
You don't have to take massive risk to go create a business
by quit your job and go create a business.
There are countless examples.
We have a user who's a product manager
at a larger real estate business.
And he built something that created 10% lift and
conversion rates, which generated millions and millions of dollars of that business.
And that person became celebrity at that company and became someone who is
lifting everyone else up and teaching them how to use these tools.
And obviously that, that is like a really great for, for anyone's
career and you're going to get a promotion.
Your example of building a piece of software for your family, for your kids
to improve and to learn more, to be better kids, as an example of being
entrepreneur in your family.
So I really want people to break away from this concept of entrepreneurship
being, this is your podcast, a diary of a CEO you started this podcast
By talking to CEOs I assume right and over time
It changed to everyone can be a CEO. Everyone is some kind of CEO in their life
and so I think that we have
On presented access to tools for that vision to actually come
to reality.
Well, it is obviously a moment of a kind of human phase transition, something that I believe
will be the equal of a discovery of farming or writing or electricity.
The darkness that I think is valid in looking at all of the possible outcomes of this scenario
is actually potentially part of a different story as well.
In evolutionary biology, we talk about an adaptive landscape in which a niche is represented as a peak and a higher niche, a better niche,
is represented as a higher peak.
But to get from the lower niche to the higher niche, you have to cross through what we call
an adaptive valley.
And there's no guarantee that you make it through the adaptive valley.
And in fact, the drawing that we put on the board, I think, is overly hopeful because
it makes it in two dimensions.
It looks like you know exactly where to go to climb that next peak.
And in fact, it's more like the peaks are islands in an archipelago that is in fog where
you can't figure out what direction that peak is and you have to reason out it's probably
that way and you hope not to miss it by a few degrees.
But in any case, that darkness is exactly what you would expect if we were about to
discover a better phase for humans.
And I think we should be very deliberate about it this time.
I think we should think carefully about how it is that we do not allow the combination
of this brand new, extremely powerful technology and market forces to turn
this into some new kind of enslavement.
I don't think it has to be.
I think the potential here does allow us to refactor just about everything.
Maybe we have finally arrived at the place where mundane work doesn't need to exist
anymore, and the pursuit of meaning can replace it.
But that's not going to happen automatically if we don't figure out how to make it happen.
And I hope that we can recognize that the peril of this moment is best utilized
if it motivates us to confront that question directly. Each one of us has two parents, four grandparents, eight great grandparents, 16, 32, 64.
You've got this long line of ancestors who all had to meet each other.
They all had to survive wars.
They all had to survive illness and disease.
Everything had to happen for us.
Each individual, all of this stuff had to happen for us to get here.
And if we think about all the people in those thousands and thousands of people, every single
one of them would trade places in a heartbeat if they had the opportunity to be alive at
this particular moment.
They would say that their life was struggle, disease, that their life was a lot of mundane and meaningless
work.
It was dangerous.
You know, every single one of us has probably got ancestors that were enslaved, probably
got ancestors that died too young, probably got ancestors that worked horrific conditions.
We all have that.
And they would all just look at this moment and say, wow,
so are you telling me that you have the ability to solve meaningful problems, to come up with
adventures, to travel the world, to pick the brains of anyone on the planet that you want
to pick the brains of? You can just listen to a podcast. You can just watch a video.
You can talk to an AI. Like, are you telling me that you're alive at this particular moment?
Please make the most of that.
Like, do something with that.
You know, you can sit around pontificating about society
and how society might work,
but ultimately it all boils down to what you do
with this moment and solving meaningful problems,
being brave, having fun,
making your little dent in the universe.
You know, that's what it's all about.
And I feel like there's an obligation to your ancestors
to make the most of the moment.
Thank you so much to everybody for being here.
I've learned a lot and I've developed my thinking,
which is much the reason why I wanted to bring us all together
because I know you all have different experiences,
different backgrounds and education, you're doing different
things, but together it helps me sort of pass through all of these ideas to figure out where
I land. And I ask a lot of questions, but I am actually a believer in humans. I was thinking
about this a second ago, I was thinking, am I optimistic about humans' ability to navigate
this just because I have no other choice? Because as you said, the alternative actually isn't what I'm thinking about.
And so I do have an optimism towards how I think we're going to navigate this,
in part because we're having these kinds of conversations.
And we in history haven't always had them at the birth of a new revolution.
When we think about social media and the implications that had,
we're playing catch up with the downstream consequences.
And I am hopeful.
Maybe that's the entrepreneur in me.
I'm excited.
Maybe that's also the entrepreneur in me.
But at the same time, to many of the points Brett's raised
and Amjad's raised and Dan's raised,
there are serious considerations
as we swim from one island to another.
And because of the speed and scale of this transformation
that Brett highlights, and you look at the stats of the growth of this technology and scale of this transformation that Brett highlights, and
you look at the stats of the growth of this technology and how it's spreading like wildfire
and how once I tried Replit, I walked straight out and I told Cosy immediately, I was like,
Cosy, try this.
And she was on it and she was hooked.
And then I called my girlfriend in Bali, who's the breathwork practitioner, and I was like,
type this into your browser, R-E-P-L-I-T, and then she's making these breathwork schedules with all of her
clients' information ahead of the retreat she's about to do.
It's spreading like wildfire because we're internet native.
We were native to this technology, so it's not a new technology, it's something on top
of something that's intuitive to us.
So that transition, as Brett describes it, from one peak to the other or one island to
another, I think is going to be incredibly destabilizing.
And having interviewed so many leaders in this space from Reid Hoffman,
who's the founder of LinkedIn, to the CEO of Google, to Mustafa, who I mentioned.
They don't agree on much, but the thing that they all agree on and that Sam
Altman agrees on is that the long-term future, the long-term way that our
society functions is radically different.
People squabble over the short term.
They sometimes even squabble over the midterm or the timeline, but they all agree that the
future is going to look completely different. Amjad, thank you for doing
what you're doing. We didn't get to spend a lot of time on it today, and
this is typically what I do here, but your story is incredibly inspiring,
incredibly inspiring from where you came from, what you've done, what you're
building. And you are democratizing and creating a level of playing field for entrepreneurs
in Bangladesh to Cape Town, to San Francisco, to be able to turn their ideas into reality.
And I do think just on the surface that that's such a wonderful thing that, you
know, I was born in Botswana in Africa and that I could have the same access to turn
my imagination into something to change my life because of the work that you're doing at Replet.
And I highly recommend everybody to go check it out.
You probably won't sleep that night because it's so, for someone like me, it was so addictive
to get to be able to do that because it's been the barrier to creation my whole life.
I've always had to call someone to build something.
Dan, thank you again so much because you represent the voice of entrepreneurs and you've really
become a Titan as a thought leader for entrepreneurs in the UK.
And that perspective, that balance,
is incredibly important.
So I really, really appreciate you being here, as always.
And you're a huge fan favorite of our show.
And Brett, thank you a gazillion times
over for being a human lens on complicated challenges.
And you do it with a fearlessness
that I think is imperative for us finding the truth
in these kind of situations, where some of us can run off with with optimism and we can be hurtling towards the mousetrap
because we love cheese. And I think you're an important counterbalance and voice in the
world at this time. So thank all of you for being here. I really, really appreciate it.
And we shall see. These things live forever, so...
I find it incredibly fascinating that when we look at the back end of Spotify and Apple
and our audio channels, the majority of people that watch this podcast haven't yet hit the
follow button or the subscribe button wherever you're listening to this. I would like to
make a deal with you. If you could do me a huge favour and hit that subscribe button I will work tirelessly from now until forever to make
the show better and better and better and better. I can't tell you how much it helps when you hit
that subscribe button. The show gets bigger which means we can expand the production, bring in all
the guests you want to see and continue to do in this thing we love. If you could do me that small
favour and hit the follow button wherever you're listening to this that would mean the world to me. That is the only favour I will ever ask you. Thank you so much for your time. Bye!