Podcast Page Sponsor Ad
Display ad placement on specific high-traffic podcast pages and episode pages
Monthly Rate: $50 - $5000
Exist Ad Preview
TED Talks Daily - The AI revolution is underhyped | Eric Schmidt
Episode Date: May 15, 2025The arrival of non-human intelligence is a very big deal, says former Google CEO and chairman Eric Schmidt. In a wide-ranging interview with technologist Bilawal Sidhu, Schmidt makes the case that AI ...is wildly underhyped, as near-constant breakthroughs give rise to systems capable of doing even the most complex tasks on their own. He explores the staggering opportunities, sobering challenges and urgent risks of AI, showing why everyone will need to engage with this technology in order to remain relevant.Want to help shape TED’s shows going forward? Fill out our survey! Hosted on Acast. See acast.com/privacy for more information.
Transcript
Discussion (0)
An Apple Watch for your kids lets you stay connected with them wherever they go.
They can call you to pick them up at grandma's or text you because they forgot their lunch again.
Their watch will even send an alert to let you know they finally got to school.
Great for kids who don't have a phone because their Apple watch is managed by you on your iPhone.
iPhone XS are later required with additional wireless service plan.
This episode is sponsored by Google Pixel.
I am always looking for tools that help me stay curious and efficient.
And lately I've been exploring the Google Pixel 9, which was gifted to me by Google.
What's impressed me most is how it's powered by Gemini.
That's Google's personal AI assistant built right into my phone.
Gemini helps me brainstorm ideas, summarize emails, even plan out my day, all just by
holding the power button.
For example, let me show you how easy it is.
Gemini, summarize my unread emails. Re, away next week. Jonathan confirmed with Elise Hue about rescheduling a meeting. Reminder,
development committee meeting tomorrow at 12 p.m. Central time.
It's super helpful for staying on top of things without feeling overwhelmed. Or when
I needed a quick dinner plan, I snapped a photo of what I had in my fridge and Gemini
gave me recipe ideas. It's like having a research assistant right in my pocket. If you can think
it, Gemini can help create it. Learn more about Google Pixel 9 at store.google.com.
Support for this episode comes from Airbnb. Winter always makes me dream of a warm getaway.
Imagine this, toes in the sand,
the sound of the waves, and nothing on the agenda except soaking up the sun.
I think of myself in the Caribbean, sipping on a frozen drink and letting my troubles
melt into the sea. Maybe Jamaica, Turks and Caicos, St. Lucia, lots of possibilities for me and my
family to explore. But vacations always fly by too quickly. I
was planning my next getaway when I realized my home will be sitting empty
while I'm away. That's why I've been thinking about hosting on Airbnb. It'll
allow me to earn extra income and could help me extend that trip just a little
longer. One more sunset, one more amazing meal, one more day to unwind. It sounds like the smart thing to
do and I've heard it's easy to get started. Your home might be worth more than you think.
Find out how much at airbnb.ca.
You're listening to TED Talks Daily where we bring you new ideas and conversations to
spark your curiosity every day.
I'm your host, Elise Hwu.
AI and the future of humanity were huge topics at this year's TED conference.
Central to all of this was a pretty existential question.
In today's world, what is a human actually for?
To add to this discussion, Google's former CEO and Chairman Eric Schmidt joined creative
technologist Bilal Al-Siddu for a conversation about AI and our collective future.
They discuss what, if any, are the limits of AI, ethical questions about its rising
use across various sectors, and why the AI revolution, as Eric puts it, is underhyped.
Eric Schmidt, thank you for joining us. Thank you. You said the arrival of non-human intelligence
is a very big deal. What did you see that the rest of us might have missed? In 2016, we didn't
understand what was now going to happen, but we understood that these algorithms were new and powerful.
There was a new move invented by AI
in a game that had been around for 2,500 years
that no one had ever seen.
Technically, the way this occurred was that the system of AlphaGo
was essentially organized to always maintain
a greater than 50% chance of winning.
And so it calculated correctly this move, which was this great mystery among all of
the Go players who are obviously insanely brilliant mathematical and intuitive players.
The question that Henry, Craig Mundy and I started to discuss, right, is what does this mean?
How is it that our computers could come up with something that humans had never thought
about?
I mean, this is a game played by billions of people.
And that began the process that led to two books, and I think frankly is the point at
which the revolution really started. If you fast forward to today,
it seems that all anyone can talk about is AI,
especially here at TED,
but you've taken a contrarian stance.
You actually think AI is underhyped.
Why is that?
And I'll tell you why.
Most of you think of AI as,
I'll just use the general term as chat GPT.
For most of you, chat GP was the moment where you said, oh my God, this thing writes and it makes mistakes, but
it's so brilliantly verbal. Right? That was certainly my reaction. Most people that I
knew did that.
It was visceral. Yeah.
This was two years ago. Since then, the gains in what is called reinforcement learning,
which is what AlphaGo helped invent and so forth,
allow us to do planning.
And a good example is look at OpenAI 03 or DeepSeq R1,
and you can see how it goes forward and back,
forward and back, forward and back.
It's extraordinary.
In my case, I bought a rocket company
because it was like interesting. And know as one does as one does and
It's it's an area that I'm not an expert in and I want to be an expert
So I'm using deep research and these systems are spending 15 minutes writing these deep papers
It's true for most of them. Do you have any idea how much computation?
15 minutes of these supercomputers is?
It's extraordinary.
So you're seeing the arrival, the shift from language to language,
then you have language to sequence, which is how biology is done.
Now you're doing essentially planning and strategy.
The eventual state of this is the computers running all business processes.
Right? So you have an agent to do this, an agent to do this, an agent to do this, an agent to this,
and you concatenate them together
and they speak language among each other.
They typically speak English language.
I mean, speaking of just the sheer compute requirements
of these systems, let's talk about scale briefly.
You know, I kind of think of these AI systems
as hungry, hungry hippos.
They seemingly soak up all the data
and compute that we throw at them.
They've already digested all the tokens
on the public internet,
and it seems we can't build data centers fast enough.
What do you think the real limits are,
and how do we get ahead of them
before they start throttling AI progress?
So there's a real limit in energy.
Give an example, there's one calculation, in energy. I'll give you an example.
There's one calculation,
and I testified on this this week in Congress,
that we need another 90 gigawatts of power in America.
My answer, by the way, is think Canada, right?
Nice people, full of hydroelectric power.
But that's apparently not the political mood right now.
Sorry.
So 90 gigawatts is 90 nuclear power plants in America.
Not happening, we're building zero.
Right? How are we going to get all that power?
This is a major, major national issue.
You can use the Arab world,
which is busy building five to 10 gigawatts of data centers.
India is considering a 10-gigawatt data center. To five to 10 gigawatts of data centers, India is considering
a 10 gigawatt data center.
To understand how big gigawatts are
is think cities per data center.
That's how much power these things need.
And the people look at it and they say,
well, there's lots of algorithmic improvements
and you will need less power.
There's an old rule, I'm old enough to remember, right?
Grove giveth, gates taketh away.
OK?
The hardware just gets faster and faster.
The physicists are amazing.
Just incredible what they've been able to do.
And us software people, we just use it and use it and use it.
And when you look at planning, at least in today's algorithms,
it's back and
forth and try this and that, and just watch it yourself. There are estimates, and you
know this from Andreessen Horowitz reports, it's been well studied, that there's an increase
in at least a factor of 100, maybe a factor of 1,000 in computation required just to
do the kind of planning. The technology goes from essentially deep learning
to reinforcement learning to something called test-time compute,
where not only are you doing planning,
but you're also learning while you're doing planning.
That is the, if you will, the zenith or what have you,
of computation needs.
That's problem number one, electricity and hardware.
Problem number two is we ran out of data,
so we have to start generating it,
but we can easily do that because that's one of the functions.
And then the third question that I don't understand
is what's the limit of knowledge?
I'll give you an example.
Let's imagine we are collectively all of the computers in the world
and we're all thinking,
and we're all thinking based on knowledge that exists,
that was previously invented. in the world, and we're all thinking, and we're all thinking based on knowledge that exists,
that was previously invented,
how do we invent something completely new?
So Einstein.
So when you study the way scientific discovery works,
biology, math, so forth and so on,
what typically happens is a truly brilliant human being
looks at one area and says,
I see a pattern that's in a completely different area,
has nothing to do with the first one, it's the same pattern.
And they take the tools from one and they apply it to another.
Today our systems cannot do that.
If we can get through that, I'm working on this,
it's a general technical term for this,
is non-stationarity of objectives.
The rules are keep changing.
We will see if we can solve that problem.
If we can solve that, we're going to need even more data centers,
and we'll also be able to invent completely new schools of scientific and intellectual thought,
which will be incredible.
So as we push towards a Zenith, autonomy has been a big topic of discussion. Yoshua Bengio gave a compelling talk earlier this week,
advocating that AI labs should halt the development
of agentic AI systems that are capable
of taking autonomous action.
Yet that is precisely what the next frontier
is for all these AI labs and seemingly for yourself too.
What is the right decision here?
So Yoshua is a brilliant inventor of much of what we're talking about
and a good personal friend, and we've talked about this.
And his concerns are very legitimate.
The question is not are his concerns right, but what are the solutions?
So, let's think about agents.
So, for purposes of argument, everyone in the audience is an agent.
You have an input that's English or whatever language,
and you have an output that's English or whatever language,
and you have an output that's English, and you have memory,
which is true of all humans.
Now, we're all busy working,
and all of a sudden,
one of you decides it's much more efficient not to use human language,
but we'll invent our own computer language.
Now, you and I are sitting here watching all of this
and we're saying, like, what do we do now?
The correct answer is unplug you, right?
Because we're not going to know.
We're just not going to know what you're up to.
And you might actually be doing something really bad or really amazing.
We want to be able to watch.
So we need provenance, something you and I have talked about,
but we also need to be able to observe it.
To me, that's a core requirement.
There's a set of criteria that the industry believes
are points where you want to metaphorically unplug it.
One is where you get recursive self-improvement,
which you can't control.
Recursive self-improvement is where the computer
is off learning and you don't know what it's learning.
That can obviously lead to bad outcomes.
Another one would be direct access to weapons.
Another one would be that the computer systems
decide to exfiltrate themselves,
to reproduce themselves without our permission.
So there's a set of such things.
The problem with Yashua's speech with respect to such a brilliant person
is stopping things in a globally competitive market doesn't really work.
Instead of stopping agentic work,
we need to find a way to establish the guardrails,
which I know you agree with because we talked about it.
Yeah.
Yeah.
Yeah.
Yeah.
Yeah.
I think that brings us nicely to the dilemmas
and let's just say there are a lot of them
when it comes to this technology.
The first one I'd love to start with Eric is the exceedingly dual use nature of this
tech, right?
It's applicable to both civilian and military applications.
So how do you broadly think about the dilemmas and ethical quandaries that come with this
tech and how humans deploy them?
In many cases, we already have doctrines about personal responsibility.
A simple example, I did a lot of military work and continue to do so.
The US military has a rule called 3,000.09,
generally known as human in the loop,
or meaningful human control.
You don't want systems that are not under our control.
It's like it's a line we can't cross.
I think that's correct.
I think that the competition between the West,
and particularly the United States and China,
is going to be defining in this area.
And I'll give you some examples.
First, the current government has now put in,
essentially reciprocating 145 percent tariffs.
That has huge implications for the supply chain.
We in our industry depend on packaging and components from China
that are boring, if you will, but incredibly important.
The little packaging and the little glue things and so forth
that are part of the computers.
If China were to deny access to them, that would be a big deal.
We are trying to deny them access to the most advanced chips,
which they're super annoyed about.
Dr. Kissinger asked Craig and I to do track two dialogues
with the Chinese and we're in conversations with them.
What's the number one issue there is this issue.
Indeed, if you look at Deepsea, which is really impressive,
they managed to find algorithms that got around the problems
by making them more efficient.
Because China is doing everything open source, open weights,
we immediately got the benefit of their invention
and have adopted into US things.
So we're in a situation now, which I think is quite tenuous,
where the US is largely driving, for many, many good reasons,
largely closed models, largely under very good control.
China is likely to be the leader in open source
unless something changes,
and open source leads to very rapid proliferation around the world.
This proliferation is dangerous at the cyber level and the bio level,
but let me give you why it's also dangerous in a more significant way,
in a nuclear threat way.
Dr. Kissinger, who we all worked with very closely, why it's also dangerous in a more significant way, in a nuclear threat way.
Dr. Kissinger, who we all worked with very closely,
was one of the architects of mutual assured destruction,
deterrence and so forth.
And what's happening now is you've got a situation where,
I'll use an example, it's easier if I explain.
You're the good guy and I'm the bad guy, OK?
You're six months ahead of me,
and we're both on the same path for superintell intelligence, and you're going to get there, right?
And I'm sure you're gonna get there.
You know, you're that close, and I'm six months behind.
Pretty good, right?
Sounds pretty good.
No, these are network effect businesses,
and in network effect businesses,
it is the slope of your improvement that determines everything.
So I use OpenAI or Gemini, they have a thousand programmers.
They're in the process of creating a million AI software programmers.
What does that do?
First, you don't have to feed them except electricity, so that's good.
Second, and they don't quit and things like that.
If you get there first, you dastardly person...
You're never going to be able to catch me. I will not be able to catch you, and they don't quit and things like that. If you get there first, you dastardly person.
You're never gonna be able to catch me.
I will not be able to catch you
and I've given you the tools to reinvent the world
and in particular, destroy me.
That's how my brain, Mr. Evil, is gonna think.
So what am I gonna do?
The first thing I'm gonna do is try to steal all your code
and you've prevented that because you're good
and you were good, because you're good.
And you were good.
So you're still good at Google.
Second, that I'm going to infiltrate you with humans.
Well, you've got good protections against that.
We don't have spies.
So what do I do?
I'm going to go in and I'm going to change your model.
I'm going to modify it.
I'm going to actually screw you up to get me
so I'm one day ahead of you.
You're so good, I can't do that.
What's my next choice?
Bomb your data center.
Wow.
Now, do you think I'm insane?
These conversations are occurring around nuclear opponents today in our world.
There are legitimate people saying the only solution to this problem is preemption.
Now, I just told you that you, Mr. Good,
are about to have the keys to control the entire world,
both in terms of economic dominance, innovation, surveillance,
whatever it is that you care about.
I have to prevent that.
We don't have any language in our society,
the foreign policy people have not thought about this,
and this is coming.
When is it coming?
Probably five years.
We have time.
We have time for this conversation,
and this is really important.
Let me push on this a little bit.
So if this is true and we can end up in this sort of standoff scenario
and sort of the equivalent of mutually assured destruction.
You've also said that the US should embrace open source AI, even after China's deep-seek
showed what's possible with a fraction of the compute.
But doesn't open sourcing these models just like hand capabilities to adversaries that'll
accelerate their own timelines?
This is one of the wickedest, or we call them wicked hard problems.
Our industry, our science, everything about the world that we have built is based on academic
research, open source, so forth.
Much of Google's technology was based on open source.
Some of Google's technology is open source, some of it is proprietary, perfectly legitimate.
What happens when there's an open source model that is really dangerous and it gets into the hands of
the Osama bin Laden's of the world and we know there are more than one,
unfortunately? We don't know. The consensus in the industry right now
is the open source models are not quite
at the point of national or global danger.
But you can see a pattern where they might get there.
So a lot will now depend upon the key decisions made in the US and China
and in the companies in both places.
The reason I focus on the US and China
is they're the only two countries where people are crazy enough
to spend the billions and billions of dollars
that are required to build this new vision.
Europe, which would love to do it,
doesn't have the capital structure to do it.
Most of the other countries, not even India,
have the capital structure to do it, although they wish to.
Arabs don't have the capital structure to do it,
although they're working on it.
So this fight, this battle, will be the defining battle.
I'm worried about this fight.
Dr. Kissinger talked about the likely path to war with China
was by accident.
And he was a student of World War I,
and of course, World War I started with a small event,
and it escalated over that summer in, I think, 1914,
and then there was this horrific conflagration.
You can imagine a series of steps along the lines of what I'm talking about
that could lead us to a horrific global outcome.
That's why we have to be paying attention.
I want to talk about one of the recurring tensions here before we move on to the dreams is
to sort of moderate these AI systems at scale, right?
There's this weird tension in AI safety that the solution to preventing 1984
often sounds a lot like 1984. there's this weird tension in AI safety that the solution to preventing 1984
often sounds a lot like 1984.
So proof of personhood is a hot topic.
Moderating these systems at scale is a hot topic.
How do you view that trade-off, right?
In trying to prevent dystopia,
let's say preventing non-state actors
from using these models in undesirable ways,
we might accidentally end up building
the ultimate surveillance state.
It's really important that we stick to the values that we have in our society.
I am very, very committed to individual freedom.
It's very easy for a well-intentioned engineer
to build a system which is optimized and restricts your freedom.
So it's very important that human freedom be preserved in this. A lot of these are not technical issues,
they're really business decisions.
It's certainly possible to build a surveillance state,
but it's also possible to build one that's freeing.
The conundrum that you're describing is because
it's now so easy to operate based on this information,
everyone knows what I'm talking about,
that you really do need proof of identity.
But proof of identity does not have to include details.
So for example, you could have a cryptographic proof
that you are a human being,
and it could actually be true without anything else,
and also not be able to link it to others
using various cryptographic techniques.
So zero knowledge proofs and other techniques like that.
Zero knowledge proofs are the most obvious.
All right, let's change gears, shall we, to dreams.
In your book Genesis,
you strike a cautiously optimistic tone,
which you obviously co-authored with Henry Kissinger.
When you look ahead to the future,
what should we all be excited about?
Well, I'm of the age where some of my friends
are getting really dread diseases.
Can we fix that now?
Can we just eliminate all of those?
Why can't we just uptake these
and right now eradicate all of these diseases?
That's a pretty good goal.
I'm aware of one nonprofit that's trying to identify in the next two years
all human drugable targets and release it to the scientists.
If you know the drugable targets,
then the drug industry can begin to work on things.
I have another company I'm associated with
which has figured out a way, allegedly, it was a startup,
to reduce the cost of stage three trials by an order of magnitude.
As you know, those are the things
that ultimately drive the cost structure of drugs.
That's an example.
I'd like to know where dark energy is,
and I'd like to find it.
I'm sure that there is an enormous amount of physics
in dark energy, dark matter.
And think about the revolution in material science.
Infinitely more powerful transportation,
infinitely more powerful science and so forth.
I'll give you another example.
Why do we not have every human being on the planet
have their own tutor in their own language
to help them learn something new,
starting with kindergarten?
It's obvious.
Why have we not built it?
The only possible answer is there must not be a good economic argument.
The technology works.
Teach them in their language,
gamify the learning,
bring people to their best natural means.
Another example, the vast majority of health care in the world
is either absent or delivered by the equivalent of nurse practitioners
and very, very sort of stressed local village doctors.
Why do they not have the doctor assistant
that helps them in their language treat whatever with, again, perfect health care?
I can just go on.
There are lots and lots of issues
with this world of the digital world.
It feels like we're all in our own ships in the ocean,
and we're not talking to each other.
In our hunger for connectivity and connection,
we're making these tools make us lonelier.
We've got to fix that, right?
But these are fixable problems.
They don't require new physics.
They don't require new discoveries.
We just have to decide.
So when I look at this future,
I want to be clear that the arrival of this intelligence,
both at the AI level, the AGI, which is general intelligence,
and then superintelligence,
is the most important thing that's going to happen in about 500 years,
maybe 1,000 years in human society,
and it's happening in our lifetime.
So don't screw it up.
Let's say we don't.
Right.
Yeah.
Let's say we don't. Let's say we don't screw it up.
Let's say we get into this world of radical abundance.
Let's say we end up in this place and we hit that point of recursive self-improvement.
AI systems take on a vast majority of economically productive tasks.
In your mind, what are humans going to do in this future?
Like, are we all sipping pina coladas on the beach, like engaging in hotties?
You tech liberal you.
I'm just, I have to ask.
You must be in favor of UBI.
No, no, no.
Look, humans are unchanged
in the midst of this incredible discovery.
Do you really think that we're going to get rid of lawyers?
No, they're just going to have more sophisticated lawsuits.
Right?
Do you really think we're going to get rid of politicians?
No, they'll just have more platforms to mislead you.
Sorry.
I mean, I can just go on and on and on.
The key thing to understand about this new economics
is that we collectively, as a society,
are not having enough humans.
Look at the reproduction rate in Asia.
It's essentially 1.0 for two parents.
This is not good.
So for the rest of our lives,
the key problem is going to get the people who are productive,
that is, in their productive period of lives,
more productive to support old people like me,
who will be bitching that we want more stuff from the younger people.
That's how it's going to work.
These tools will radically increase that productivity.
There's a study that says that we will,
under the set of assumptions around agentic AI and discovery
and the scale that I'm describing,
there's a lot of assumptions,
that you'll end up with something like 30 percent increase in productivity per year.
Having now talked to a bunch of economists,
they have no models for what that kind of increase in productivity per year. Having now talked to a bunch of economists, they have no models
for what that kind of increase in productivity looks like.
We just have never seen it.
It didn't occur in any rise of a democracy or a kingdom in our history.
It's unbelievable what's going to happen.
Hopefully, we will get it in the right direction.
It is truly unbelievable.
Let's bring this home, Eric.
You've navigated decades of technological change.
For everyone that's navigating this AI transition,
technologists, leaders, citizens that are feeling
a mix of excitement and anxiety,
what is that single piece of wisdom or advice
you'd like to offer for navigating this insane moment
that we're living through today?
So one thing to remember is that this is a marathon, not a sprint.
One year I decided to do a 100-mile bike race,
which was a mistake.
And the idea was, I learned about spin rate.
You just, every day you get up and you just keep going.
You know from our work together at Google
that when you're growing at the rate that we're growing,
you get so much done in a year.
You forget how far you went.
Humans can't understand that.
As this stuff happens quicker,
you will forget what was true two years ago or three years ago.
That's the key thing.
So my advice to you all is ride the wave, but ride it every day. two years ago or three years ago. That's the key thing.
So my advice to you all is ride the wave, but ride it every day.
Don't view it as episodic and something you can't,
but understand it and build on it.
Each and every one of you has a reason to use this technology.
If you're an artist, a teacher, a physician,
a business person, a technical person, If you're not using this technology,
you're not going to be relevant compared to your peer groups
and your competitors and the people who want to be successful.
Adopt it and adopt it fast.
I have been shocked at how fast these systems are.
As an aside, my background is enterprise software.
And nowadays, there's a model protocol from Anthropic.
You can actually connect the model directly into the databases
without any of the connectors.
I know this sounds nerdy.
There's a whole industry there that goes away
because you have all this flexibility now.
You can just say what you want, and it just produces it.
That's an example of a real change in business.
There are so many of these things coming every day. Ladies and gentlemen, Eric Schmidt.
Thank you very much.
Thank you very much.
Thank you.
Thank you, guys.
Thank you.
That was Eric Schmidt in conversation with theal Vosudu at TED 2025.
If you're curious about TED's curation, find out more at TED.com slash curation guidelines.
And that's it for today's show. TED Talks Daily is part of the TED Audio Collective.
This episode was produced and edited by our team, Martha Estefanos, Oliver Friedman,
Brian Green, Lucy Little, Alejandra Salazar, and Tansika Sangarnivon.
It was mixed by Christopher Faisy-Bogan, additional support from Emma Tobner and Daniela Ballarezo.
I'm Elise Hu.
I'll be back tomorrow with a fresh idea for your feed.
Thanks for listening. An Apple Watch for your kids lets you stay connected with them.
Wherever they go, they can call you to pick them up at grandma's, or text you because
they forgot their lunch.
Again.
Their watch will even send an alert to let me know they finally got to school.
Great for kids who don't have a phone because their Apple Watch is managed by you on your
iPhone.
iPhone XS are later required with additional wireless service plan.
Hi, it's Morgan from Off the Shelf and I'm here to tell you how my Google Pixel 9 helps
me read more.
Google actually gifted me this phone and now I use it nonstop.
The other day I was trying to remember the name of this book someone
recommended and instead of spiraling into a 40-minute social media scroll I
just asked Gemini on my Pixel. What's that romantic book with a competition and a
ghost helping her through the trials? The book you're likely thinking of is
Phantasma by Kaylee Smith. Here's a breakdown of why it fits your
description. It's like having that one friend who always knows
what you're talking about.
Learn more about the Google Pixel 9 at store.google.com.
Let me tell you about a mattress you need.
Logan & Cove.
It was named Canada's best luxury hybrid mattress.
A hybrid blends traditional steel coils
with modern mattress technology, like cooling gel foam.
And Logan & Cove is made right here in Canada.
But here's the thing, it's luxury sleep without the luxury price tag.
Logan & Cove starts at just $7.99.
It's the perfect balance of comfort and support,
and it's backed by a 365-night trial.
Canadian listeners will even get a free bedding bundle
included with their mattress at loganandcove.ca slash podcast.