The Journal. - Artificial: Episode 3, ChatGPT
Episode Date: January 7, 2024OpenAI launched ChatGPT with low expectations and little fanfare. But the chatbot was an instant hit and went on to become one of the fastest growing consumer apps in tech history. ChatGPT’s surpris...e success gave OpenAI its first shot to make big money, and the company moved quickly to cash in — even as critics called out some very real problems with the company’s hit product. Further Reading: Outcry Against AI Companies Grows Over Who Controls Internet’s Content The Awkward Partnership Leading the AI Boom Further Listening: Artificial: Episode 1, The Dream  Artificial: Episode 2, Selling Out Learn more about your ad choices. Visit megaphone.fm/adchoices
Transcript
Discussion (0)
Hey, it's Kate.
This is the third episode in our series, Artificial,
the OpenAI story.
If you haven't heard episodes one and two,
go back and listen to them.
They're in your feed.
They're also linked in the show notes.
Last June, Eric Davis got a call from his brother.
And right off the bat, he says he had a bad feeling.
My brother and I have never talked on the phone once in my entire life.
We usually text back and forth, but we have never talked on the phone once.
So when I saw my brother's name pop up on the caller ID,
my heart sank.
I knew something was wrong.
And when I picked up the phone,
he just said,
Eric, mom's dead.
Eric's mom, Andrea,
had passed away unexpectedly.
She was my best friend throughout all my life.
She was the type of person who could walk into a party and by the end, everybody would know who she was.
The feelings you get after losing a parent,
I think is indescribable,
especially when you're as close to them as I was with my mom.
Grief does very strange things to people,
and I'm no exception.
And that's kind of where ChatGPT came into play.
How do you mean?
I was sitting at my computer.
I was putting together a slideshow for her funeral.
And I got the idea that,
hey, I never got to say goodbye to her.
What if I use ChatGPT to say goodbye?
Why can't I have ChatGPT
pretend to be my mom?
Eric had been playing with OpenAI's new chatbot for a few months at this point.
He works at a company that sells stuff online, and he'd gotten ChatGPT to write some code for him.
What the script does is I can just click a button.
The script will see which products are out of stock, and it will remove them from the purchase orders.
And it saved me probably four hours every
Monday morning. That's nice. Yeah, it's a really great thing to have. It's like having someone
much smarter than you looking over your shoulder. But this time, Eric would be asking ChatGPT to do
something very different. He'd be asking it to role play. ChatGPT would play the part of his mom, and Eric would get
to have one last conversation with her. For this to work, Eric would need to teach ChatGPT all about
Andrea, and so he started looking back through his phone. Before she died, Eric says he and his mom would text back and forth multiple times a day.
He still had hundreds of her messages.
I selected the ones that I thought showed how she talked, her personality, her sense of humor.
And I plugged them into ChatGPT along with some instructions on,
hey, you're going to act as my mom who just passed away.
And after that, I pasted in the text messages and hit enter.
I wrote, hi, mom.
And when it started writing back, I just burst into tears.
Its first paragraph that it wrote back, it was almost indistinguishable from how my mom
talked. I asked it, you passed away so soon, what do I do? And it wrote back a paragraph as her
basically telling me to be good, don't be boring, and think of the memories that
I had with her.
Did it feel like her?
It did.
It was almost as if she didn't pass away.
Did this conversation with ChapGPT as your mom feel like the closure you were looking for?
It kind of did.
When I was done, I felt a sense of relief that I was able to say what I needed to say.
Immediately after that, however, I felt almost gross, like I had done something wrong.
I remember asking myself after I was done, what the hell did I just do?
And why did that feel gross?
Because I think that once someone has passed, you honor their memory.
You think of the good times. You don't try to turn an algorithm into them and try to talk to them.
So did you have more conversations with it?
No, that was the only one I had.
Have you been tempted?
Oh yeah, every day. Every day I think of it.
Oh, yeah. Every day. Every day I think of it. AI has operated in the background of our lives for years. Algorithms decide what ads we see, what websites were served, what insurance rates were offered.
But ChatGPT was different.
You could have a conversation with it.
It could sound like your mom.
And that was a game changer.
Not just for us, but for OpenAI.
ChatGPT catapulted the company into the spotlight.
But with that attention came tough questions about how OpenAI's systems are built and
who they benefit.
From The Journal, this is Artificial, the OpenAI story.
I'm Kate Leinbach.
Episode 3, ChatGPT. We'll see you in Trollhalla. Oh, very flavorful. Yodeling with them. Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Ew.
Must be mating season.
And hiking with them.
Is that a squirrel?
Bear!
Run!
Collect more moments with more ways to earn.
Air Mile.
Summer's here and you can now get almost anything you need delivered with Uber Eats.
What do we mean by almost? You can't get a well-groomed lawn delivered, but you can get chicken parmesan delivered, OpenAI's president, Greg Brockman, told us he didn't have particularly high expectations for ChatGPT.
Inside the company, the bot wasn't seen as a big leap forward.
ChatGPT was basically a language model OpenAI had already released, GPT 3.5 with a chat interface stuck on the front of it.
In fact, Greg and his colleagues were already hard at work building a bigger and better language model, GPT-4.
GPT-4 was clearly going to be the chat product.
Like, we knew that for sure.
3.5, we just didn't think was going to be that interesting of a product.
And so it was like,
yeah, like, let's release it.
It was better to, like,
get it out and shake the bugs
out ahead of time.
Internally,
they talked about the release
of ChatGPT as a, quote,
low-key research preview.
I remember telling people
the night before that,
because we were all kind of asking,
like, is this going to take away
the thunder from GPT-4 launch?
But I was like,
no one's going to pay attention.
It's all good.
We're not going to have more than like,
you know, like one tweet thread
with more than like five or 10,000 likes,
whatever the number was.
Made a very specific prediction,
which was totally wrong.
Chat GPT was released
on November 30th, 2022.
And it exploded.
A popular new artificial intelligence application
is making content creation a whole lot easier.
It is called ChatGPT.
AI programs surpassed 1 million users in just five days.
If you haven't heard of ChatGPT,
you've been living under a rock.
ChatGPT became one of the fastest-growing consumer apps in history.
Suddenly, OpenAI was sitting on something the whole world wanted.
And that meant the potential to make a lot of money.
This was new territory for the company.
Just a few years earlier, CEO Sam Altman had been quizzed in an interview
about how OpenAI planned to bring in revenue.
His response?
You know, the honest answer is we have no idea.
We have never made any revenue.
We have no current plans to make revenue.
We have no idea how we may one day generate revenue.
We have made a soft promise to investors that once we've built this sort of generally intelligent system,
soft promise to investors that once we've built this sort of generally intelligent system,
basically, we will ask it to figure out a way to generate an investment return for you.
Build AGI, this super intelligent machine, and then ask the AGI how to make money.
That was the business plan.
It sounds like an episode of Silicon Valley.
It really does.
I get it.
But with the release of ChatGPT, OpenAI had a product to sell, and it moved quickly to cash in.
A few months after ChatGPT's launch, OpenAI unveiled a new, souped-up version of its chatbot. This version of ChatGPT was powered
by GPT-4, that new better language model Greg and his colleagues had been working on. Users now had
two options. Use the old 3.5 version of ChatGPT for free, or pay 20 bucks a month to use the turbocharged GPT-4 version.
OpenAI also began licensing its language models to businesses.
Companies in healthcare and hospitality, publishing and accounting
partnered with OpenAI and began rolling out products based on ChatGPT.
OpenAI was riding high, and so was its biggest partner.
What was the response at Microsoft to ChatGPT's success?
Joy and surprise.
And I think they felt like, oh, wow, okay, so this bet really paid off.
That's tech reporter Deepa Sitharaman.
By the time ChatGPT came out, Microsoft had poured about $3 billion into OpenAI.
It had been a bit of a speculative investment.
No one could be sure OpenAI's tech would take off. But after
ChatGPT, you see like the executives just be very proud of their correct bet and then think trying
to look for ways to double down. You know, at this time, Microsoft and OpenAI are talking about the next round of funding for the company. OpenAI
needs more money because they have a lot of compute needs if they really want to build AGI
and keep going in that direction. And Microsoft liked what it saw. And then you see an even
larger check cut from them in the months, in the months after chat GPTs launched.
Microsoft committed to invest about $10 billion in OpenAI, according to Wall Street Journal
reporting. And this partnership propelled Microsoft to the front of the AI race. Rivals
like Google and Meta were scrambling to catch up, and Microsoft was putting OpenAI's
technology into nearly all of its products, from Office to its search engine Bing. Here's CEO Satya
Nadella. It's a new day in search. It's a new paradigm for search. Rapid innovation is going
to come. In fact, a race starts today in terms of what you
can expect. And we're going to move. We're going to move fast. And for us, every day...
ChatGPT had gone from a low-key research preview to a ubiquitous hit product in record time.
And all of this was happening despite some pretty glaring problems with ChatGPT itself.
That's next. Need a great reason to get up in the morning? Well, what about two? Right now, get a small, organic Fairtrade coffee
and a tasty bacon and egger breakfast sandwich
for only $5 at A&W's in Ontario.
DQ presents...
Picture this.
Picture the DQ freezer,
home to all the Blizzard flavours of the past.
Picture it opening... to bring back the Salted Caramel Truffle Blizzard for a limited time.
Picture that Salted Caramel Truffle Blizzard in your hands.
It's all yours. No, really, it's all yours.
This treat is too good to share. Everyone will have to get one for themselves.
Hurry in to get this flavor before the DQ Freezer closes.
DQ.
Happy tastes good.
On August 9th, don't miss the Borderlands movie starring this summer's biggest cast.
Everybody buckle up.
With Cate Blanchett, Kevin Hart, Jack Black, Ariana Greenblatt, and Jamie Lee Curtis.
Wow, you never see that.
Borderlands. Hello.
Hello. How are you?
A few months ago, I phoned up a guy halfway around the world to hear about his experience with ChatGPT.
His name is Brian Hood. I live in a little town called Trentham
that's just outside Melbourne, Victoria, Australia. I'm the mayor of Hepburnshire Council at the
moment. What is Hepburnshire like? Hepburnshire, it's only about a one-hour drive out of Melbourne. It's a farming community.
Picture rolling green hills, very nice countryside.
Brian definitely looks the part of mayor of the Shire.
He sports a flat cap and has a handlebar mustache.
When he's not working, he says he likes to garden,
zipping around his yard on his ride-on mower.
He really doesn't have much use for chat GPT.
But Brian found himself getting pulled into chat GPT mania anyway because he discovered that the chatbot
would sometimes put out false information about him.
It all had to do with a big financial scandal in Australia
known as the Securrency Scandal.
Two currency companies controlled by the Reserve Bank
had caught up in a foreign bribery scandal this lunchtime.
Before becoming mayor, Brian had worked for a company that printed banknotes.
Polymer banknotes.
And to cut a very long story short,
And to cut a very long story short, I uncovered that some of our people were paying bribes in order to win contracts.
Oh.
So it was white collar crime.
Yeah, that's not good.
So what was your role in this?
I was the whistleblower.
Ah.
So my role was to rock the boat, spill the beans.
Brian's bean spilling helped expose the bribery.
But when Brian asked ChatGPT about the history of the scandal,
this was not the story the chatbot came up with.
What did it say? It would say that I had been charged with serious crimes, that I'd been imprisoned for three years, all of which was completely untrue.
I've never been charged with anything, and I certainly have never, ever been to jail.
This made Brian angry.
never, ever been to jail. This made Brian angry. OpenAI did have a disclaimer about the chatbot's accuracy, but Brian didn't think it was enough. He worried the false information could harm him
in his current job. I'm an elected official. I'm representing my community. The town I live in is fairly small. People know me, and I'm the mayor of the whole
shire. So my reputation, my integrity counts for a lot. So I thought, well, I've got to do something
about this. And initially, I didn't know what to do. I thought, well, I can't send a message back
to ChatGPT. So I pretty quickly concluded, well, I just got to go to the lawyers.
Brian's lawyers told him ChatGPT's responses could constitute defamation, and they sent a letter to
OpenAI. Among other things, they demanded that the company stop ChatGPT from generating false
information about him. OpenAI's lawyers rejected Brian's defamation claim,
but the company did make a change.
If you now type my name in, you get back an error message
and it doesn't tell you anything.
Not a very elegant solution in itself, but they've done those other things.
Not very elegant or effective.
Because when I asked ChatGPT about the scandal without naming Brian, it still generated false information.
The bot told us that Brian had been, quote, accused of involvement in corrupt practices.
ChatGPT's tendency to mix fact and fiction?
It's been a problem ever since the chatbot's debut.
fact and fiction, it's been a problem ever since the chatbot's debut. People would head over to ChatGPT looking for a quick answer to a question, some help brainstorming, or homework help. What
they got back could sound convincing, but wasn't always factual. There was the time, for example,
when ChatGPT wrongfully included a famous law professor on a list of people accused of sexual harassment.
There was this allegation of harassment on a trip with students.
That trip never occurred.
I've never gone on any trip with law students of any kind.
It had me teaching at the wrong school, and I've never been accused of sexual harassment.
Then there was the time a lawyer used ChatGPT for work. He asked the bot for some cases
that he could cite in a briefing. However, the report was filled with over half a dozen
bogus court decisions created by ChatGPT. These mistakes weren't that surprising when you consider
how ChatGPT works. When the chatbot generates an answer,
what the underlying language model is doing
is predicting which word is most likely to come next.
But what's likely isn't always true.
What does it mean to have such a stubborn problem like this?
And like, why?
Yeah, I think it's actually a really, really instructive example. When I spoke with Greg Brockman, OpenAI's president back in October,
I told him about Brian Hood's frustrations with ChatGPT, and I asked Greg about the bot's
persistent accuracy problems. In this case, you know, this is clearly something that is not a
desired property of the system. And the question is, well, how do you patch it? And so we have a combination of different
approaches, right? One is that we try to actually fix it at the model level, but that takes time
and we may not be perfect at it. So then we also try to have more classic sort of safety approaches
where you just have some keyword block list or you add some other model on top that's able to
detect different scenarios that we're not good at yet. And I think that, again, the core thing here is this question of, well, how should you approach it?
Should you say, we're not going to deploy until all this stuff is perfect?
Or do you deploy when it's imperfect and you learn from feedback?
But why would you build off of a system that has flaws?
Like, you're saying you're working on it, but you're also pushing forward.
But, well, I guess, do you know of any
systems that don't have any flaws? OpenAI says it takes the accuracy problem seriously and that it
devotes a lot of time before each release to making sure its systems are as factual as possible.
One way the company does that is by enlisting legions of humans to essentially teach its chatbot.
Before releasing ChatGPT, for example, OpenAI had people chat with it and rate its responses.
Researchers then used those ratings to steer the bot toward more helpful and accurate answers.
This approach helped make ChatGPT better, but it hasn't solved the problem.
Critics of the chatbot have also pointed to a different issue, ChatGPT's training data.
When OpenAI's language models predict the next word, they're making those predictions based on
all the words they've seen before, all the words
in their training data. And some people are questioning where all those words came from
and whether OpenAI has the right to use them. Some of the biggest and best-selling authors
in fiction are suing OpenAI. Game of Thrones author George R.R. Martin and 16 other writers are suing the creators of ChatGPT.
The lawsuit filed with the Authors Guild
claims that OpenAI reproduced copyrighted books
without the author's consent to train ChatGPT.
Will you take legal action?
I might.
I might. I might.
Well, I'm a member of the Authors Guild, and I support their lawsuit.
That's Min Jin Lee, author of the best-selling novels Free Food for Millionaires and Pachinko.
I'm writing this other book right now, and it's taking me f***ing forever.
I promise I will not curse on the podcast.
But it's making me really emotional.
And my husband's like, well, you were really upset when you're writing Pachinko, too.
And I was like, oh, that's right.
Yeah.
We called Min at her home in Harlem to talk about the issues authors are having with chatbots.
We asked ChatGPT to write a short story in your style.
Do you want to hear what it came up with?
Wow.
Sure.
I asked, can you write a five-sentence story about a dog in the voice of Min Jin Lee?
Okay.
Here it is.
Okay. Here it is. Okay. In the hushed alleyways of Old Soul, Bori, a wise-eyed Jindu, remembers tales of her grandmother whispered under starlit nights.
The Li family, bound by tradition and memories, saw in her the legacy of a land once whole. Each evening, as lanterns lit the streets, Young Soo, the family's elder,
shared stories of ancestors in honor while Bo-Ri listened intently. Through her, the Lees felt the
pulse of their history, a heartbeat echoing through generations. In the intertwining of
their tales, Dog and family found unity in shared memories.
What do you think?
I think it's really bad fiction writing, and I say that as a teacher, and also I say that as a former judge of the Pulitzer Prize and the National Book Award.
It really makes a mockery of Asians and Asian Americans who write fiction about Asia.
It's almost like when you hear like chimes being used when we think about the audio for
Asian things or thinking about chop suey when we think about Chinese food. It's a bastardization and a caricaturization
of what I care about. It's hard to say how ChatGPT learned what writing in the style of
Min Jin Lee means. One possibility is that OpenAI used Min's work as training data for its language models.
But whether or not that's true, only the company can say.
And the company hasn't said.
That's a shift.
OpenAI used to publish papers about its models, which included their code and information
about their training data sets.
But as OpenAI started to look less like an academic research lab and
more like a company, it's been less open about some of those details. Now, OpenAI will only say
that ChatGPT is trained on licensed content, publicly available content, and content created
by users and human AI trainers. And that lack of specific information
has left authors like Min Jin Lee wondering whether ChatGPT's success was built off the
back of their copyrighted work. I think that authors don't write to get rich. The average author in America lives below the poverty line. So clearly, we are
a bunch of people who just do this because we care so intensely about the proprietary information.
And I think that these companies who care intensely about their proprietary information
are being deeply hypocritical in the way they treat our property.
When Greg and I spoke in October, I asked him about authors' concerns
and about OpenAI's recent secrecy about its training data.
Why not disclose your training data?
Well, I think that, you know, there's some answer for, well,
it's proprietary because it's something that helps us, you know, sort of build the next system. It's something that we spent a lot of time on to get here. The selector is really for any given fact that we reveal, how much does this help people who are going to use the system? How much does this help people choose what to use it for versus how much does it help people build the exact same thing that we're trying to build. And how do you respond to this, like, growing list of authors who are saying,
you're using my copyrighted work as training data and you need to pay us for that?
I think that there's, look, I think there's a really important question for society to answer
here. You know, I think that there's a legal question, which is definitely for the lawyers.
I am not a lawyer, you know, but I have an opinion.
You know, I do think that, first of all, that we as society, like, I think that there should be some way for compensation to flow back to people who are producing, you know, works.
We're definitely very interested in that.
We've been talking about that for some time.
We're definitely very interested in that.
We've been talking about that for some time.
But I also think that there is something really interesting to step back and think about.
Like, these systems are just not what we're used to, right?
That what we're training them to do is to help us solve problems that no one's ever solved before.
Like, that's the actual goal.
I ran Greg's response by Minjin Lee.
The argument is sort of a greater good argument that AI is going to solve problems for the world that no one has solved before.
And that in the process of creating this new thing,
there'll be some troubles along the way
that we'll just have to sort out.
So I can say that my book and my work
is a way of advancing humanity
and it serves a greater good.
I have absolutely no proof that their work will solve the greater good.
And I'm not in any way cowered by their theory of advancing humanity.
Also, what they're doing is not for free.
So why is their idea more valuable than mine?
I don't think that it is.
In the year after ChatGPT's release,
public pressure continued to build on OpenAI.
But it didn't keep the company down.
OpenAI put out a steady stream of new releases.
There was GPT-4
and the latest version of its image generator, DALI.
The company's revenue was growing fast.
By mid-November, the company was closing in on a deal that would value OpenAI at nearly $90 billion,
about triple what it had been worth at the beginning of last year.
ChatGPT had made OpenAI one of the most valuable tech companies in the world.
And then came a shocking announcement
that put the entire future of the company in jeopardy.
Breaking news, Sam Altman is out
as CEO of OpenAI, the company just announced.
Coming up, an exclusive interview with one of the four board members who fired Sam Altman.
Yeah, my name is Helen Toner.
How would you describe that weekend for you?
We knew that it would not be a joyride, and it wasn't a joyride.
That's next time on Artificial, the OpenAI story.
Artificial is part of The Journal, which is a co-production of Spotify and The Wall Street Journal.
I'm your host, Kate Leinbaugh.
This episode was produced by Annie Minoff and Laura Morris, with help from Pierce Singey and Kylan Burtz.
The series is edited by Maria
Byrne. Fact-checking by Amelia Schoenbeck, with consulting from Arvind Narayanan. Series art by
Pete Ryan. Sound design and mixing by Nathan Singapak. Music in this episode by Katherine
Anderson, Peter Leonard, Emma Munger, and Nathan Singapak. Our theme music is by So Wiley and
remixed by Nathan Singapak. Special thanks to Katherine Brewer, Jason Dean, Tom Doughton,
Karen Hao, Berber Jin, Sarah Platt, Sarah Rabel, and Lisa Wang. Thanks for listening.
Our final episode will be released later this month.