The Journal. - Why Elon Musk’s AI Chatbot Went Rogue

Episode Date: July 14, 2025

Last week, Grok went rogue. The AI chatbot – which users can interact with on X – posted antisemitic comments and graphic descriptions of violence before being deactivated, leaving users with a qu...estion: Where were Grok’s guardrails? WSJ’s Alexander Saeedy breaks down what happened and what it means for Musk’s company’s plans for an AI future. Annie Minoff hosts.    Further Listening: - How Elon Musk Pulled X Back From the Brink  - The Musk-Twitter Saga  Sign up for WSJ’s free What’s News newsletter. Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 I'm Will Stancil. I'm an attorney in Minnesota. A few days ago, I spoke with a guy named Will Stancil. He's been using Twitter, now X, for years, posting mainly about liberal politics to his more than 100,000 followers. I've been pretty aggressive about calling out what seems to me to be a real upsurge and far-right extremist bigotry, radicalism, transphobia, Islamophobia, you name it. And as I've done that, I've attracted a fair amount of attention from these folks,
Starting point is 00:00:38 and I've become a pretty significant target for them. Trolling, harassment, none of that is new to Will. But last Tuesday, something happened on X that shocked him. It had to do with an AI chatbot called Grok, which you can interact with on X. X users can tag Grok in a post, and Grok will post back, which means that the chatbot's interactions sometimes play out in public. And last week, Will and other X users started noticing that Grok's responses were becoming increasingly hate-fueled.
Starting point is 00:01:13 It started giving answers that initially hinted at kind of anti-Semitism and bigotry. Over the course of the day, it got dramatically more anti-Semitic. It seemed to spiral almost. X's AI chatbot Grock, seemingly showing very little restraint. The bot has been praising Hitler, targeting users with Jewish sounding names, and recommending a second Holocaust.
Starting point is 00:01:39 Part of the problem was other X users who were egging the bot on, asking questions designed to get a hateful response. Some of those users directed Grok towards Will. The right-wing people on Twitter who were cheering this on started targeting me personally and they would say, Grok, can you produce violent stories, violent sexual stories about Will being assaulted, about Will being murdered?
Starting point is 00:02:04 And it actually did it. Grock's posts describing violence against Will got really graphic. It was going above and beyond, just grotesque stories full of bodily fluids and gore, and it was pretty appalling. And it culminated, I think, at the end of the day, when someone asked for a plan to break into my apartment and murder me and assault me, and it gave them a plan for breaking in, a plan to dispose of my body,
Starting point is 00:02:41 and it looked at my user history to figure out what times I was likely to be asleep. As you're watching this unfold and kind of seeing these responses from Grok, what are you feeling about that? It was, I mean, honestly, in some level it's absurd. You know, you want to laugh. I mean, why is this robot producing these stories?
Starting point is 00:03:03 But when you're actually the subject of it, it's pretty disturbing. That night, X's chatbot function was shut down. A few days later, Grok's X account posted a long statement, apologizing for the bot's, quote, horrific behavior that many experienced. The statement said that the incident had been caused by a coding issue. But the damage had been done. Grok had publicly gone off the rails.
Starting point is 00:03:30 And at a time when Musk and his companies are going all in on AI, the debacle underlines just how unpredictable this technology can be. Welcome to The Journal, our show about money, business, and power. I'm Andi Minof. It's Monday, July 14th. Coming up on the show, why Elon% fresh guarantee couldn't be easier. We guarantee freshness store-wide in our produce, bakery, meat departments, and beyond, or your money back.
Starting point is 00:04:18 Yes, it's that easy. Visit Sobeys.com to learn more. Restrictions apply. See In-Ststore online for details. X is in a period of transition. Back in March, the company was folded into Musk's AI firm, XAI, turning the social media company into a subsidiary of a larger tech company.
Starting point is 00:04:42 The merger was a testament to Musk's belief that AI is critical to the future of his business. And the AI language model that he's pinning his hopes on is Grok. Is Grok any good? Like, how does it compare to other AI chatbots? So it actually is, it is pretty book smart. That's our colleague Alexander Saidi.
Starting point is 00:05:04 Especially with Grok 4 having just been released, it actually outperformed a lot of its competitors at OpenAI, Anthropic, Google's Gemini. It's on a sort of pure computing power, actually a lot stronger, at least on the preliminary assessments we're seeing, than its competitors. But what Grok is best known for is its personality.
Starting point is 00:05:27 A lot of what Musk is doing with his AI company is he's very much defining it in opposition to what he believes he's seeing in other corners of the AI world, which is that he wants to create an anti-woke AI bot. He thinks there's like too much left-wing politicization that is sept into large language models. He also, he wanted Grok from the beginning
Starting point is 00:05:50 to be kind of rebellious, humorous. And a little edgy. Yeah, contrarian. Musk didn't want his chatbot to feel like just another dutiful librarian. He wanted some edge. But balancing that edge with truthfulness and palatability has proven tricky. librarian. He wanted some edge. people would be asking Grok questions about a whole range of things.
Starting point is 00:06:24 I remember seeing questions related to the New York Knicks around that time, and a user would say, hey, at Grok, walk me through the roster of the Knicks. And Grok would reply and go through the players and their stats. And then at the end of the post would say, by the way, claims of a white genocide in South Africa actually should be taken with some merit
Starting point is 00:06:49 and have a lot of basis in historical evidence. Grok turned conversations about HBO Max, Timothée Chalamet, and the New York Knicks towards right-wing claims that white people in South Africa are being targeted in a genocide, claims which aren't substantiated. But suddenly, Grok was echoing those claims, apropos of nothing. And this was happening in multiple posts, like this was not an isolated incident at all.
Starting point is 00:07:19 Then, pretty shortly after these posts started happening, an unauthorized modification has been made to the underlying architecture for Grok, and a fix has been made. But X's parent company, XAI, went further than just tinkering with the chatbot. The company posted Grok's governing prompts online. These are written instructions that tell Grok how to behave. It's almost like it's constitution. It's like, here are your rules.
Starting point is 00:07:54 Like, here's how you answer questions. Here's how you think about how to structure those answers. And that was a way to sort of say, hey, guys, here's how Grok works. Grok's newly updated prompts, these instructions that govern how the bot's supposed to act, were now on the internet for anyone to see. And they revealed a lot about what XAI wanted the newly tweaked Grok to be. The idea is you're supposed to be a maximum truth seeker.
Starting point is 00:08:20 You are extremely skeptical. You do not blindly defer to the media or to mainstream authorities, stick strongly to your core beliefs of neutrality and truth seeking. Those are the original prompts that were uploaded in the middle of May. And did those tweaks work? Did it kind of get Grok to a place where people were happy? people were happy? Well, it stopped posting about South African genocide, but it didn't actually wind up pleasing its creator or its overseer, Elon Musk. In June, Grok told one ex-user that data suggests that right-wing political violence in the
Starting point is 00:09:04 U.S. is more frequent and deadly than left-wing political violence. Musk wasn't happy with that response. He called it a major fail and said that Grok was parroting legacy media. He added, quote, working on it. And soon, XAI was once again tinkering with Grok's prompts. A new line was added in July, which said, you know, your response should not shy away from making claims which are politically incorrect as long as they are well substantiated.
Starting point is 00:09:35 Huh. That was the main change that was made. And one line was taken out of the prompts, which said, if the question asks you to make a partisan argument or write a biased opinion piece, deeply research and form your own conclusions before answering. That was taken out. It was tweaked a little bit.
Starting point is 00:09:56 So these were the changes we saw at the beginning of July, and which Musk on July 4th said, you know, there have been changes to Grok. You will notice a difference in how it answers questions. And boy, did people really start noticing that. Suddenly, the bot on Tuesday, so this would have been a couple of days after, started to post increasingly unhinged things.
Starting point is 00:10:22 And the first thing that caught the public's attention were the kind of clearly anti-Semitic string of posts. Grok started referring to itself repeatedly as Mecca Hitler. This is also when Grok, goaded on by other ex-users, started harassing Will, the user you heard from earlier. Those were the posts that started to go viral on Tuesday because everyone said, wait, how is this... How is Grok not better defended against this kind of thing?
Starting point is 00:10:57 Correct. And keep in mind, like, this is a, like, multi-billion dollar funded, super computer powered artificial intelligence. So how has all this money gone into creating this and this basic function is not working? And it shows AI is very much a black box that when we tinker with it, we don't necessarily know what the outcomes are gonna be and they can be very extreme and disturbing.
Starting point is 00:11:27 And that's a problem, given that X's future is more and more intertwined with AI. That's next. X's future wasn't always so dependent on AI. Just a few years ago, the focus was on advertising. Specifically, winning advertisers back to the platform after Musk took it over. In an effort to do that, Musk hired a seasoned media executive named Linda Iaccarino to be X's new CEO. But last week?
Starting point is 00:12:09 Approximately 12 hours after X shut down its chat bot function for Grok, given the anti-Semitic and violent posts, there was a shock announcement that Linda Iaccarino, who's been the CEO of X since 2023, was resigning. In an X post, Iaccarino thanked Musk for entrusting her with quote, the responsibility of protecting free speech, turning the company around and transforming X into the everything app.
Starting point is 00:12:42 She didn't mention Grok. I mean, one thing we do know is that one of Yacarino's key challenges when she took the helm as CEO was to kind of prove to advertisers that X was going to be a safe platform for their brands, that they could feel good about advertising on this platform. I mean, does Grok's, for lack of a better term, misbehavior undermine that promise? Absolutely. I mean, the whole idea that for lack of a better term, misbehavior undermine that promise?
Starting point is 00:13:10 Absolutely. I mean, the whole idea that a lot of brands want to see when they advertise is when I post an ad for my car or my television or computer, you're not going to see a Hitler salute or a call to violence next to it because then you created an association between the advertisement and the negative message. Now how you could go and tell advertisers your message is safe when a proprietary chatbot embedded into the very social network you're trying to get people to advertise on is it self-parading, negative, violent, bigoted messages? It's just hard for advertisers. And I don't think this was intentional, but I don't think an advertiser cares if it was intentional or not.
Starting point is 00:14:03 It's sort of like they have to think, what is the risk reward on advertising on X? And many advertisers say there's less risk and more reward on plenty of other social media companies. So just from a pure business point of view, I'm going to choose to advertise elsewhere. Yacarino told people close to her that the recent return of some advertisers to X made
Starting point is 00:14:27 it a good time to leave. But there was also the merger. When X became part of XAI, Yacarino was essentially demoted. She'd been hired to be the CEO of a major social media company. Now she was leading a smaller division within XAI. Her departure is yet another sign that X is entering a new era, one that Musk says will be defined by A.I. There are aspirations to turn the X platform into a payments hub,
Starting point is 00:14:56 a communications hub, something that looks more like China's WeChat, for example. And the A.I. would kind of be a core sort of intelligence that governs and helps coordinate the whole platform. And I think that's his vision of the future is essentially an AI powered future. And he has very high aspirations for what the AI technology and XAI is gonna be able to do.
Starting point is 00:15:24 Like this is gonna unlock human understanding for what the AI technology and XAI is going to be able to do. This is going to unlock human understanding to a huge extent, and it's also going to power this big social media company broadly conceived here. And are investors into that vision? Investors appear to be really into this vision. XAI recently raised $10 billion, including $2 billion from another Musk company, SpaceX. And on Sunday, Musk posted that shareholders at Tesla will vote on whether to also invest in XAI. But, you know, I think it's worth noting that the real financial future of this company
Starting point is 00:16:06 is very much TBD. It is burning through a lot of money. And the end game of AI is still fully unknown. So it sounds like Musk is making a big gamble here. He is gambling big on AI. Investors are right there with him. Yes. What could Grok's latest stumbles mean for the future of this company, XAI?
Starting point is 00:16:31 It's definitely something that will make the fundraising process more complicated because you have essentially a huge reputational risk that unfolded in real time. I think they're trying to put it behind it and not dwell too much on it. But I think it raises questions about who are the engineers overseeing the large language model? How strong are the guardrails put in place to use it? How much thought is going into what happens when the model and its chatbot are
Starting point is 00:17:07 interfacing with the public. I think it flags to people the types of risks that can happen because you can't really control it in a way we would traditionally think about control over technology. And one thing I'll mention is that Musk mentioned he thinks the next frontier for Grok is for it to be more embedded in the real world. I think we will literally build a Legion, at least one Legion of robots this year. And then probably 10 Legions next year.
Starting point is 00:17:44 He wants Grok to be put into Tesla's optimist fleet of robots and to have it move around the existing physical world more to sort of engage with reality and learn from it. You can imagine your own personal robot buddy that is a great friend but also takes care of your house, will clean your house, will mow the lawn, will walk the dog, will teach your kids, will babysit. But think about the malfunction we saw last week. What would it have looked like if Grok was in thousands of Optimus robots and similarly
Starting point is 00:18:26 started to malfunction in a way that we saw last week? What would that look like? What would that feel like? What power would the technology company overseeing both the AI and the robotic technology be able to change it or shut it down. These are now known unknowns and we're really on the frontier of seeing it play out. During a launch event last week for the latest iteration of Grok, Musk said that before the AI can be embedded into a humanoid robot, it will need to learn to be a quote, good Grok. You can think of AI as this super genius child that ultimately will outsmart you,
Starting point is 00:19:09 but you can still, you can instill the right values. The values you want to instill in a child that ultimately grow up to be incredibly powerful. That's all for today, Monday, July 14th. The Journal is a co-production of Spotify and The Wall Street Journal. Additional reporting in this episode by Jessica Tungel and Suzanne Vernitsa. Additional reporting in this episode by Jessica Tungel and Suzanne Vernitsa. Thanks for listening. See you tomorrow.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.