Front Burner - Iran and AI on the battlefield

Episode Date: March 6, 2026

For decades we have been hearing about the possibility of AI-driven warfare, and now it’s here.Anthropic's AI platform Claude has been reportedly central to the U.S.-Israeli war on Iran. It was used... during the attack that killed Iranian Supreme Leader Ayatollah Ali Khamenei, which involved strikes on nearly 900 targets dropped within the first 12 hours, including on a girls’ elementary school that killed at least 165 people – mostly students.Today we’re talking about AI military capabilities: how companies like Anthropic and OpenAI are working with the military, and what happens when these companies and governments start building systems that help decide who lives and who dies in a war.Heidy Khlaaf, the Chief AI Scientist at the AI Now Institute and an expert on AI safety within defense and national security, joins the show.

Transcript
Discussion (0)
Starting point is 00:00:00 This ascent isn't for everyone. You need grit to climb this high this often. You've got to be an underdog that always overdelivers. You've got to be 6,500 hospital staff, 1,000 doctors all doing so much with so little. You've got to be Scarborough. Defined by our uphill battle and always striving towards new heights. And you can help us keep climbing. Donate at lovescarbro.cairro.com.
Starting point is 00:00:30 This is a CBC podcast. Hi everyone, I'm Jamie Poisson. For decades, we have been hearing about the possibility of AI-driven warfare. Now it's here. Anthropics' AI platform, Claude, has reportedly been central to America and Israel's war on Iran. Within the first 12 hours, there were more than 900 strikes, including one that killed the Supreme Leader Ayatollah Khomeini. An air strike on a girls' elementary school killed at least 160.
Starting point is 00:01:09 people, mostly students. Today we're talking about AI military capabilities, how companies like Anthropic and Open AI have become or are on their way to becoming deeply enmeshed in the military. And what happens when these companies and governments start building systems that help decide who lives and who dies in a war? I'm joined today by Heidi Klaff. She is the chief AI scientist at the AI Now Institute and an expert on AI safety within defense and national security. including an autonomous weapon systems. And she previously worked at OpenAI. Heidi, hi, thanks so much for making the time.
Starting point is 00:01:56 Thank you for having me. And it's really great to have you. So let's start with what we know about how AI is being used in the war with Iran right now. What rule is it playing? So Cloud is currently being used as what we call a decision support systems, which means that it brings together a lot of different types of data that has been gathered. So you're looking at satellite images, social media. feeds, intercepted phone communications, and then it uses them to make recommendations,
Starting point is 00:02:23 including target recommendations, prioritizing them, and even providing coordinates for those targets. They're also likely used in other phases of the kill chain, so that it includes things like intelligence gathering, surveillance, and even collateral damage assessment, as sort of have been confirmed by research reports. And this is all happening, as I understand it, because of this pairing with the military's Maven smart system, which is built by the company Palantir, which is mining all this data, including classified data, and then it's been mesh with Claude. And can you just like describe to me what's going on there? Yeah. So Palantir is a data analysis company and they work a lot
Starting point is 00:03:05 on intelligence and surveillance. And back in 2024, Anthropic and Palantir had an agreement where essentially Anthropics model that being clod would empower sort of this recommendation engine, which takes all of the enormous amounts of data that Palantir has sort of analyzed, collected, and sort it through, and then feed it to an AI to then make some sort of decision. And so I mentioned that strike on the girls' school in Iran. The Pentagon is denying targeting civilians. Israel says that there was no IDF operations in the area. But is it possible that AI was used in some form here? I think that's actually a very good question because it brings the point forward that AI is actually being used to evade accountability. And that's because when you're using these types of systems, it makes it difficult to distinguish if some attacks were in fact deliberate due to indulgence failures or due to the lack of AI accuracy or the parameters that were even set for these systems in terms of how many casualties they were willing to accept. So it actually muddies the accountability altogether while also obscuring what the LLM is actually doing here. And so, you know, it's very difficult for us to say if this was due to an AI mistake or if this was deliberate or not.
Starting point is 00:04:29 And that's exactly why a lot of militaries are using AI. It obscures that. Right, right. Because I had been reading that like the school was at one time a base for the IRGC, right? And so is it possible that that information was just scraped from some kind of system and then analyze using these tools? It is a possibility, but it's impossible for us to know. And also, you know, when we're talking about these systems as well, a lot of the times they provide these recommendations and it could be on outdated intelligence. So it could have been a combination of both intelligence failures and also the model really not understanding that that data was, you know, no longer relevant.
Starting point is 00:05:08 And that's exactly why these models shouldn't be used for targeting because they have a very difficult time making these types of distinctions where a human might be able to. It also very well could be that this was deliberate or not the fault of the AI. Right. But as I said, it's very hard to know. Claude through this partnership with Palantir was also used by the Pentagon during the U.S. rate to capture Venezuelan President Nicholas Maduro, according to the Wall Street Journal. although there is not a lot of details on the precise role there either. And just like what other examples of military AI use do we know about in recent years in the U.S. or elsewhere, really? So there's a lot of different types of AI.
Starting point is 00:05:59 And I just like to always make it clear that we've had AI be used in the military, including within the U.S., maybe even since the 1960s, as far back as that. That's when a lot of research started on using these types of systems. But those are very different types of AI. than what's being deployed today, and that being large language models, right? Even Maven initially used very different types of AI before large language models came in,
Starting point is 00:06:25 and then they sort of had the partnership with Anthropic. And the thing about these different types of AI is that military purpose-built AI is what we call sort of the former category tends to be very task-specific. It's trained on specific sets of data and very specific sets of tasks. LLMs, on the other hand, or generative AI, are very general purpose, and then they're kind of like fine-tuned to be repurposed for military purposes. But the thing to remember is that military purpose-built AI or task-specific AI tends to be more accurate than general-purpose models like LLMs.
Starting point is 00:07:01 LLMs have an incredibly low accuracy rate. You're looking at a lot of hallucinations. You're looking at something like 25 to 50 percent accuracy, and yet they're still being deployed. And that's not to say that other types of AI weren't that accurate either, but at least instead of us going towards a direction of maybe stepping away or understanding a little bit more how to responsibly deploy AI in the military, instead we went to the complete opposite direction. We're deploying something that's even worse than what we had before. So I would say, you know, if we're talking about different types of conflicts, yes, we have
Starting point is 00:07:38 seen AI. There's been a lot of investigations on drone strikes in Afghanistan. and, you know, Syria and Iraq even. But that's very different from what we're seeing today, where we're essentially having a system really make a final decision on everything. And they're just simply not capable of doing that. Right. And just to be super clear here about what you're saying is it's the same kind of system as these chat bots that we're using. Yes.
Starting point is 00:08:07 Right? Like OpenAI's chat cheap, T that regularly hallucinates and get stuff wrong. Exactly. What they actually do is they take these everyday models that you and I have been using and they input new data on top of it. And essentially, you basically have the same baseline model with a bit more context on military targeting, for example. But it's still the same type of hallucinations. It's still the same lack of accuracy. So these are an improved model as they would like us to believe. We've talked on the show before about this really explosive reporting from The Guardian in 972 magazine that revealed through intelligence sources how Israel used AI at one point to identify like 37,000 potential human targets based on their apparent links to Hamas during the war in Gaza. This AI system was called Lavender. The reporters interviewed people who said they would take the information and drop a dumb bomb on a target off. and killing an entire household of people. And is this what we're talking about here? Was this generative AI?
Starting point is 00:09:23 So actually, Gospel and Lavender did not rely on large language models, but they used different other types of AI to essentially have the same outcome. But after that, we did receive confirmation that both GPT4 and Google Gemini, which are language models, they were then eventually used to generate and validate targets, And, you know, we started seeing hints of that when we saw the cloud contracts with Google and Microsoft that were ramped up after October 7th. So even though they had Lavender and Gospel prior to that, as soon as they saw the opportunity to use large language models, they went ahead, you know, sought out these tools to deploy them very quickly. And regardless of the type, the AI algorithms are all looking to do the same thing, right? They're looking to generate as many targets as possible.
Starting point is 00:10:11 And they all ingest like similar types of information. And then they give recommendations. So they're used towards the same purposes. And as I mentioned, both types of AI really have accuracy problems here. And just big picture, how would you say the use of these LLMs is changing modern warfare, what we're seeing today? I'd actually argue that it isn't changing it. I think when you keep in mind the abysmal accuracy rates and the speed of LLMs, it's almost just a high-tech version of carpet bombing. And what we're actually seeing is that AI is being used to evade accountability. The very use of these systems, like with the school, will make it very difficult to say, okay, whose fault was this?
Starting point is 00:11:03 And so, you know, when you're looking at these types of issues together, the evasion of accountability and the fact that these systems aren't accurate, we're just going towards, you know, a world where we're no longer following the traditional legal rules of war. And I think that's very, very concerning. So to me, we know, we're just seeing us move away from that accountability from understanding the type of targets that we have, from us understanding the decisions that are even being made. because with AI, we don't know why some of the decisions are being made. So it's not, to me, it's not an advancement, but rather a regression. Right, right. Like, I've just seen legal experts talk about how they worry that what this will ultimately result in is rubber stamping automatic strikes with no meaningful human oversight.
Starting point is 00:11:52 Would you say that that's a pretty good summary? Yes. And I think it's really important to remember here, especially, you know, if you're looking at, and specifically Anthropic because they had this fallout with the Department of War. A lot of people are talking about, well, Anthropic has a strong red line because they're talking about they don't want their models to be deployed for autonomous weapons systems and they don't want less human oversight. I don't consider that a moral high ground because actually in practice, the difference between decision support systems, which is how they're currently being used.
Starting point is 00:12:23 And AI-driven autonomous weapon system is this human in the loop that you're talking about. And that's impacted by what in our field we call automation bias. And automation bias is this idea based on decades of research showing that humans often trust the recommendations of algorithms without corroborating with other sources or checking if that recommendation is correct or not. So it does end up end up being rubber stamping. And then again, when you combine that with the knowledge that these things are very accurate, you're looking at the normalization of this technology that shouldn't be used for targeting. And I think that has huge implications of the worst to come. And we already have a very good example of that already, which is what happened in Gaza, where these types of AI systems were also used by the IDF.
Starting point is 00:13:10 Right, right. In Gaza, I know a lot of people called Ressa as well, just for people listening. Yeah. What's your worst fear here? Like, do you have a nightmare scenario that keeps you up at night? Yeah, I think we are in the nightmare scenario when it comes to military use. I think something I'm really worried about is also giving these models access to nuclear weapons. And when we're talking about things like autonomous weapon systems and then we're talking about,
Starting point is 00:13:37 you know, we saw scale AI, which is a different AI company, recently land a contract on nuclear command and control. It really makes you question what's the next step here. We are already in such like a terrible, you know, worst case scenario in terms of giving these models, ability to target, right, with very little human oversight. And now we're seeing just the expansion of that. The fact that we're even having the conversation about autonomous weapon systems is what scary for me, that no one took a step back and say, wait, hold on, they shouldn't even be used for decision support systems. And here we are talking about autonomous weapons systems. And now there's nuclear command and control. And I think, you know, we're actually deploying these
Starting point is 00:14:20 systems in a very, very, very reckless way. And it seems like there's no stop. think. And, you know, a lot of people are hoping that Anthropic is a hero here and they'll draw the red line. But at the end of the day, they're back on the negotiating table with the Pentagon as of today. And at the end of the day, Dario Amadeh, who's the CEO of Anthropic, came out and said, oh, no, I'm not opposed to autonomous weapons systems. We just don't think they're reliable enough and we wanted to work on developing them, you know? So they don't have those red lines. So if the governments aren't doing and the companies aren't doing it, it's supposed to send a very, very bad position of what's to come next. This ascent isn't for everyone.
Starting point is 00:15:11 You need grit to climb this high this often. You've got to be an underdog that always over delivers. You've got to be 6,500 hospital staff, 1,000 doctors all doing so much with so little. You've got to be Scarborough. defined by our uphill battle and always striving towards new heights. And you can help us keep climbing. Donate at lovescarborough.ca.
Starting point is 00:15:38 At Desjardin, we speak business. We speak equipment modernization. We're fluent in data digitization and expansion into foreign markets. And we can talk all day about streamlining manufacturing processes. Because at Desjardin business, we speak the same language you do.
Starting point is 00:15:55 Business. So join the more than 400,000 Canadian entrepreneurs who already count on us and contact Desjardin today. We'd love to talk, business. I want to dig into this whole anthropic, open-a-I kind of fight, I guess, with you just in one minute. But first, I wonder if I could ask you, like, what I think might be an incredibly obvious question, but also I think might be helpful. Like, what would an autonomous weapon system do? This is actually a really good question because I think a lot of people think of killer drones. That's the first thing that they think of.
Starting point is 00:16:30 I mean, I am thinking of killer drones. Exactly. There's many different types of autonomous weapons systems. In fact, you have organizations, international organizations, that consider minds autonomous weapons systems, right? They could look like many things. When you're talking about drones, that's a very different type of AI. You need an AI with specific skills like object recognition. These are not general purpose models.
Starting point is 00:16:52 and then based on that and some data that's fed into them, they then sort of lock in and target something. That's very different from how we're using frontier AI models, which run on these enormous cloud systems in just a huge amount of information. And based on that, give recommendations. And if we're thinking about a targeting recommendation, you can then use any type of weapon to strike that target. So it could be a drone.
Starting point is 00:17:18 You could send a drone off after that target has been identified. It could be a missile if it's a missile. It's an infrastructure target. So there's many different levels of autonomous weapon systems. And I think most people really have focused on this, like, you know, killer drone, drone swarms. But that's actually not where Frontier AI or Cloud is being used. They're using on a much bigger scale where they're given many different types of targets.
Starting point is 00:17:42 But typically a human then takes that recommendation and decide, right? Like, I'm going to send a missile after that, right? Or I'm going to send a drone there. Yeah. But the idea isn't to eliminate that human that when we're talking about Frontier AI. To eliminate all that together, so whatever recommendation is given, the model itself can even decide that type of weapon that's being used. We saw Scale AI sort of give demos on that using large language models. Like, it's a different company that also had a lot of contracts with the Department of War.
Starting point is 00:18:08 But, you know, it's like what would be the best type of weapon to use for this recommendation? And so the idea is to just eliminate the human from that loop and just to automate that process. Got it. And just make it so that maybe a human just, as we were talking before, kind of rubber stands. Yeah. Okay, so given that, I mean, given this request that the Pentagon had of Anthropic, that they wanted the company to remove restrictions on things like autonomous weapons and also large-scale surveillance. And so we have said to the Department of War that we are okay with all use cases, basically 98 or 99% of the use cases they want to do, except for two that we're concerned about. One is domestic mass surveillance. They're worried that things may become possible with AI that weren't possible before.
Starting point is 00:19:00 Case number two is fully autonomous weapons. This is the idea of making weapons that fire without any human involvement. I just wonder if you could tell me a little bit more about what that told you about what the Pentagon ultimately wants to do here, what direction they want to go in with AI. I think they very clearly want to use these technologies as an alibi for whatever actions they want to carry out. Like I said, when we go back to the point that AI really helps you evade accountability, it makes people question, was this deliberate, that they do this on purpose? And it also makes it very easy to say things like, well, the AI determined it so it must be true, right? There's, I think, a group of people who view AI decisions as being more objective than humans, and that's not the case in any way.
Starting point is 00:19:47 And so it's much easier to hide behind AI decision and not have to justify them. So I think when they want no guardrails whatsoever, it generally shows that they want to be able to use these systems in whichever capacity they feel like. On the surveillance stuff, the Atlantic reported that Anthropic was told the Department of Defense, I guess the Department of War now, wanted to use Claw to analyze bulk data collected from Americans. So everything, I think you type into a chatbot, right? All your search, your credit card history. And just what could be done by the military with that level of information? So I think it's really important to note that these models are what we call dual use. So they can be used for civilian purposes and they could be used for military purposes.
Starting point is 00:20:36 And when you essentially start thinking about how they already started using these model for intelligence gathering, right? In intelligence sort of analysis, you start to think, right, this can also. be used on, you know, the citizens of that country themselves because it's the same feature. So if you collect data from different sources, so that things like data brokers, location data, Internet habits, and that's not just the things you're using the model for. That's really commercial data that you can buy from anyone, from data brokers, as we call them. And then you train or input these specific data points into the LLMs. You can then draw inferences on whom these individuals are and track them according.
Starting point is 00:21:17 And again, as I mentioned, it might not be accurate. But you might be making decisions off of that, which, again, is really, really problematic. The U.S. Undersecretary for Defense, who is negotiating the deal with Anthropic, was defending the position. And he said, well, at one point he said, quote, at some level, you have to trust your military to do the right thing. And if anything, we're the biggest, you know, organization in the world with the most rules of any organization of the world, but we do have to be prepared for the future. He also made the argument that what Anthropic was afraid of or what people were afraid of was already barred by the law.
Starting point is 00:21:59 And just, does he have a point there, I guess? I would say that the Department of War conducts its own legal reviews and compliance with international law, and they likely see that Anthropic is overstepping its role in these determinations, including their legal judgment or whether or not to deploy these systems. However, from an international law perspective, there's going to be different interpretations of what the lawfulness and predictability of these AI-based autonomous weapons systems, the difference from how the current Department of War currently chooses to interpret them. So the international community output is necessary here, and we haven't seen any consultations either from orgs like the UN or the ICRC on these types of decision, especially when you have Dario raising concerns that AIA, too unreliable, and that would support an interpretation, actually, that LLM-based autonomous weapon systems are not in line with international law. And either way, we shouldn't be at the
Starting point is 00:23:01 winds of a private corporation redlines on whether or not this dangerous technology should or should it be deployed. But I think it's at the end of day also important to remember that there is international law, right? And it's not just, is a private corporation versus a Department of War, right? There's already, you know, restrictions on what types of autonomous weapon systems can be used. And currently, when we're looking, when Daru admits himself that these models are reliable, it says a lot about sort of their legal status, if that makes sense. Yeah, though I guess some people might be listening to this right now and feeling like maybe they don't have a lot of confidence in international law at the moment. Just state that obvious fact.
Starting point is 00:23:43 Look, like I take your point that, you know, you don't want to make Anthropic out here to be the hero and that CEO to be the hero because they are back at the table now negotiating with the Pentagon. But just I wanted to ask you about the threat to declare them a supply chain risk. Yes. And then also obviously that coming to fruition, right? This label has never been used against the U.S. company, as I understand it, and would bar the company from any work with the government. And just like how big of a deal is that for a company like Anthropic? I mean, it's quite a lever for the government to pull. Definitely, definitely, because ultimately this is the first frontier AI company that's been working with the military and they're embedded with them.
Starting point is 00:24:30 So to suddenly use that label clearly doesn't align with what's already been happening. But on the point of supply chain, I actually do believe that all LLMs are a supply chain threat to, national security and defense. And a lot of people internally in the military also believe the same things. And, you know, this is about the companies themselves being foreign adversary in any way, which they're American companies, are clearly not. But this is about the nature of LLMs themselves that I think is really important for people to understand, is that they're trained on the open internet and publicly available information. Again, very different from purpose military built models where every data point and software decision is traceable.
Starting point is 00:25:12 So as a result of these LMs being built on an unprotected supply chain, there's a lot of new and undetectable attack factors which include things like poisoning web training datasets or building back doors into these models, which actually may intentionally or inadvertently lead to the subversion of their behavior, which includes in military applications. And this isn't something that can be patched. And in fact, a study by Anthropic themselves showed that you only need about 250,000. military documents to produce a backdoor into one of these models. And so when you consider that, with the news that China and Russia have been compromising online information to influence the answers of large language models, it's likely that these backdoors already exist. So I think it's very difficult for me to speak on the Department's Award decision of why they would choose to say this about Anthropic. But I actually do think large language models as a whole do have supply
Starting point is 00:26:10 I'm reticent to like repeat what people have just heard themselves, but I just, I do think it's really important to just put a line underneath this, that essentially what you're saying here is that it's non-anthropic, that is the threat to national security. It's the data that all this stuff is trained on and that an enemy of the United States, you mentioned China or Russia, can mess around with the data and it can affect the outcome. Exactly. What did you make of all the United States? What did you make of Open AI's decision to kind of swoop in and make a deal with the Department of War after this falling out with Anthropic. Is that deal materially different from what Anthropic had wanted to do? You worked at Open AI. I would be very curious to hear your reaction to this.
Starting point is 00:27:13 Yeah, I think it was very clear to me that it was not the same deal, although they claim that it was. It really wasn't, especially when you're familiar with military contract and these open-ended words that are being used, right? And so it's very different language from what Anthropic was fighting for. And initially they came out as saying, oh, we just managed to get the same deal. And then it ultimately came out that that wasn't the case, right? And so it seems that they threw Anthropic under the bus while also trying to sort of benefit from the same PR.
Starting point is 00:27:42 And also in their announcement, because they talk about a safety stack, and this is what I work and I work on safety. I would say the safety guardrails that open the eye is referring to are not operationally feasible because when you have a model and it gives you an output, you can't monitor after that what that individual does with the output. They can't double check that that's being overseen by human. It could just very well be taken to then select and engage a target without any further oversight, making it an autonomous weapon, even if this is deployed on the cloud, because this is an operational matter, not a technical one, whether or not you put a human in the loop.
Starting point is 00:28:17 Right. I think Sam Altman is like basically admitted that, right? Like that they don't, they won't be able to actually know what the Department of War was doing with their technology. Exactly. So when they say we have the safety sack is going to be on the cloud and that's going to somehow limit it, I think that's very, very much misleading. How would you describe how a company like Open AI approaches safety? I would call it safety theater, safety co-option, because ultimately a lot of these AI companies like to use the same terms that safety engineers use and reference. my background is safety engineering, which is we work on safety critical systems like nuclear plants, aviation, you know, autonomous vehicles, where if they fail, human lives are at risk. They take a lot of that same terminology and essentially kind of safety wash everything with it.
Starting point is 00:29:09 Every sort of mission that they have, they eventually like rolled back on like military use, right? That was eliminated in 2024 when it was initially part of their charter or their terms of service. you know, we have a lot of safety incidents that occur all the time with people using their models. There's no guard whales, really. And so I would say for a lot of people, you know, it's, they believe, they often believe the messages like, oh, we care and so on, so forth. But they can write as much as they want. If you just look at the actions of these companies, these AI models are causing a huge amount of harm. Contracts with militaries are very enticing for AI companies, right, for a couple of different reasons. And could you just lay those out? for me. Well, these AI models are incredibly expensive to train. And it's actually likely the case that they're not even breaking even on being able to generate the revenue needed to keep producing these big models, training them and deploying them. So when you have the military, that's a really big money
Starting point is 00:30:09 pot for you to be able to make that money back. And it also embeds you within military and safety critical infrastructure, which means that you're then too big to fail. So for them, you know, when there's a huge amount of backlash right now against the uses of AI. This is a very secure position for them to have monetarily and also in practice because, you know, as we've seen with the reporting of Claude, the entire military infrastructure now relies on Claude for this type of analysis. And it's very hard for them to disconnect from it. So although they wanted a ban, they were like, oh, we're going to try to phase it out and they still use it in Iran, right? It just shows the type of dependence and this dependence is very, very strategic for them. I want to kind of go back to
Starting point is 00:30:48 state power. You know, we've been talking about how much power these private companies have in influencing how wars are fought. But when you look at the pace of AI development, when you see the military interest in this, the money being spent on it, is one of the real main worries here that what we are witnessing right now is AI becoming a central tool of state power? Absolutely. It is the perfect tool to concentrate power. because you're collecting essentially all data possible on humans, on our behavior, on anything, really. It even goes beyond humans, right? You're also collecting data on states, on everything. And then allows them to use those same various systems that are able to collect that.
Starting point is 00:31:34 Huge amounts of information be trained on it. So then make any sort of decision. It becomes very centralized to them. And there's no accountability because these models are black boxes. No one now can question, why did you make that decision? Like we're seeing these models be used for immigration rates by ICE. We're seeing use in judicial systems and military. And yet they're black boxes. And now it feels like no one can question that power.
Starting point is 00:31:57 This is an unfortunate side effect of something like these very large general purpose models. Okay. Seems like a good place for us to end. Heidi, thank you. Thank you so much. All right. That is all for today. Front burner was produced this week by Joytha Shen Gupta, Shannon Higgins, Matthew Amha,
Starting point is 00:32:22 Lauren Donnelly and McKenzie Cameron. Our intern is Riley Cunningham. Our YouTube producer is John Lee. Our music is by Joseph Shabbison. Our senior producers are Imogen Burchard and Elaine Chow. Our executive producer is Nick McKay Blokos. And I'm Jamie Poisson. Thanks so much for listening.
Starting point is 00:32:48 For more CBC podcasts, go to cbc.ca.ca slash podcasts.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.