Risky Business - Risky Business #837 -- GitHub Actions footgun claims TanStack

Episode Date: May 13, 2026

On this week’s show Patrick Gray, Adam Boileau and James Wilson discuss the week’s cybersecurity news. They cover: Mini Shai-Hulud and the TanStack compromise u...sing Github Actions Instructure pays Canvas elearning platform data extortionists More Linux privilege escalation 0days! CISA helping critical infrastructure operators rearchitect their networks so they work offline This week’s episode is sponsored by email security platform Sublime Security. Bobby Filar chats with Patrick about how agentic AI is being evaluated by buyers in a marketplace that’s experiencing “AI fatigue”. This episode is also available on Youtube. Show notes ‘Mini Shai-Hulud’ malware compromises hundreds of open-source packages in sprawling supply-chain attack | CyberScoop Hardening TanStack After the npm Compromise | TanStack Blog Canvas Breach Disrupts Schools & Colleges Nationwide – Krebs on Security Instructure pays ransom after Canvas incident as Congress announces investigation | The Record from Recorded Future News When DNSSEC goes wrong: how we responded to the .de TLD outage Adversaries Leverage AI for Vulnerability Exploitation, Augmented Operations, and Initial Access | Google Cloud Blog Mythos smythos! How to find 0day with lesser models - Risky Business Media GitHub - V4bel/dirtyfrag · GitHub retr0.zip NVD - CVE-2026-42511 Flaw in Claude’s Chrome extension allowed ‘any’ other plugin to hijack victims’ AI | CyberScoop Ivanti customers confront yet another actively exploited zero-day | CyberScoop Palo Alto warns of critical software bug used in firewall attacks | The Record from Recorded Future News Where Have All the Complex Windows Malware and Their Analyses Gone? Meet Rassvet, Russia’s Answer to Starlink | WIRED DOJ says ransomware gang tapped into Russian government databases | TechCrunch Iranian government hackers using Chaos ransomware as cover, researchers say | The Record from Recorded Future News Foxconn confirms cyberattack impacting North American factories | The Record from Recorded Future News New CISA initiative aims for critical infrastructure to operate offline during cyberattacks | The Record from Recorded Future News ‘HELLO BOSS’: Inside the Chinese Realtime Deepfake Software Powering Scams Around the World How to Disable Google's Gemini in Chrome | WIRED FCC pushes ban on security updates for foreign-made routers, drones to 2029 | The Record from Recorded Future News

Transcript
Discussion (0)
Starting point is 00:00:00 Hey everyone and welcome to risky business. My name's Patrick Gray. Adam Barlow is back on deck and he'll be joining James Wilson and I in just a moment to talk through the week's security news. And there's lots of like awful and funny things happening. So that's going to be fun. And then after that in this week's sponsor interview, we're going to be hearing from Bobby Filler, who heads up artificial intelligence over at Sublime Security. Sublime Security makes the most modern sort of contemporary iterative.
Starting point is 00:00:33 of an email security platform. So if you are, you know, looking to get the best in-class email security platform, you want to hit up Sublime. And we're talking to Bobby about, I guess, how customers these days are evaluating AI features in products. It's an interesting conversation. They are very AI-heavy, Sublime. And, yeah, he's going to walk us through the conversations they're having with customers and the questions they're asking, which seem to be the right questions, if I'm honest. And then we can sort of, we also have a bit of a chat about how,
Starting point is 00:01:03 you know, LLM AI, selling that compares to selling machine learning based AI, if you want to call it that from, you know, a decade ago. So it's a all in all, it's a very interesting conversation. It is coming up after this week's news, which starts now. And look, we got so much wonderful, delicious chaos to talk about. But we're going to start off by having a chat about this mini Shia Lude worm. We've seen this worm originally pop up last year sometime. We talked about it at the time. You know, it's a self-propagating NPM worm. In this case, though, the initial access is a really interesting thing, the thing that started this all off.
Starting point is 00:01:43 And it wound up infecting TANstack, which is an extremely wide-used thing in the Dev ecosystem. I mean, James, you're the engineer among us. We're going to start off with you on this one. What was the interesting vector here? And can you give us a bit of background on TANstack? Yeah, let's start with TANstack, because this. is a complicated machinery in a complicated landscape. So if you're building a React app, you know, React is like one of your two fundamental decisions. React is the framework that you're
Starting point is 00:02:12 working with and you're probably writing the code in TypeScript. But that's kind of akin to saying, well, I've bought my block of land and I've got my plans for my house, but there's a heck of a lot of other decisions you need to make about how you're going to assemble that app. And it's things like, you know, what's going to handle the API routes, what's going to handle the state management, and all these things are an entire separate ecosystem of components that have built up. And TAN stack is a set of those components. They've become wildly popular, and they've sort of forged their own paradigm within the React community.
Starting point is 00:02:45 So that's what it does. It is a very integral part of building a React app. Now, the thing that is super interesting about the way that this initial attack vector happened here is there was no leaked credential, there was no fished credential, there was none of the traditional sort of ways in that you would expect for an initial compromise. Instead, it relied upon a malicious pull request making its way through a GitHub action. And that GitHub action was to the admission of the TANSTAC folks improperly configured. But it's just delicious how they did this.
Starting point is 00:03:22 We talked a little bit about this. We ran through the run sheet before we got recording. We do this every week, right? And the misconfiguration here was quite subtle. Like this was not like Tanstack did something completely suicidal and dumb here. They did slip up a little. But like why don't you walk us through the mechanics of how this like malicious GitHub action would wind up giving these attackers access to Tanstack's repos? Yeah, it is funny, right?
Starting point is 00:03:49 Because you pulled me up on the fact that I was approaching this from my software engineering perspective, which is like they tried so hard and they did almost everything right. And you said, buddy, you're in security now, and there's no prize for getting just about everything right. And sure enough, there is a particular GitHub action that happens on a pull request trigger, which is it's kind of one of the most dangerous areas where a GitHub action can operate, because it's essentially the moment when someone says, I've got a pull request, you know, a set of changes that I would like you to bring into your repo. And this action fires within the context of the repo that the change might potentially, be merged into, but can, if so configured, also pull in those untrusted changes that have been
Starting point is 00:04:34 proposed by the external third party in this pull request. The advisory from GitHub that's actually been out for a couple of years now, to be fair, does say this is a very dangerous footgun. If you're using this pull request trigger action, be very, very careful about bringing in the untrusted part of the code repo, because this is what can happen. And, you know, they they didn't follow that advice, and that's exactly what this GitHub action did, and that's how the attacker got their foothold. They got code run, they poisoned a cache that was then used during the deployment, and it was actually the legitimate deployment step that then pulled that cash out and resulted in the bad binaries being uploaded to NPM.
Starting point is 00:05:17 I mean, it does feel a little bit, though, that GitHub has given everyone a footgun and said, don't shoot yourself in the foot with this footgun. you know, Adam, let's bring you in on this one. I mean, this GitHub action stuff, it just seems like perhaps GitHub could be doing more, but I don't know because I'm not an expert in this field. Yeah, they absolutely could have avoided shipping people a footgun. I think we kind of hear what we are seeing is that
Starting point is 00:05:43 I don't know that anyone really expected the whole industry to coalesce around GitHub as the way of building everything. And, you know, one of the things that I thought was interesting about the story is that... Well, remember, sorry to cut you off there and ruin your flow, but you remember when Microsoft bought GitHub and everyone's like, it's over for GitHub, that's it. No one's going to use GitHub anymore. GitHub's dead, right?
Starting point is 00:06:07 Yeah, no, it's weird how much GitHub has become a critical part of everyone's infrastructure and flow. But yeah, so in this particular case, I feel like they did so many things right, as James was saying. but the GitHub action kind of set up like this, I don't know that anyone really expected this to become so important, right? And this particular footgun to be such an important thing. I mean, the idea of running a little bit of automation
Starting point is 00:06:36 when you push stuff in and out of a Git repo or make pull request or whatever else seems like a good idea. And all these little things, they all seem like good ideas until you start really crossing important trust boundaries and then building infrastructure that in very kind of near, yearish real time pulls in dependencies from other people and so on and so forth. And we end up with the impact of these, you know, kind of little choices turning into, you know, quite widespread compromise of code.
Starting point is 00:07:03 But I thought the trick of using the kind of shared cache between this way, the GitHub action can cache artifacts that they've built or whatever else they want to speed up later steps in the process in which when you are redownloading, you know, heaps of infrastructure or heaps of, um, dependence. and stuff. That can make a pretty material difference to how responsive your build pipeline feels. And developers love having their build process be really snappy because then they can turn things around really fast and it feels like you're doing stuff. That, you know, that, you know, kind of desire for everything to be snappy and responsive was the second kind of part of this,
Starting point is 00:07:39 right? I mean, the first thing is GitHub actions, you know, executing kind of macros, I guess, or triggers automation events that potentially have untrusted inputs. One problem, you know, the side effects of those are. untrusted inputs being processed. In this case, poisoning a cache that's then used later on is the second part of the puzzle. And then the third part, of course, is once it was executing in the release context, then it drops this worm that will propagate into other people's repos, steal creds. Well, let's go there because we've covered the TAN stack part of this.
Starting point is 00:08:10 But then from there, as you say, it dropped a worm which went off then and self-propagated. And it's like, you know, you just get on social media and it looks like people are having a bit of a hard time containing this thing. Like it is running real quick. You know, and we went through this last year and everyone's like, oh, we've all taken steps to slow it down and GitHub's like, oh, you know, we're going to, this ain't going to be a problem in the future. And like, here we all are. And this time, the people behind this worm have added some real nasty features. And one of them is if you try to rotate cred. So it will, it will, you know, hand off all of its API tokens and stuff on your machine to the attacker. If you try to rotate them, it detects that they've been invalidated and
Starting point is 00:08:48 RMRFs your whole drive, which is just like the nastiest thing to do. I don't even know why you would do that unless you're just like, I don't know, it's like a real anti-social personality disorder kind of thing to do with your NPM worm. But like this thing's out there now and causing all sorts of drama. But what to what end is this, is this, you know, sitting above this whole thing? Is this going to be something really dumb like someone trying to steal cryptocurrency or something? Because that's the vibe I get here. It's very hard to tell because again, this all amount. to stockpiling creeds that we then go for what and for where
Starting point is 00:09:22 and so similar to PCP it might be a little while until we see either the follow-up actions that they decide to do or they just kind of farm the creds out and let other actors have a go with them but it like it is odd that we don't know where this is going but it'll go somewhere
Starting point is 00:09:38 the other thing I just wanted to add in here is I had a real cold sweat moment of panic this morning when I read this because I thought to myself it's okay I haven't done like an NPM install or a bun install for a while now. So I know I didn't pull in packages during this, you know, six minutes, six whole minutes where this was ablaze. But then it dawned on me that the way that coding agents work now, both codex and also Claude, they work in Git work trees. And so when you farm your agent off to go and do a task, it's working in a new directory that doesn't have
Starting point is 00:10:10 all these modules installed. And so time after time after time throughout my day as I'm interacting with agents, they are pulling these packages over and over. again. So yeah, folks need to think you might not have done an NPM install during this window, but which one of your agents was working on something that went and did it? You might not know that you did that. You might not know. That scared the hell out of me this morning. Yeah. All right. We'll talk about our incident response later. But yeah, probably time to move on to the next story now. And funnily enough, the next story we're going to talk about is this breach of Instructure and the learning platform canvas,
Starting point is 00:10:48 which is used by K-12 schools and colleges worldwide. I mean, it's reported as being an America thing, but I can tell you that universities here in Australia and across in your country, New Zealand, Adam, have been like delaying exams and things and dealing with this. Funnily enough, though, it being the big story, I don't know that we've got much to add here. I mean, it was a small breach that initially instructor was like,
Starting point is 00:11:11 oh, yes, you know, attackers tried. to compromise us but we have contained it and yeah it turned out not so much a lot of their data got walked including according to the attackers billions of messages between students and their educators and then of course there was a you know ransom note kind of dropped on a shiny hunters ransom note dropped on the login page for this system reporting suggests that they have actually paid now to get the data deleted and you know I mean that's pretty much the end of it looks like shiny hunters managed to rack up a win here. Adam, any thoughts? Yeah, I mean,
Starting point is 00:11:45 just paying them out feels bad, but on the other hand, you see the amount of pain it was causing you can understand why that was a decision that they were going to consider. And especially in cases where you've got a supplier and their customers, and the customers are all applying pressure to the supplier to do something, anything, and the only fast option that a supplier has is to go pay. And that feels bad, but you understand why. And, you know, I don't know, hopefully very little of it actually ends up in the pockets of a shiny hunter. But, you know. Well, you hope they slip up doing their money laundering or something, right?
Starting point is 00:12:20 Like, that's the vibe I get here is that that could be, that could well be how this ends. Because shiny hunters definitely has like UK teens vibes, right? Yeah, and I can't imagine their money laundering slash, you know, kind of money handling slash not going out and immediately spending it on, you know, on trash that's immediately kind of draws attention to them. doesn't feel super likely. Like, they may be good at the tech stuff, but that doesn't necessarily translate the good at, you know, long-term crime. I mean, everyone always criticizes people who pay, but I've said it on the show a million times, right?
Starting point is 00:12:51 Like, sometimes it's existential, and they've got to pay. Like, and it's really hard to say that it's uniformly wrong in every circumstance, which is why I was against legislative proposals that would have outlawed paying ransoms. Yeah, and I guess that's why I cast it as feels bad as opposed to is the wrong thing to do, you know? Yeah, exactly.
Starting point is 00:13:10 feels gross but what are you going to do well we're going to stay with you on this next story Adam because this one has you written all over it I thought of you as soon as I saw this um how how how how to begin so um someone screwed up in Germany uh rotating a key signing key for uh DNS sec which meant that the entire dot d dot dd dot d TLD was returning serve file on DNS queries where there were they were they were signed zone files right So they broke the chain of trust for all of .de. And Cloudflare basically its decision, and I agree with this decision. I think it was the correct decision.
Starting point is 00:13:50 And apparently there's an RFC backing this decision as well, but they just let the whole thing fail open, right? So basically they just switched off for the 1.1.1 cloudflare resolver. They just switched off DNS sec validation for all of Germany because of this screw up. And I think, look, absolutely the right decision. But it kind of goes to show you that like, this isn't really news. You know what I mean? Like I'm reading this straight off the Cloudflare blog. The fact that an entire TLD and a big one at that just turned off DNSSEC validation and everything was like no one noticed.
Starting point is 00:14:25 Kind of tells you what you need to know about whether or not the DNSSEC juice is worth the squeeze in my view. But I really came for your opinion on this one. Yeah, it's a pretty interesting tale, this one. And I think you kind of summarized the guts of the technical aspects. The interesting way it with Cloudflare is survey their 1-1-1-1-1-resolver does enforce DNS and validates the domains that it's answering questions about. And the correct behavior in the situation was to return an error, right? Return serve, fail and, you know, not answer the query.
Starting point is 00:14:58 And Cloudflare very rapidly realized that, you know, that was worse than just answering the queries and marking them as not secure. you know, saying like, here's the answer, but we're not, you know, it's not DNSSEC validated, which is what they ended up deciding to do. Well, and thankfully, Adam, thankfully, all of the software that we use and rely upon out there is set up to really take note of that note in the Return Zone file that says this isn't secure, you know? And, you know, this really changes a lot of things.
Starting point is 00:15:26 We, but I know we play with a bunch of DNSSEC stuff over the years and, you know, making the, like it's one thing to sign your zone files and publish it for other people to validate. It's a whole other thing to say, we are also going to make all of our queries fail if DNSSEC isn't available or, you know, should be available and isn't. Because it's just going to break stuff. And the amount of breakage that DNSSEC causes versus the amount of, you know, kind of impersonational, you know, um, cache poisoning or whatever other things that's trying to prevent, really the, you know, the impact of the, you know, the impact of DNSSEC is mostly about bad availability and not about integrity. And yeah, does the juice seem worth
Starting point is 00:16:10 the squeeze for DNSSE? Really not. And especially now that we have so much other crypto layer over the top with TLS. Let's encrypt one. Let's encrypt one. And the thing is like, I guess one of the reasons that I like to beat up on DNSSEC is the proponents of DNSSEC, like the real rabid ones, are among the most annoying people you'll ever meet in your life. Right. I mean, DNSSEC clearly grew out of the kind of cyphapunk way of thinking, right, where we should make it perfect without really accepting the reality of the world that we have to live in. And yeah, it's, I mean, DNS itself is just old tech and then bolting crypto into old tech, you know, we just, you know, brittle is the end result, right? We have let's encrypt. We have modern browsers. Let's move on. Yeah, and it's
Starting point is 00:16:54 not perfect, right? I mean, the let's encrypt world and the browser, like, letting, delegating this all out to TLS, you know, isn't the best solution. but it's the solution we've got and it's the only one that's you know kind of viable in the real life I'd argue Adam that actually in that everybody uses it it is actually the best solution unlike the NSSEC which people don't actually use but anyway anyway we can argue this one we could argue this one more at great length over a beer one day moving on and Google's threat intelligence group has released a report all about you know what's happening out there with adversaries and whatever and not surprisingly AI features very heavily. I guess the item that they spoke about here that's been talked about most is they discovered some threat actors had used AI to uncover an ODA in some sort of web IT administration tool.
Starting point is 00:17:50 I don't know. That sounds like C panel to me. I don't know. But it's like an MFA, something like that, I guess. But it's an MFA bypass bug that they found. And they were able to, I guess, disrupt the active. from being able to do widespread exploitation against that. So that's great, a wonderful win for Google.
Starting point is 00:18:06 But the thing that's remarkable to me is looking through the executive summary of this report, and it's all stuff we've been talking about, like, at length, on the show for months and months and months. James, you've been through this one. That was your take as well. Yeah, exactly. You know, they open with that exact statement that this is really just a trajectory from Nassent AI usage by attackers, which is where we were from the last report to now this is, you know, to use their term, industrial scale application of generative models within adversarial frameworks.
Starting point is 00:18:37 But the nice thing about it is they sort of break it down into six headings, and each of them is a very sort of targeted look at where AI is being used for vulnerability discovery, AI augmented development of defense evasion, autonomous malware operating end-to-end with AI. There's a good section in here as well about the obfuscated LLM access, and I think that's something that needs to get a whole. whole lot more attention throughout industry is just like how do we prevent the large-scale use of LLMs in an unauthorized sense from these bad actors through things like, you know, chat interfaces
Starting point is 00:19:09 or other half-baked LLM's being shoehorned into this product, and it's accidentally a really nice distillation vector for an attacker. But overall, not like a, it's not a super thrilling report, but it's just really great to see this all condensed down in one place that says, yes, this is happening, and the trajectory is we're now at industrial scale, and let's see where this goes from here. Yeah, I mean, Adam, you would, as someone who has spent your entire career, basically working in off-sec, I'm guessing you would have found this one pretty interesting. Yeah, yeah, it's a good summary of kind of where things are at on, all the various places you can use the tooling. And, you know, I think much like James, the idea that we can control access to models as a viable kind of strategy for mitigating the various ways that it's being used doesn't seem like a really robust path forward. but I thought it was just, you know, this is a great roundup of the state of the world,
Starting point is 00:19:59 and obviously they have insight by virtue of being both incident response, but also operating one of the big models, and they can look at the arts being used and so on. So, yeah, it's always interesting reading their work because they have that kind of both ends perspective on it. Yeah, and if you want to know about the state of the art in terms of using LLMs to do vulnerability discovery, last week we spoke about some work from Niels Provost, who's an old school, you know, security head, who did some work in instrumenting and orchestrating LLMs to do Voldev in a way that was like as effective as mythos, even using local models and older models.
Starting point is 00:20:36 That was some very interesting work. We did talk about it last week, but since then, James did a 90-minute interview and discussion with Niels all about that work, which is available in the risky business features feed. So again, I know I've been banging on about it every week, but if you are not subscribed to that feed, you are missing some really good stuff. So head over to either risky.biz to get the links, the subscribe links,
Starting point is 00:21:00 or you can just fire up your podcatcher and search for risky business features. But that is a fascinating discussion. And I've also linked through to it in this week's show notes. Now, a big thing that happened to Adam while you were away for a few weeks is every week, it was like the agenda seemed tailor made to the special guests that we had that week, right? Just incredibly well tailored. And then you come back and it's the same, right? because we got a whole bunch of really interesting bugs to talk about in stuff that you know very well.
Starting point is 00:21:28 So first of all, there's the Dirty Fragg bug. We spoke about Copyfail last week, and James, I believe you're going to correct something that you said last week that was incorrect about that. But there's this new one called Dirty Fragg. We've also got bugs popping up in like free BSD and whatever. A lot of this feels very AI driven. And I just wanted to get your thoughts on these bugs, Adam. I mean, I really enjoyed copyfail. It's a beautiful bug. and I went through, read some of the coverage of that when I got back from my holiday and wanted to refresh my memory about what the cybers was all about, and it just felt so familiar in so many great ways.
Starting point is 00:22:05 Dirty Fragg is essentially just another variant of the same bug. The guts of copy fail were that you could write, you know, corrupt the disc cache in the kernel. You could, you know, override data stored in the kernel's idea of cased files off disk. and that was done through in that particular case, so, you know, something into, like, encryption plumbing somewhere. This is another couple of vectors in the kernel that you can use to write to the page cache in ways that are surprising
Starting point is 00:22:36 and use that for local prevesque. And, like, they're beautiful local previsk bugs, right? You don't have race conditions, you don't have memory corruption. Like, it's targeted, repeatable. It's exactly what you want in a kernel local previsk. Because the last thing you want is, you know, bugs that are going to cause instability. You want things that are super reliable.
Starting point is 00:22:55 And so we love a Linux privisque. We've had to think about Linux being essentially single user from a security point of view, probably for the last 20 years. It's never really been safe to have multi-user Linux boxes. And it's nice to see that sort of reinforced for everybody. But I enjoyed both of these bugs. Actually, Dirty Frague technically is kind of two different instances
Starting point is 00:23:19 of a repeat of the bug, that one that works well on Ubuntu, one that works well on everything else. But yeah, they're well worth reading and understanding. And I felt, yeah, it felt nice seeing something so near and dear to my heart, you know, on the run seat this week. Well, and then there's the free BSD one too, which is also pretty hilarious. And I mean, these are all AI discovered bugs. As far as I know, I think this free BSD one was a mythos discovery. And I think one thing when I was doing a little bit of research on it, I, you know,
Starting point is 00:23:47 plugged it into Google or whatever. and I think Forbes was running a story saying Mythos has found a bug in one of the world's most secure operating systems and I'm like, man, it's 2026. Like, come on. Who did that headline? But yeah, walk us through this one as well. Yeah, so the free BSD book absolutely feels AI discovered and it feels a little unsporting, honestly, letting an AI look at 26 year old or whoever old this, you know. But it's the most secure operating system in the world according to Forbes. They may have confused their BSD variants there perhaps. Anyway, the particular bargain question is that a malicious DHSCP server can set a value that gets written, the set of value that's given to the DHCP client, that when the DHSDDDB client writes it into a cache file on disk for later use, you can kind of, you know, this is incorrectly escaped meta characters, you can inject, you know, more directives to the DHCP client. And then next time it runs and repasses that file, it interprets those directives and you get code exec.
Starting point is 00:24:46 And that feels like an AI discovered bug because, you know, you know, sort of chaining that logic together of how you would use it and what it's good for makes a lot of sense. The thing that I really liked, though, is this bug, so writing into the lease cache file on disk, dirty frag, writing to the cache files, writing to the cache in memory, and that bug way out front with Shai Hulud, all three of those are cache poisoning. and I'm reminded of that like classic amorphism about there being two hard problems in computer science one is naming things the other is cation validation and that definitely felt like yeah that that is ringing through this week for sure yeah yeah
Starting point is 00:25:30 meanwhile James you wanted to correct something when you were talking about copy fail last week you said you got one of the technical details role yeah so when I read the the publication from the theory folks there was mention of using the what was the IPSEC encrypted sequence numbers, there was a bug in there that basically resulted in a predictable
Starting point is 00:25:49 four-byte right outside of boundaries. And I assumed that that meant that that was just a really simple like buffer overflow in that. And on one hand, you think, okay, neat that they found that, but you also think there's so much tooling and stuff that should have caught things like that. And so I think we'll link to it in the show notes, but there's a great write-up from retro. dot zip retro with a zero where they actually sort of did a bit of a record scratch of like it's not your average four bite right out of bounds and they go deep into this and it is crazy the level of sort of hoops that were jumped through to just get this four bite right into the page cache so superb work and glad they took the time to really explain it yeah that right
Starting point is 00:26:32 up is absolutely worth the read if you want to understand the specific details because like it explains it so well i yeah i definitely recommend that one yeah and we've lived through to that write-up in this week's show notes. Now, look, we're talking about AI discovering flaws in other things. There's a fix just gone out for the Claude Chrome extension, which would have enabled other plugins, I'm presumed they mean extensions by that, to hijack the Claude extension there.
Starting point is 00:27:00 Is that about right, James? Yeah, it's, you look at this one, and you just go, Claude does not belong in a Chrome extension right now. because it's just so simple how this was done, right? So if you've got the Claude extension running, someone else, or sorry, someone could load another extension into the browser that had no permissions, no elevated permissions whatsoever,
Starting point is 00:27:25 but just had some nasty code in there that was interacting with the page. And then, you know, if the browser goes to claw.com, for example, the malicious extension, all it has to do is just inject something into the DOM, which is the bread and butter thing that all of these extensions do, and then Claude sees that thing that's been injected into the DOM and reads it as just a prompt. It's like if it's this easy to trick Claude into reading a prompt out of the DOM in a browser, get that thing the hell out of an extension.
Starting point is 00:27:55 The only safe way to really use a model these days is for you to be the sole source of input into the initial prompt, right? That is like the cleanest guard rail and surface we have at the moment is because that's when you as the human express, your intent, your instructions. Yes, we pull in a bunch of skills and other things along the way, but you've bootstrapped that you've set the task and that's generally what the model will follow. But when you take Claude and put it in an extension that is reading the DOM and that can become the prompt, nothing good is going to come off this, my friend. That's going to be a bad time. And meanwhile, Google copying a bit of flack this week for shipping everybody, a four-gig, like, local version of Gemini with Chrome,
Starting point is 00:28:38 where really people just woke up and their computer had seriously just downloaded this four-gig update. Funnily enough, I mean, look, this is to be expected, right? This is the way the world is going. And I've included a story in this week's show notes from Lily Hay Newman over at Wired. The headline is, you can disable Gemini and Chrome if it's freaking you out. And the reason I really wanted to include this is this is going to be a future historical artifact, this article, where people will be like, wow.
Starting point is 00:29:03 People thought they could avoid AI. It's like saying, you know, here's how you can have a clean install of Windows and disable all web browsers, right? Like, you don't want to use those web browsers. In a browser, you used to be able to turn off JavaScript, and that was a legitimate decision that people would make. It's the same sort of lineage. Yeah, exactly, exactly.
Starting point is 00:29:20 But, I mean, what do we think about Chrome doing this? Adam, I'd love your thoughts here. I mean, I think the argument for having that tiny model in there was actually pretty reasonable, like being able to do certain things that you don't necessarily they want to shove off to the cloud. Obviously, Google, by making you run the model, even with a really little one, saves them quite a lot of compute, I would imagine, rather than having to, like, make a safe browsing, make a call off to a Google AI service
Starting point is 00:29:46 every time someone visits a page. No, no, no, no. So your argument is reasonable, but the counterpoint is, which. It's a witch. Yeah, I mean, the amount of things that are going to have, like, AI models stuck in them, like, as you say, it's just like, we may as well say, let's not put DL files and things anymore. You know, that's the kind of the level that we're at these days. So, like, I understand that some people have, you know,
Starting point is 00:30:10 concerns about AI just generally as a concept, like that the training of those models of the first place was unethical or whatever. And that's kind of, I can respect that kind of point of view of it. But if you're a technical user and you want to turn off this piece of functioning out in your browser, you're kind of on a losing Wicked, I think. Yeah. I mean, I think there were more solid arguments. You remember when Sony started putting, like, basically malware on its,
Starting point is 00:30:33 music CDs back in the day where if you played one of their CDs, it would like Trojan your box to put all of this crazy like, you know, kernel level like anti-copy stuff on your computer without asking you. Like, you know, that one, okay, I think that we can say that one's over the line, but like a browser shipping a model, get used to it. Yeah, pretty much. Yeah. Now, look, we've just got to like, I've got to rub our temples. Now, we've talked about like AI bug discovery. That's a very big deal. And we've also. talked earlier in the year about like how AI is being used to orchestrate attacks and scale them up and whatever and you know while we're on that topic bugs in Palo Alto and Yvanti still I mean
Starting point is 00:31:15 this is the sort of stuff that's going to get absolutely auto owned by orchestrated AI agents which I am now referring to as infinity E script kitties because that's basically what these agents are but my god man like at this point you know running stuff like Yvanti running stuff like a lot of the Palo Alto gear. It's just, you just, it was risky enough before and now it's just suicidal. Yeah. I mean,
Starting point is 00:31:42 they're being yet more bugs in a Vanti endpoint manager. Like, how is that even possible? How is there any code left that is, like surely they must have got rid of all of it by now. There can't be anything left. But I don't know what the specifics. Well, as Gruct said,
Starting point is 00:31:55 and you weren't here and I doubt you listened to it because you were on holiday, but as Gruck said here a few weeks ago, when talking about Mozilla patching 271 bugs that were found with mythos in Firefox. He said infinity minus 271 is still infinity. And I think that just applies here. I think so. Yeah, gruck with the wisdom as usual. Yeah. I mean, ultimately, the thing that stood out to me about the Avanti story was the C-so of Avanti came out and said, look, we just want people to understand that we are trying to do the right thing. It's like,
Starting point is 00:32:24 nobody, the time to do the right thing was 30 years ago when you stopped investing in the security of this product, all the people who bought it, who bought it, who bought it, you know, like many corporate acquisitions ago. Like that's the problem is that we're running 30-year-old trash code and expecting it to be robust against modern internet, modern AI, much like that 3BSD DASHC client being unsporting as a target. Like at this point, Avanti, like it's just, like it feels like you're kicking a puppy at this point.
Starting point is 00:32:51 Yeah, well, and this palo bug too, Adam, is quite awful, right? So I haven't seen, I was looking around for a POG. Apparently there is a POC, so this was a Palo Alto remote code exec. It appears to be. memory corruption in a content length header in the year 2026 and a thing that parses web content for a living. I mean, that's Babi's first exploit, really, isn't it? It really is.
Starting point is 00:33:15 So there's a slight kind of mitigating factor in that this is in their like captive portal bit, which there's probably no reason to have internet facing, but of course people will because, you know, why wouldn't you? But yeah, come on, like memory corruption in a security appliance in the year 2026. And like, there should be defense in depth and exploit mitigation and all these things, but security appliance vendors haven't had to update their products for 20 years. And so, you know, why rel row and pie and do all of the other export mitigation stuff in here that would let you get away with having mem corruption in your continent theater?
Starting point is 00:33:48 But no, that's just palo Alto life. So everyone's at least used to patching this stuff. So that's good, right? Yeah, James, you look like you had some feelings there. To that point, everyone's used to patching it. but the thing that I got a good giggle out of was when I read the Avanti advice, it sort of is they bifurcate what they tell you to do based on how well you responded to the last time you got owned on this box.
Starting point is 00:34:11 So it's like, you know, if you responded correctly to our advisory in January and rotated your credentials, you need to do these steps. If you didn't, you are in this bad state. And if your box wasn't compromised in January, you need to actually do this instead. So I just love that it's now like a, you've got to look back over your history of, you know, Not just what do you do with this box now, but... So they released a choose-your-own-adventure to accompany their advisory, basically. Decision tree of fail.
Starting point is 00:34:35 It's great. Moving on. And Russia is launching kind of its own version of Starlink. It looks like there's a great write-up of this in Widen. We've linked through to it in this week's show notes. But it looks like it's going to take them a while to get this thing up to being quite reliable. But even an intermittent satellite connectivity. over a battlefield is actually going to be quite useful.
Starting point is 00:35:01 But what's amazing is like how quickly this sort of capability has been understood to be very important by major companies. I mean, we've got the Europeans essentially launching their own thing and doubling down on their own thing because they're a little bit worried about continued access to Starlink because of Elon Musk and the United States and its attitudes towards various things. And no doubt the Chinese will be working on their own version of this. But it just seems like, you know, this is absolutely a very important capability. and they're going to do it.
Starting point is 00:35:28 I mean, there's some delicious details, though, in this write-up about how they're, the hardware for this Russian version of Starlink is like multiple times bigger than anything that comes out of SpaceX, right, which shouldn't be all that surprising. But, you know, they're giving it a crack. James, I know you've been through this one, and I guess one of the things that you and I both zeroed in on is the orbit path for these satellites is quite different to those from the other companies, which, I don't know, there's some. That could be interesting in terms of, well, could you shoot these things down and have it not actually damage your own satellites, for example?
Starting point is 00:36:05 Exactly. Yeah, there's two very interesting differences. One is just the count of the satellites, you know, whereas Starlink has thousands and thousands of these things up there to get full global coverage, especially around densely populated areas. The plan here is to only launch, well, only 300 satellites by 2030. That alone will require them to churn out a satellite or two a week, which seems a little bit ambitious. So it's a much smaller number of satellites, but the orbit path, because they want to essentially cover Russia and its various territories around there, they can get away with an almost polar orbit on a certain incline that means that they're going to be out of the way of the other satellites. And they're also operating at a much higher altitude, about 800 kilometers versus 500 kilometers. So it sort of settles an argument you and I were having because I said, you know, there's no way that that, that.
Starting point is 00:36:57 an adversary would shoot one of these down because you'd take out all the satellites and then, you know, we'll never go through space travel again. But, you know, if it's only 300, that doesn't sound like you actually need to disable too many of them. And even if you had to do something kinetic, it's probably in a flight path where that debris won't be too dangerous. So, yeah, gosh, when everyone's got these, it's going to be interesting to see what happens if there's a kinetic conflict around them. Well, the next World War is going to happen in space, I guess, as well as, as everywhere else, I guess is where we landed on this. But I think also, you know, there's probably ways
Starting point is 00:37:30 to disable these types of satellites without creating a debris field, whether that's just putting a hole in them with a laser or burning out some of their sensors or antennas, you know, there's probably ways to do it. And no doubt the very smart people over at US Space Force working on exactly this problem right now. But yeah, just an interesting sign of the times.
Starting point is 00:37:50 Moving on, we've got a report here from TechCrunch from Zach Whitaker, which is looking at a Latvian hacker named, Dennis Zolotajovs. This guy was doing ransomware stuff, working for a Russian ransomware gang called Karakurt. The interesting thing is here, though, that the indictment sort of spells out links
Starting point is 00:38:11 between this group and the Russian government, basically. So it's really good to have some of those links spelled out in a more explicit way than we've had previously, and that's just why I've flagged it and put it in this week's show notes. We've also got a report from John Grieg over at the record talking about how the muddy water crew, which is an Iranian APT crew, essentially, they've been dropping chaos ransomware to kind of cover their tracks, which I don't know.
Starting point is 00:38:38 That's not so surprising. I mean, it's right there in the name, you know, muddy water, trying to muddy the waters on attribution. Although James, you pointed out to me that this does not look like a particularly, you know, effective way of obfuscating attribution, given that they're signing YouTube. with certificates that people know they use, for example. Yeah, it did seem like well-intended, but perhaps poorly executed. But I think, you know, you made a good point to me, which is that even if you get to the
Starting point is 00:39:09 point of realizing, huh, this, you know, ransomware was actually signed by, you know, muddy water, you've already had to deal with that ransomware and all the impact, and that's what they're aiming for, is just to slow you down to distract you so they can get on with doing what they want to do otherwise without being noticed. So, yeah. I feel like this will actually work as an obfuscation against people who don't call mandiant, which is most people. But if you do call mandiant, they're going to pick it apart and figure it out.
Starting point is 00:39:34 I mean, is that your vibe too, Adam? Yeah, yeah, pretty much. But I think like the specific value of this, you know, in, you know, in terms of it being believable or whatever, is kind of less important than, you know, the goal is to get people onto the ransomware off the playbook, right? They might have a playbook for how to respond to ransomware. if you can shunt them over under that, you know what they're going to be doing
Starting point is 00:39:56 for the next three days, and it buys you time. And like anything where you can like manipulate how your adversary, well, how your victim, I guess, I'm still thinking my off-sec life, anything where you can manipulate
Starting point is 00:40:09 how they respond gives you predictability. And that's just really important. It doesn't have to hold the long. It might just be long enough for you to, you know, action on objectives or whatever else. So, yeah, I think it's, you know, it's still a worthwhile trick. Yeah.
Starting point is 00:40:21 And speaking of, Ransomware, Foxcon has confirmed there's been some sort of attack against its factories in North America. Looks like a limited impact, though. I mean, from what we can tell so far, it looks like they had a bit of drama with, you know, the Wi-Fi didn't work and they couldn't use their computers and whatever. And I don't know if that's ransomware or if it's people like, you know, if they're pulling stuff down and responding or whatever. But it looks like they're all back up now. Many such cases these days, right, where you actually see, you know, there's an initial foothole. There's a bit of drama with a ransomware attack and then things are quickly brought.
Starting point is 00:40:53 back to normal, so it looks like perhaps that's what's happened here. But one story I wanted to talk about in a bit more depth. John Greig has this write-up in the record. There is a CISER initiative where it's called CI Fortify. And the idea here is that they're going to help critical infrastructure operators to figure out how to operate offline. Now, this could be because there are DDoS attacks happening, targeting critical infrastructure.
Starting point is 00:41:19 It could be because a campaign like Volt Typhoon is starting, to do unthinkable things to US critical infrastructure. Could be for any reason like that. But I think this is a really good idea. I think in terms of like resilience, you know, going to a utility, going to a bit of critical infrastructure and saying, look, how dependent are you on US East One? Right. And that link being active in order for you to actually be able to operate your service.
Starting point is 00:41:44 And like, can we build some contingencies here? Funnily enough, James, at one point in your career, you kind of went through this. with a critical infrastructure provider. And where it's landed is, okay, so for the really insecure utilities and whatever, going through this process is going to be a really good thing to do. But you made an interesting point to me, which is for the people who really know what they're doing, for the highly sort of secure environments,
Starting point is 00:42:12 having to go through an exercise like this actually winds up introducing a fairly large amount of new attack surface because there's all of a sudden a lot more new equipment and a lot of redundancies. So, yeah, there is a downside, I guess, for some operators was the point you were trying to make to me. Yeah, I was looking at it from the perspective of if an organization has gone through, you know, their cloud transformation and moved a lot of on-prem workloads up into the cloud and they're heavily dependent on, you know, telco infrastructure that is managed for them. That's the kind of stuff that is straightaway is going to be in the crosshairs if they're, you know, challenged to say, well, you know, how do you operate if all of that stuff is gone away? inevitably, and this is the situation I went through, is we had to start bringing some versions of software,
Starting point is 00:42:57 really key critical things back on-prem, in a hybrid cloud and on-prem model. Now, that introduces complexity, right? It's one thing to go from cloud to on-prem, but if you're having to manage both active standby or active-active on-prem and cloud at the same time, you've got complexity, you've got additional equipment, you've got an additional configuration,
Starting point is 00:43:18 and all of those things when you're not in the offline mode become really excellent places for an attacker to get access to and to dwell. My point is I think this is still, I mean, you have to go through this, right? It would be my point, which is we live in a world where these links, these data centers, I mean, we saw in Iran, right? Iran actually attacking Amazon data centers, for example, right? When there's going to be a conflict, this infrastructure is going to be targeted. And we need to make sure that our critical infrastructure is resilient. Yes, it does cut both ways.
Starting point is 00:43:48 I don't really see that we have another option but to do this. Adam, what's your take on this? I think this kind of exercise is really useful and important because we build this stuff so quickly, all this infrastructure, everything is all high tech very quickly, and we haven't really thought about the failure modes. Now, what you do with your, you know, if you go ahead and do this kind of exercise
Starting point is 00:44:07 and you understand now what the potential failure modes look like, what you do with that is another thing, right, building redundant infrastructure or offline, bringing stuff, you know, back out of the cloud. like those are things, powers that you can go down, but at least understanding what the potential impact could be so that you're not exploring that the first time when it actually happens. And I think in New Zealand, for example,
Starting point is 00:44:28 we have one, essentially, one or two kind of big fibre links in and out of the country. If those go away because someone does a little submarine snippy-snippy, which, you know, great powers have been known to do in times of conflict, little snippy-snippy, and like we don't understand. And like, sure, there's a bunch of things we could do. it's probably prohibitively expensive. We're just going to accept the risk. But at least understanding that all our national payment systems are going to stop working
Starting point is 00:44:51 without that piece of fibre, like that's good knowledge. And it's worth doing these kinds of exercise to know, are we going to have power? Are we going to have water? You know, because you don't want to be doing this for the first time for real. And of course, this will happen because the first objective of a superpower during a great power conflict will be to take New Zealand out of the war. strategic dagger pointed at the heart of Antarctica, yes. That's right.
Starting point is 00:45:17 Now, we're just going to stay here with our sheep, and because we're not exporting all of our food anymore, we're going to have plenty. It'll be fine. I hope. Just no diesel. Yeah. Yeah.
Starting point is 00:45:27 What is it, the New Zealand defence strategy of, well, they have to go through Australia first. Thanks, funny. Your taxpayer dollars buying F-18s, thank you very much. F-35s, thank you very much. We get the good stuff. I wanted to link through to a report, which unfortunately, unless you're a 404 media subscriber, you won't be able to read it on web. If you are an email recipient, you would have received this one so you can dig it out of your inbox.
Starting point is 00:45:55 It did go out for free as an email. But Joe Cox managed to get his hands on the software, a bit of Chinese software that's used by people doing scams to do deepfakes, right? So you can grab pictures of people, send them to these guys for 500 bucks. they will create like a model of this person you're trying to impersonate and it will do it in real time on Zoom, on WhatsApp, on whatever you want. A really fun write-up of like Joe's adventure in like going and procuring the software and they even like remotely set it up for him. They set up a partition on his on his computer and in they came.
Starting point is 00:46:31 Remote support. Very slick operation. James, I mean, you know, I know you enjoyed this one as well. I know I did. A couple of things really jumped out of me. The first is, yes, it was such a white glove, high-touch service. So it's beautiful to see they've considered customer service. But didn't take a lot in terms of hardware specs.
Starting point is 00:46:50 They demanded a I-7 processor 16 gig of RAM and an Nvidia 4080. That's not the kind of spec that I thought would be needed to pull this off. So that was quite surprising. But the thing that is like just this moment when you go, oh my goodness, we're in trouble is when you read the part that says, you know, exception spelled with an X, which is a deep fake detection model, it struggled, they say in inverted commas, but struggled was actually almost 100% of the samples from this software they acquired was mistakenly labeled as authentic despite this research being the state of the art of deep fake detection.
Starting point is 00:47:30 And the videos look good. There's still a little bit of uncanny valley, but the fact that it works when, there's something in front of the face it tracks lighting differences it's gosh it's it's getting good yeah I wonder I wonder I mean we've had persona on a couple of times right and they talk about how they do real-time video and stuff and that's a big part of how they actually do like K YC style you know verification of identities and it's being used in the enterprise and whatever and you you worry how enduring that sort of approach is going to be and now I'm sure they've got
Starting point is 00:48:02 labs and they're cooking up all sorts of ideas and detections and whatever but But this, you know, remotely verifying someone's identity is correct is a wicked problem. It has always been a wicked problem and it's going to remain a wicked problem. And I feel like we've had an easy run of it with video recently and now that's kind of gone. And our final piece this week, which is, you know, I guess we'd call it our skateboarding dog. We spoke about how the FCC in the US is going to ban foreign-made routers and it looks like they're pushing through with this. But it looked like they were also going to ban foreign routers. from patching and issuing patches from like March next year,
Starting point is 00:48:40 they've now realized this is not a great idea. And they have pushed that patch ban out to 2029. Still a bad idea, but it's further away. They've also reversed the ban on patches for drones. So well done, FCC. So many bad ideas, but yeah, bad idea two years down the track we've got a chance with. It's the bad ideas this week that these days we are reduced to struggling with. So I guess good news.
Starting point is 00:49:06 And we'll check in on a couple of years and see, you know, what they do then. Yeah. Right, guys. Well, that is it for this week's news. Great to have you back, Adam. James, great to chat to you as well. And, yeah, we'll do it all again next week. Yeah, we certainly will, Pat.
Starting point is 00:49:20 I'll see you then. Yeah, thanks, Pat. See in a week. That was Adam Boyalo and James Wilson with the check of the week's security news. Big thanks to them for that. It is time for this week's sponsor interview now with Bobby Phila, who heads up AI over at Sublime Security. If you are not familiar with Sublime Security,
Starting point is 00:49:39 It is the modern Whizbang AI-enabled secure email platform or email security platform. So, you know, if you need to filter out BEC, if you need to filter malware, fishing links, things like that, you know, it is that sort of platform. It is the most modern iteration of one of them. It's also highly inspectable. You can write custom rules for it. Or you can just get their AI agents to do that for you. That stuff actually works really well. Crazily well.
Starting point is 00:50:04 In fact, but Bobby joined me to have a bit of a broader discussion about AI in the cyber security. marketplace. You know, when people are evaluating cybersecurity solutions that use agentic AI, what sort of questions are they asking? What are the things they want to know? And let's just start it there. But we go on to talk about a few other things like how this is a bit similar to the machine learning craze of like 10 years ago. It's a fun interview. So here's Bobby Philar talking about how customers go about evaluating agentic cybersecurity platforms. Enjoy. Yeah, I think they go about it a few different ways. And honestly, the easiest one is just asking questions, right?
Starting point is 00:50:44 How has this agent been trained? What is its background of knowledge? Do you use evaluations offline, online evaluations to monitor performance? When we're talking about things like agentic use and autonomy, has it been red-teamed, right? Like that's a real, real situation that folks need to consider at this point is these agents can do a variety of things. They have different skills, different tools they can reach into. And if that hasn't been thoroughly tested internally and externally, customers are understandably wary of that. And then as you move down the line, I think it turns into what's your methodology?
Starting point is 00:51:26 What is the reason for building this agent in the first place? Like, what problem did you identify where you felt? like I needed this. And those types of questions, I think, really suss out whether or not a vendor is bolting on, something that is just kind of an afterthought, checking the box, or whether or not they're building it with good intentions, this idea of up-leveling the customer, giving them an opportunity to grow with the product until the point where they feel comfortable releasing some of their day-to-day responsibilities to it. Now, do you feel like there is, And I definitely feel like this is something that's happening out there.
Starting point is 00:52:06 Do you feel like there's some AI fatigue among buyers at the moment? Because I feel like everybody has like bolted some sort of AI function onto their thing. They're like, where an AI platform? Because like that's what you have to do at the moment, right? It's just what you have to do. So I feel like if I'm a buyer at this point, I'm like, oh, you want to pitch me your AI solution, do you? Oh, great. Like I haven't had like 10 of them this week.
Starting point is 00:52:31 Yeah, it's interesting. So I grew, I kind of got my background in the heyday of early machine learning being introduced in security products. I'm from the endgame. A lot of my security AI friends were in like CrowdStrike and silence and things like that. It was always really funny because you would go to RSA, you'd go to Black Hat and be like, I'm the person who uses math to catch malware and they're like, no. Just like, and you probably remember snake oil booths and things like. like that popping up. And it was it was kind of a joke for a while. And it was a it was a tough one.
Starting point is 00:53:07 I mean, it always it always worked though. And I think I think AI is kind of the same. It's interesting that you mentioned this, right? Because like Ryan Permay, I know still I'm in touch with Ryan who was a co-founder of silence. And I remember running into him at Black Hat once, like just before they launched that product. And I'm like, hey man, you know how you beat. He's like, I've been working on something. It's really cool. And silence, you know, look, in the end, they had an exit to, you know, wherever they wound up. And it wasn't really that spectacular. But the product was interesting. I think where they messed up is they missed the EDR train, right? But as an anti-malware engine doing machine learning classification, man, it worked well. Like, it worked really well. The
Starting point is 00:53:42 problem with all of that machine learning stuff was always going to be the edge cases. It was like, how do you handle like anti-cheat on that ships with a game or how do you handle some enterprise products that look like Trojans? But that was the thing is fundamentally, this technology was incredible. But then all of a sudden, everybody's like, it's got machine learning. Right? It's It sort of does feel like a repeat of that whole. It does. I feel like it kind of like ebbs and flows. And it's fascinating now when I'm when I'm in a lot of these,
Starting point is 00:54:11 these customer meetings and talking to folks, I don't get, I don't feel the same pushback that I used to in like 2016. What I get instead is there's usually some pressure from a higher up being like, look, if you get funding for this project, we need the latest and greatest. latest and greatest is AI. And it's like, okay, so there's some self-fulfilling prophecy there that kind of takes place.
Starting point is 00:54:38 So people are slapping AI on stuff so that like people can get authority to buy it because there's a mandate from heaven that says that they need to find efficiencies in AI. Yeah, yeah. And then I think on the flip side, like in 2016, people weren't using even accidentally using machine learning in their day to day. Whereas now it is so, it is just so. pervasive in everything that you do where I just kind of wonder there's just like a general malaise or a general comfort around like, okay, I'm already familiar with what a lot of this is.
Starting point is 00:55:12 So, you know, what does that mean for me and in the product that I'm trying to buy? Like, maybe I'm not pushing back as much as I should be. That's a really interesting point, which is that if people, if the people making the purchasing decisions or running these programs already familiar with chatbots, they've got a general familiarity with what they can and can't do, right? they've got a feeling for it, right? Yeah, and I find that, I find that to be the most interesting because it's like, yeah, but me, like us cybersecurity experts are often on the sidelines,
Starting point is 00:55:42 watching people use these tools external to cybersecurity being like, wait, whoa, wait, you know, don't put your medical records on here. Be careful what you hook your machine up to and allow it to do. And then on the same note, we're building tooling that's being like, yeah, you can probably take your hands off the wheel. it's fine. We'll remediate things. We'll catch things. And it's, it just seems to be, the message just doesn't seem to be, I don't know if it doesn't resonate or if we're just not thinking through the potential impacts. But I feel like there's an opportunity for the entire industry to kind of take a
Starting point is 00:56:17 step back and be like, what are we actually trying to sell here? And what does that look like? And that's kind of what I've been trying to communicate internally with this, this idea of like, SAE levels for autonomy, for lack of the better way of putting it. But I mean, just going back to what we were talking about earlier, which is like how do people gain trust in these systems? I mean, it sounds like what you're saying is the starting point, the starting level of trust is already kind of high because people are familiar with this, you know,
Starting point is 00:56:44 with basic chatbot technology. So there's already that starting level of familiarity. And then they're working through this stuff mostly with questions. What are the questions that people see most, concerned with. I mean, you mentioned red teaming as being a big, big concern. Like, what are some of the other ones where people are really like, you know, this is a, this is a deal breaker question for us when we're looking at evaluating an AI-enabled security technology? I think one of the the bigger ones I hear is just about data flow kind of through these agents, right? There's a lot of,
Starting point is 00:57:20 I think, misunderstanding about what tends to happen. And I feel like the, the main thudge around AI use in general is like, oh, these frontier providers are going to take all your data and train on it. It's funny how that became just an accepted truth when it is just not the case at all. That's not how this is. Right, right.
Starting point is 00:57:42 And that's, you know, I do elements of that happen probably on some level, but cybersecurity industry in general has so many policies and guidelines that we need to adhere to with regards to data. It's like we're not just vacuuming all this stuff up and then shipping it off to a frontier provider and being like, give me a response back, charge me money, and keep the data. It's yours.
Starting point is 00:58:11 That isn't really how it works. So, you know, part of it is education, right, with these customers and be like, look, this is the way this flows. This is what these models actually do. When we say we're learning from your feedback or from mistakes, like this is what we mean. It's, we're not going back as well. Oh, sorry. What's interesting here, what's interesting here is that, you know, you're talking about where there's pushback.
Starting point is 00:58:34 It seems to be, well, is it safe? They're like, is it red teams? Is our data contractually protected against being included in a training set? So it seems like people are not so much pushing back on, can this thing do what it says it does? They're more pushing back on, is it safe to use it? Do you think that's the dynamic here? I think that's the start, right? So that's, I can like pitch this more is those are the questions you get pre-POC or POB.
Starting point is 00:59:04 And then once it's in their environment, that's when you get more of the operational questions. Hey, what was this trained on? Do these things look like my environment? If they don't is an approximation. If it's not an approximation, how do you learn? Like at what point should I feel comfortable hitting the toggle button saying like I need to be in an active kind of feedback loop as opposed to a passive feedback loop. And that is a really interesting thing to navigate because it really can be a choose-your-own-adventure.
Starting point is 00:59:37 What level are you comfortable with? And you can chip away at that as a vendor by giving them explainable, kind of transparent reasoning along with any decision that it makes. Or you could just say, look, you could treat this as any other feature, it's just slightly, other machine learning feature, but it's just slightly more intelligent. And we've found personally that it's a, it's kind of a back and forth, a give and take, where you're showing them evidence, trust is built up. You make a mistake. Trust can degrade, but then how quickly do you turn that around? Or is the explanation around why that mistake occurred strong enough where that sort of
Starting point is 01:00:19 trust did not evaporate or go down? just said something fascinating there which made me yeah change my thinking I guess about how all this works right which is you said oh it's just like a machine learning thing but it's smarter and in many ways that's true right because LLMs are just machine learning but like at ridiculous kind of scale that was thought to be like impossible previously right you know you could do that but you'd need so much compute it's ridiculous it's like yeah here we are like hundreds of billions of compute later um yes that's what we got no one expects machine learning solutions which are just bought and sold like without any question. No one expects them to be
Starting point is 01:00:56 perfect and never to make mistakes. But it seems like when it comes to a lot of these like contemporary AI solutions, that expectation is very different. I'd never thought of that before. But like if your machine learning based like IDS or mail filtering thing makes a mistake, like no one's even complaining at that point. Like they might grumble about it a bit if the mistake's really bad. But like why is it that there's such a higher expectation that these AI solutions are going to have to be, you know, perfect. Like, people will look to, like, point to them making a mistake and say, see, this technology's rubbish. They don't do that without a tech. Like, what, why is that? Yeah. I, I chalk it up to, you know, the hype, like, the, the marketing
Starting point is 01:01:39 hype around AI in general is, is it like a, it's such a level that I feel like it's, it's very hard to walk back. And, you know, I recall the, the days where it was like, our machine learning catches 99.999%. It's like, that's probably not true. But I think now there's just this expectation that even when you make a mistake, these things are so smart that it's just going to pick it up the next time around. It's like when you're talking to a frontier model via chat, and you're like, no, no, no, that's a mistake. And it's like, it takes on the persona of a human being and you're like, oh yeah, that's actually, that's a sharp question or that's a good point. And I think human beings take that feedback and they're like, oh, it's learning.
Starting point is 01:02:24 So now I shouldn't see that mistake ever again. And I think where you run into problems, particularly with being a security vendor, is you're pulling in these frontier models. You're not actively adapting them, right? Like nobody's sitting and burning through $30 million, right? Yeah. Yeah. Yeah. Like you're saying, oh, people, you know, are they learning?
Starting point is 01:02:45 But they don't. You know, and even if you put, you know, like even if you prime them with the right instructions and prompts and whatever, they still ignore you every now and then. We saw this Twitter thread recently where someone lost their entire production environment because they thought they're like text-based, you know, instructions to a model of never ever do this were guardrails and they're not. Yeah. Yeah. And that's, I love that consumers are getting a little bit more savvy and they're learning more the nomenclature and kind of what to ask. So it is cool to get things like, you know, what guardrails do you have around that? And it's like, well, you know, we,
Starting point is 01:03:20 here's kind of what we're doing and this is what we give it access to. And sometimes it satisfies things, but other times they pull the thread a little more. And they're like, all right, well, talk to me about tool use. Like, what tools do they have? And they're coming at it. And it's getting, I want to say, maybe more precise the way they're thinking about it. And they're starting to pull the right threads. So as you go through prop, Pock, they start asking you, well, how, you know, why do I need the AI? What does the AI actually do? Yeah, yeah, exactly. And it's just like, and then they get that taste, right? And they're like, oh, wow, this like takes care of this problem that I have or I'm throwing too many people at.
Starting point is 01:03:59 You're like, great. And then it's usually at that point, they're like, could we put it over here? And I'm like, it took us so long to get to this point. Like, let's take a breath. Let's learn. And then we can start to move it over. And it's, yeah, the parallels, I feel like, with self-driving cars and kind of what we went through in the late 20th. 2010s is like, certainly not lost on me.
Starting point is 01:04:23 Where it's just like, yeah, kind of funny, kind of funny. Bobby, we're going to have to wrap it up there. We are out of time. Great to chat to you about all of this. And yeah, for those interested, they can check out Sublime Security. A great email security product. Thanks again. No, thank you.
Starting point is 01:04:38 Take care. That was Sublime Securities. Bobby Fila there. Big thanks to him for that. Big thanks to Sublime for being a risky business sponsor. And that is it for this week's show. I do hope you enjoyed it. I'll be back soon with more security news and analysis.
Starting point is 01:04:54 But until then, I've been Patrick Gray. Thanks for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.