Your Undivided Attention - Decoding Our DNA: How AI Supercharges Medical Breakthroughs and Biological Threats with Kevin Esvelt

Episode Date: July 18, 2024

AI has been a powerful accelerant for biological research, rapidly opening up new frontiers in medicine and public health. But that progress can also make it easier for bad actors to manufacture new b...iological threats. In this episode, Tristan and Daniel sit down with biologist Kevin Esvelt to discuss why AI has been such a boon for biologists and how we can safeguard society against the threats that AIxBio poses.Your Undivided Attention is produced by the Center for Humane Technology. Follow us on Twitter: @HumaneTech_RECOMMENDED MEDIASculpting Evolution: Information on Esvelt’s lab at MIT.SecureDNA: Esvelt’s free platform to provide safeguards for DNA synthesis.The Framework for Nucleic Acid Synthesis Screening: The Biden admin’s suggested guidelines for DNA synthesis regulation.Senate Hearing on Regulating AI Technology: C-SPAN footage of Dario Amodei’s testimony to Congress.The AlphaFold Protein Structure DatabaseRECOMMENDED YUA EPISODESU.S. Senators Grilled Social Media CEOs. Will Anything Change?Big Food, Big Tech and Big AI with Michael MossThe AI DilemmaClarification: President Biden’s executive order only applies to labs that receive funding from the federal government, not state governments.

Transcript
Discussion (0)
Starting point is 00:00:00 Hey, it's Tristan. And this is Daniel. When you strip down all life on Earth to its most essential parts, it's a language. A language with four characters, A, G, C, and T. These are the four base pairs that make up every strand of DNA. And all of biology is fundamentally controlled by this language. Even though it wasn't until 2022 that we sequenced the complete human genome, We're still in the relative infancy of understanding how this language works.
Starting point is 00:00:34 And interacting with it is like pouring over ancient hieroglyphics. But now we have a Rosetta Stone, a key that can translate those hieroglyphics for us. And that is AI. We've seen so much in the news about how AI has been able to ingest the language of human interaction and create chatbots, or ingest the language of coding and create GitHub co-pilot. But what gets less attention is that AI is also able to understand the language of our DNA. and understanding our DNA means that we can go from creating powerful language machines to powerful biology machines. Research has been able to use AI to model thousands of molecules and proteins in seconds to develop promising new drug candidates.
Starting point is 00:01:11 But these advancements are a double-edged sword. The same technology that enables rapid drug development can also allow bad actors to rewrite the language of biology for malicious purposes. AI is rapidly opening up new frontiers in biology that sound like science fiction, but they're not. And we need to take these advancements seriously. That's why I'm so excited to have Kevin Esfeldt on this episode. Kevin's dedicated his career to this topic. And he's one of the world's top experts in synthetic biology. He's a professor at MIT and director of the sculpting evolution group.
Starting point is 00:01:43 Recently, Kevin has devoted a lot of his time and attention to the threat of AI-enabled biological threats. And he's come up with some really pragmatic and achievable solutions on how we can minimize those threats. Kevin, welcome to your undivided detention. It's lovely to be here. So let's dive right in. What is it about AI that makes it such a powerful tool for biologists? That's a great question. It's really hard for us to learn to speak the language of life, because we just don't natively read A's and C's and G's and T's, let alone the 20 amino acids. That's not how our brains evolved. But if we can train an AI to do that, then it gives us access to a capability we wouldn't otherwise have. So we can train an AI, a biodeign tool, with a lot of data that we've gathered, either on different proteins and how they fold, or different variants of a particular gene
Starting point is 00:02:38 and the corresponding activities of the protein that that gene encodes. Now, proteins are chains of amino acids, typically hundreds of amino acids long, and there's 20 possibilities at every position, which means that a short protein of just 100 amino acids, there are more possible proteins of that size than there are particles in the universe. So we can't possibly learn on our own how to predict the activities of proteins.
Starting point is 00:03:09 But with AI, we can now predict how proteins fold. And if we get enough data, we might eventually be able to design new molecular tools with useful functions in order to fight disease. One of the things that you did relatively early on is talk not only about all the incredible benefits that this could give us, but also the new risks that these tools end up creating, specifically in the way that the DNA synthesis machines could be used.
Starting point is 00:03:35 Can you talk a little bit about that duality? Like there's all these benefits, but then there's all these risks as well, right? So I'm cursed to always think about what might go wrong. And I'm fairly concerned that as we learn to understand in program biology, we might learn to make pandemic viruses before we can adequately defend against them. And that suggests that we might not want to make it extremely easy for everyone to acquire the DNA that allows them to synthesize an infectious virus, because that's one of the things that we can do now. So you can go online and look at the sequence of pretty much any virus that we've ever discovered.
Starting point is 00:04:18 And virologists have come up with ways of turning that se, sequence into an infectious virus. It's a critical tool in understanding how viruses and disease work and coming up with countermeasures. And it depends on the exact kind of virus, the details of how it works. But what it amounts to is you need to turn those A's and Cs and Gs and T's on your screen into physical DNA sequence. So to do that, to turn bits into atoms, you place an order with a DNA synthesis provider, which is just a company whose business it is to turn that string of symbols into a string of molecules. So you order that from them, sometimes with other DNA, with instructions for to help make the virus. That depends on the details of the virus.
Starting point is 00:05:08 And they send you this DNA sequence in the mail. And I think a lot of our listeners might not even know that these are fully automated at this point. They're machines. I don't know how much they cost. Are we talking, you know, desktop machines? Or are we talking a mobile? you know, $5 million things in the back of a lab? Like, what are these machines? Well, that's a great question. So the DNA synthesis machines, there's a whole bunch of them, different varieties. Some of them specialize in making lots and lots of DNA at once, the industrial scale.
Starting point is 00:05:34 And those are a few million dollars, as you said. But you can also pay several tens of thousands of dollars for one that will sit on your desk, and it will let you make DNA. So a desktop DNA synthesizer can produce enough strings of DNA to make. an influenza virus in about a week. Now, it's just going to make the DNA in pieces. It's not going to stitch them together for you. There's people making machines now that are the assemblers that will do the stitching together automatically. Right now, there's usually some humans involved, to some extent, but that's gradually vanishing.
Starting point is 00:06:14 In your research, I think you say 30,000 people are already able right now to assemble a flu virus, 5 to 1,000 people could make coronaviruses, and 100 to 500 could make a pox virus right now. That's way more people than can create nuclear weapons in the world. So this already just sounds incredibly dangerous. Like, why hasn't this just been locked down to biology labs that have access and you have to be some kind of licensed biology researcher? Why are we even in this state as is? Like, it sort of seems crazy to me.
Starting point is 00:06:45 That is a phenomenal question to which, despite being... being in this field for multiple years, I just have not yet worked out an answer. Given the seriousness with which we take nuclear weapons, it doesn't make sense. Given the seriousness with which we take infectious versions of these pathogens and the controls that are necessary to work with them, to store them, to get anywhere near them, it doesn't make any sense. But if you think about the select agent program in the United States, which is the program that regulates control of all these kinds of pathogens,
Starting point is 00:07:14 like smallpox, like the 1918 pandemic influenza virus, In order to work with them, you have to have security background checks. You have to show that your lab has adequate security precautions, and it's all under lock and key at all times when you're not actively working on it. And they're far-sighted enough that you cannot order the DNA that could directly be used to make the virus. But you can order it in two pieces that any undergraduate student in biology could stitch together on their own and then have the illegal thing.
Starting point is 00:07:47 So you're saying, if I right now say, hey, I want smallpox, I want to order it, the mail order synthesis providers, they will not send me back smallpox if I order smallpox. But if I say, hey, I want half of this, half of smallpox, but I don't call it smallpox, and this other half of smallpox and I don't call it smallpox, they will happily send me both those things. I can't say that for sure with smallpox, but we did run a test recently with 1918 influenza. Now, it's not a perfect test because we did place the order from a biosecurity org that doesn't do any wet lab research. So it has no reason to order DNA whatsoever. So we wanted to check how many of these companies would detect our orders and send us pieces of 1918 influenza virus. And it turns out
Starting point is 00:08:29 that most of them did. That's really surprising to me, that you can just split this up into two pieces. It feels like a pretty basic failure of security. Why does it exist like that? Why do you think that the security is so bad? Well, there's two reasons. One, we don't always secure the right pathogens. So, for example, there are some viruses that really should be on the select agent list, but that list was made with biological weapons for state actors in mind, like anthrax, the stuff you can spray over a city and kill a bunch of people that way, not really geared towards pandemic viruses. So that's problem number one. We just didn't come up with the list right. But then the main one is they came up with the list when you couldn't make a virus from synthetic DNA.
Starting point is 00:09:11 And so they've had to update the regulations over time to deal with advances in technology. And so it was honestly pretty impressive that they got as far as saying, yeah, you're not allowed to have the virus and you're not allowed to have the DNA encoding the virus either. But they didn't quite go as far as you're not allowed to have it in pieces. And because of that, the synthesis providers are entirely within their rights. It is 100% legal for them to ship you, that DNA. If we required everyone to screen their orders to make sure you're not ordering something like 1918 pandemic influenza virus, and check to make sure that people aren't ordering different pieces from different providers,
Starting point is 00:09:55 and ensure that the only people who can get that DNA have permission to work with the actual virus, then we would be a lot safer than we are now. And it wouldn't be very costly because the tools to detect these things exist. But here's where the AI comes in. Not now, but eventually we will be able to use AI tools to make variations on these existing viruses that won't have the same sequence, but will have the same function. So that means our detection tools need to start detecting AI-designed biological weapons. And in fact, until the recent executive order on AI, there were no
Starting point is 00:10:41 legal restrictions on DNA synthesis at all. And just for listeners, President Biden's executive order on AI from last October did include restrictions on DNA synthesis for all labs that get government funding. But for all other labs, it just suggested a framework, nothing legally binding. But Kevin, there's a really basic and essential question we haven't actually asked yet. Why do researchers want to use AI tools to synthesize viruses in the first place? Yeah, it's true that a number of biodesign tools are actually optimized for discovering new variants of viruses that are likely to cause pandemics. And you might say, what?
Starting point is 00:11:24 Why would researchers ever do that? But there's an excellent reason why they want to do that. They want to design vaccines that will not just hit the current variant of the virus. They want to close off the paths to all the ones in the future, which means they need to predict how that virus is like. to evolve, and this helps them design the vaccine to prevent that. But the flip side is, you just made a tool that tells you how to change those sequences, and not just one at a time, but together in a way that will escape all current immunity while still getting into ourselves.
Starting point is 00:12:00 In other words, once they're successful at this, and they're still getting there, right? They're not good enough yet. But if they're successful, then there will be this tool out there that will tell you, how to make the next variant of COVID or influenza, such that your version would spread and infect people throughout the world. So this is where access to those AI tools becomes an acute problem. You're bringing up a few different things here that I want to lay out as separate, distinct points of interest.
Starting point is 00:12:29 So one is AI is omnibus. The same AI that can be used to optimize a supply chain can tell you where to break supply chains. The same AI that can tell you how to synthesize, you know, the vaccine can also tell you how to synthesize a million variations of the virus. And we always say we want to maximize the promise and minimize the peril. And I think in an uncomfortable truth about AI is the promise is just the flip side of the peril.
Starting point is 00:12:52 And I'm thinking of the chemistry lab that was using AI, and they said, I wonder if we can use AI to find less toxic drug compounds. The AI came back and it had found all these more healthy, safer versions of these drug compounds. But then they had the thought, well, I wonder if I could flip that around and say, instead of finding less toxic compounds,
Starting point is 00:13:10 I'm just literally going to flip one variable, say, more toxic. And they walked away from the computer, they came back six hours later, and it had discovered 40,000 new toxic molecules, including VX nerve gas. How do we deal with that? How do you reckon with that equation of the omnuseness of AI? Well, I suppose if I had an answer to that, this would be the place to disclose. I guess I'm optimistic about AI because in a way I'm pessimistic about humanity. It does not seem to me possible to build a civilization in which,
Starting point is 00:13:40 which every human alive could release a new pandemic whenever they want, because some people would. Whereas it seems quite possible to me that we could figure out how to ensure that every AI would never help them do that. Could you then talk about the challenge of open source AI? Because while the closed models from Anthropic, when you say, I'd like to build this biological weapon, please give me instructions how, and just to be clear for listeners, it'll say, know I'm not going to do that. But the problem of open source AI is if meta trains their next Lama 3 model on a bunch of open biology data and biology textbooks and papers on a bunch of things in this area. And then they open up that model, whatever safety controls can be just stripped
Starting point is 00:14:23 off because it's open source code. How do you think about the intersection of open source AI models and this area of risk? Right now, it's not much of an issue in the sense that the models are not powerful enough to do a lot of harm. The thing about the LLMs, which are moving towards being general-purpose research assistance, and perhaps eventually researchers themselves, is that they'll be able to help people gain the skills of a human researcher. And for all the reasons we've discussed, human researchers today can make pandemic viruses. So the principle of defense-first technology development just said,
Starting point is 00:15:05 don't learn to make pandemics until you can defend. Right. And defense is hard when the AI can keep making new variants that get around whatever vaccines you made. And it's a struggle to produce vaccines enough to protect everybody. When someone malicious could just make a new pandemic, that suggests you need to move outside of biology for the defense. You need to prevent people from being infected in the first place.
Starting point is 00:15:33 So all we need to do is invent protective equipment that will reliably keep us from getting infected with any pandemic virus. But we don't want to wear protective equipment moonsuits all the time. I was going to say a vision of the world where everyone is wearing a moon suit doesn't feel like a good stable state to land in. No, definitely not. But imagine the lights in the room you're in don't just provide light that you can see to illuminate the area. They're also germicidal. They're emitting what we call far UVC light, and far UV light is effective at killing viruses and bacteria, but it's harmless to us. So we definitely would want this in public spaces, coffee houses, workplaces, stadiums, assembly halls, you name it.
Starting point is 00:16:23 Maybe we want it in our homes, maybe we don't. But this is the kind of intervention, along with just better ventilation in our homes in general, that can just protect us from getting infected with anything without having to wear clumsy protective equipment. There's lots of ways you might do this. But the goal is to just ensure that we never encounter biological information in a form that could get inside our bodies and hurt us.
Starting point is 00:16:52 And this looks like a very solvable problem. You know, there's this sort of vision of, like, like, okay, well, we have AI that can, Times Biology, that lets us uncover all these amazing positive things, but in which basically, like, a single person on any given day can hit a button and then, you know, really horrible stuff is suddenly permeating the world. And that just feels like I would prefer to just not live in that world. Like, I don't know if I would take the trade of all those benefits if the cost side of that equation is a world in which that's true. even if those benefits were just literally the most seductive ever,
Starting point is 00:17:32 it just feels kind of like a deal with the devil where the other side isn't worth it. And I'm just curious your reaction to that. Well, given that I do run a biotechnology lab, I clearly think that biotechnology is a net positive for humanity, even though I acknowledge that once we learn enough about how to program it, once we understand how it works, we, by definition, will be able to be able to be able to be,
Starting point is 00:17:58 to create new pandemics. I'm not saying we're anywhere close to that today, but I just don't see any way out of that dilemma. You acquire enough power, then you can rearrange the world, and you can use it for good or you can use it for ill. And I'm not sure that humanity as we are today is collectively responsible enough
Starting point is 00:18:24 to deal with that level of power on an individual basis, which suggests you just don't let individuals have that. Don't let everyone order synthetic DNA. You know, you can reduce the risk tremendously, but at the end of the day, the capability will be there if people can turn that information into atoms.
Starting point is 00:18:49 And it looks to me like you can remove pretty much all the pandemic risk and keep all of the benefits. We just have to get serious, about doing it? Let's talk about that, because on one hand, I'm actually imagining there's just several big fixes. In the same way that we locked down photocopiers
Starting point is 00:19:07 to make sure they're not printing $100 bills, we might be able to do very similar things and lock down the DNA synthesis providers, lock down a whole bunch of stuff and end up in a world that's not perfectly safe, but more or less as safe as we are now, or more or less as safe as we were before genetic synthesis technology rolled out. Help me calibrate the level of risk and the level, furthermore, the level of intervention that you think is necessary.
Starting point is 00:19:33 What you need to do is, yes, screen all the DNA that gets sent out as an order to a mail-in synthesis provider. You need to make sure that they're not going to just send it anyway, even though it showed up as dangerous. That means you need to actually check them, make sure that they're licensed, make sure that folks try to get harmful DNA from them without authorization, and make sure that that does. doesn't happen. Is it something that takes a few million dollars in a year? Or is this something that is a massive society-wide project? How complex is this? The former.
Starting point is 00:20:07 There's less than 100 DNA synthesis companies in the world, including those who make the DNA synthesizers themselves. There are more that sell short fragments of DNA, which eventually will need to screen, but still you're talking a few hundred companies in the world. And there's multiple free screening systems that do a pretty darn good job of of detecting hazards. And there's even ones that will turn the permissions that you have as a biologist. So my lab has permission to work with some, actually quite a number of different biological organisms.
Starting point is 00:20:41 And there's software that can turn that list of organisms we can work with into what's called a public key infrastructure certificate. So if you send that into a synthesis provider with your order, you could order any DNA from anything that you have permission to work with, and it would just go through seamlessly. So we can basically automate away this whole problem by encoding the screening system into the DNA synthesis machines themselves. This would be really basic. Can you imagine a system where you're selling what could be the equivalent of enriched fissile
Starting point is 00:21:17 materials for nuclear weapons? And you don't even check to see whether or not the person who ordered it has authorization. And so are there any barriers preventing? these kind of fixes today? I mean, what's preventing us from being in the world where we just have good biosecurity? Policymakers actually making it happen. One of the amazing things about the synthesis providers is that they recognize that this was a problem, and they got together, and the biggest companies formed what's called the International Gene Synthesis Consortium, and they said, hey, we're all going to screen to make sure that we don't
Starting point is 00:21:51 send nasty things. And they do their best, and they do this at cost to themselves. Because when someone orders a hazard, they then need to decide, should we send it or should we not send it. Which takes time, and time is money. But they do this because it's the right thing to do, which is a rare example of people acting against their incentives to do the right thing. But the problem is there's only so far they can go. These technologies already exist. We just need people to use them. In order to get people to use them, governments need to require them.
Starting point is 00:22:25 because the IGSC does their best, but if they're too much of a pain to customers who, for good scientific reasons, order hazards, those customers are going to place the order from a company that doesn't give them as much of a hassle. So it's just a problem for governments, and that means we need action from multiple governments, particularly the U.S., China, and EU,
Starting point is 00:22:50 which are far and away the largest markets, virtually all DNA synthesis providers and manufacturers. are in one of those countries or 90% of their business is to one of those three markets. If they just all get together and require this, which again costs almost nothing, then we have mostly solved this problem. Well, I mean, that's so exciting.
Starting point is 00:23:12 I have to admit that usually at CHT we end up depressing people. We bring them into a problem and we say, no, no, it's so much more complicated than you imagine. There's so many more bad incentives than you imagine. But I have to admit, I'm actually kind of cheered up by this call because in the end of the day, actually, this could be solved for the low millions of dollars. It's just we need to act now. We need to change the incentives, change the regulation, and get the protections in place.
Starting point is 00:23:33 But from what you're saying, this is actually a solvable problem, and this is a problem we should actually just lock down. Yeah, we could make DNA as secure as enriched fissile materials that we lock down because otherwise they could be used to build nuclear weapons. This is totally achievable. Someone in your position, I imagine one of the difficult things is some of the risks that you're looking at you both want public awareness
Starting point is 00:23:56 for because you want to patch these holes. You want to like spend the tens of millions of dollars, recall all the synthesis machines and install these protections. And at the same time, you kind of don't want to shout from the rooftops how vulnerable we are. Can you talk a little bit about that tension and what you see the sort of information risks
Starting point is 00:24:12 are? You've just clearly articulated the basic horror and burden of my life. Which is if you think you see a way, to cause catastrophic harm, how do you prevent it without articulating it and making it more likely?
Starting point is 00:24:28 Which is why I'm very careful with my words. I'm not saying that we're at risk of a pandemic right now that would be deliberate because there's detailed instructions. What I will tell you is that I'm concerned about 1918 influenza because there
Starting point is 00:24:43 is probably a, I don't know, maybe a 10% chance if you ask a bunch of virologists, that it might actually cause a pandemic today. and it would probably be COVID level. It wouldn't be nearly as bad as in 1918, because a lot of us have immunity. To me, a 10% chance is enough
Starting point is 00:24:59 to get people concerned about it, but if you're a bioterrorist, are you really going to go out of all this trouble for a 10% chance? Seems pretty unlikely. Now, I could be wrong about that calculation. Maybe we'll roll the dice and 10-sided die and comes up a 1.
Starting point is 00:25:15 That's always possible. But if we don't lay out some concrete scenario of something that could go wrong, I've found nothing happens. And you can't just talk to policymakers behind closed doors either. You have to lay out something there
Starting point is 00:25:34 for people to talk about and think about and understand where we're going in order to build momentum to get something to happen. I find it incredibly striking that the only effective governmental action we have on getting DNA synthesis screening in place came with the executive order on AI
Starting point is 00:25:58 because people were concerned enough about AI eventually creating biarrisks that we put that in there to close off this bits to Adams transition, as we should have. Do you feel confident in the executive order, which is not fully legally binding in the way that it will shift
Starting point is 00:26:17 towards more of these solutions? and policy, or do we need a lot more action from Congress? We need action from Congress. The executive order only covers institutions that take federal money. So if you take federal money or your institution does, then you're supposed to buy DNA from a company that will attest that they screen their orders. And there will be audits, red teams that check to see.
Starting point is 00:26:42 I can't guarantee that they're going to be robust by my standards, but they're supposed to be there. But still, is there going to be any kind of, is there going to be any kind of stick forcing companies that say they screen to actually up the standard? I don't know. It's very unclear. But if you have a startup or you're not accepting federal funds for whatever reason, it doesn't apply to you. You can buy DNA from whoever you want. And what's more, it doesn't apply to the providers at all. So we need action from Congress to close that loophole. And while they're at it, they could
Starting point is 00:27:18 fix this select agent program loophole as well, where it's fine if you buy the DNA in two pieces. As soon as you put it together, it becomes illegal. Just fix that one and improve the executive order a little bit. Get that in law, and we'll be off for the races. And I can sleep better at night. So, Kevin, about a year ago, the CEO of Anthropic Dario Amadai testified before Congress about AI risk and he talked about this exact issue. A straightforward extrapolation of today's systems to those we expect to see in two to three years
Starting point is 00:27:54 suggests a substantial risk that AI systems will be able to fill in all the missing pieces enabling many more actors to carry out large-scale biological attacks. We believe this represents a grave threat to U.S. national security. So a lot of people hear this and they think that Anthropic is just trying to scare people
Starting point is 00:28:13 and hype up these dangers that, US government will regulate AI, stop open source, and then only anthropic open AI and Google will be able to sort of be stewards of this AI future. What do you think of that? Is this a regulatory ploy? Do you agree with Dario's timeline that he gives here? I definitely do not think that it's a regulatory ploy. I believe it is sincere. My concern is that we will fail to take action because we believe it is unlikely that these capabilities will exist within three years. I don't like giving timelines
Starting point is 00:28:49 because it's so hard to anticipate capabilities, and I do feel that, at least in biodesign tools, we are very data limited right now for most things. But even conceptually, it seems plausible. That means we do need to evaluate capabilities on newly trained models before they're released. This can mostly be done by working with the industry leaders. We do need some kind of policy framework with teeth to ensure that those systems are not rolled out and made available in a form that can still be jailbroken.
Starting point is 00:29:24 If we truly do believe that they can describe ways of causing harm much worse than COVID, much worse than 1918 influenza, as bad as smallpox or even worse. It seems like there's this tension where as we do science, we start off in an open way, we publish all the papers, we publish our discoveries, you know, we're doing quantum physics in 1930s, and we publish everything we're blowing. And then at some point, some people realize, oh, shoot, there's some really dangerous things you could do. And we stop publishing. And that's happened a little bit in AI. And I'm just curious to ask you about what should happen with biology going from this open approach to do we need to go to a slightly more closed model. I'm actually going to push back on your framing, because I think we're not nearly as open in science as we could be. indeed, as we should be, right? We don't disclose our best ideas until we've largely realized them ourselves and can tell the story ourselves in order to get all the credit.
Starting point is 00:30:18 I'd love to change the incentives to disclose our best ideas early so that we can better coordinate how to realize them. I very strongly defaults to total transparency. It's just that I also recognize that there is at some point a capability that if misused
Starting point is 00:30:36 could be destructive enough that it's not worth the risk anymore. And I think it flips about the time when you can cause a COVID-level pandemic. And I'm tremendously sympathetic to the position of the folks that say, look, this just seems like a power play. We're really worried about the centralization of power. Well,
Starting point is 00:30:55 I am too, but I'm also worried about too much power in the hands of individuals who might be mentally unstable or captured by a horrific ideology. So the way that we deal with those people is we structure society so they can't do very much harm. And we impose checks and balances on government so that no central power source
Starting point is 00:31:17 can do that much harm. Kevin, I think it's been an incredibly sobering conversation about some real risks. I just want to thank you for all of the incredible work that you do, and there's a number of solutions that you've outlined from far UVC light to metagenomic testing and secure DNA screening. There's a number of things that policymakers, if they took this conversation seriously, could do. And I just want people to focus on that and get the work done.
Starting point is 00:31:45 And thank you so much for coming on your undivided detention. Thank you so much for everything that you're doing to mitigate not just this risk, but all of the different ways that the shining source of our greatest hopes could be turned against us. I'm deeply grateful. Thank you. Your undivided attention is produced by the Center for Humane Technology, a non-profit working to catalyze a humane future. Our senior producer is Julia Scott. Josh Lash is our researcher and producer,
Starting point is 00:32:17 and our executive producer is Sasha Fegan. Mixing on this episode by Jeff Sudakin, original music by Ryan and Hayes Holiday. And a special thanks to the whole Center for Humane Technology team for making this podcast possible. find show notes, transcripts, and much more at humanetech.com. And if you like the podcast, we'd be grateful if you could rate it on Apple Podcast because it helps other people find the show. And if you made it all the way here, let me give one more thank you to you for giving us
Starting point is 00:32:44 your undivided attention.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.