Cautionary Tales with Tim Harford - "I get in trouble when I say things like this" - Michael Lewis on Sam Bankman-Fried

Episode Date: July 11, 2025

Acclaimed author Michael Lewis discusses his time with Sam Bankman-Fried and why he thinks both high finance and Effective Altruism shaped the 'Crypto King's' worldview, ultimately landing him in jail.... Plus, we hear about the people fighting terrorism, cave-ins and brain-eating amoeba from Michael's new book 'Who Is Government?'. For a full list of sources, see the show notes at timharford.com. Get ad-free episodes, plus an exclusive monthly bonus episode, to Cautionary Tales by subscribing to Pushkin+ on Apple Podcasts or Pushkin.fm. Pushkin+ subscribers can access ad-free episodes, full audiobooks exclusive binges, and bonus content for all Pushkin shows. Subscribe on Apple: apple.co/pushkinSubscribe on Pushkin: pushkin.fm/plusSee omnystudio.com/listener for privacy information.

Transcript
Discussion (0)
Starting point is 00:00:00 This is an iHeart podcast. On the 10th November 2022, Sam Bankman Fried tweeted that he had F'd up. As we explored in our last episode, that was something of an understatement. FTX, the digital currency exchange platform where individuals could trade cryptocurrencies and store their funds for safekeeping, imploded when it emerged that Sam Bankman-Fried had been diverting funds into his crypto hedge fund, Malameda Research. Fraud, yes, but as an effective altruist, he was doing it for a good cause, wasn't he? In his own misguided way, he thought he was just following the teachings of Wilma Caskell, the founder of the effective altruism movement.
Starting point is 00:01:15 That can't be bad, can it? If you've not listened to that episode yet, please do before listening to this special cautionary conversation with Michael Lewis, the man who had a ringside seat for the spectacular fall of the cryptocurrency wunderkind. Michael, welcome to cautionary tales. Tim, good to see you again. It's great to see you. I think the last time we saw each other, I was trying to teach you how to play an obscure
Starting point is 00:01:42 German board game. No, that's exactly right. I can't even remember what book led you to get interested. Maybe the Big Short. I think it was the Big Short. But the FT were like, oh no, you can't just talk to Michael Lewis about his book. You have to talk to Michael Lewis about his book while you teach him a weird German board game. I thought it was actually ingenious because I try to do this with subjects.
Starting point is 00:02:02 And especially if you only have a short time with them. You're so much better off doing something with someone than just talking to them Yeah, if you're trying to kind of get some insight into how they tick the best job interview I ever had it was to lead teenage girls through Europe when I was 22 years old I went to the tour agency and the guy who ran the tour agency said God I forgot you had an interview this day We're supposed to move the furniture from this office down the hall to the other office And the guy who ran the tour agency said, God, I forgot you had an interview this day. We're supposed to move the furniture from this office down the hall to the other office. Could you just help me do that?
Starting point is 00:02:29 And then if we have time left over, we'll talk. And I spent an hour moving furniture with this guy. And I left, he said, I'll call you about the interview later. And I left bewildered. And like a week later, he said, you have the job. And two months later, I'm in a hotel room with my fellow leader in like, Bruges. And I said, you know, this was odd. I was never interviewed.
Starting point is 00:02:47 And I explained what happened. He said, I moved that furniture too, back the other way. And it was a really smart way to kind of figure out whether someone was collaborative, whether they could pick up stuff quickly, how they interacted with people. So I took it as a sign you were a clever journalist when you did this. Yeah, well, or I have a clever editor. But yeah, Michael, I'm already feeling bad that I haven't brought a board game this time, but I'm sure we'll be fine.
Starting point is 00:03:10 I'm sure we'll be fine. We are going to talk about the time you spent with Sam Bankman Fried, what you made of him, what you made of the influence that effective altruism had on him. And we're also going to be answering listener questions. And we'll be talking about your new book, Who is Government? Before all of that, here is the Cautionary Tales theme. The most remarkable thing about Going Infinite is that when you started working on it, you had no idea that you were going to be covering the fraud of the century. You were just intrigued by Sam Bankman Fried and this book turned into a very different book during the course of writing it. And it feels like this is not the first time you've been in the right place at the right time. I'm curious how is it that you find your subjects?
Starting point is 00:04:21 You know, I stumbled into this. I had no idea that there was like fraud going on. So I was asked to evaluate him by an investor who was doing a deal with him. And it was a friend. And I said, sure. I had no idea who he was. He turns up in my doorstep in Berkeley and we spent a couple of hours together. In the couple of hours I realized I had this character who was, one, he thought about the world in a very unusual way,, and a very persuasive way, but two he would possibly give me access to several places
Starting point is 00:04:50 I wanted access one was Jane Street in the world of high-frequency trading because he'd spent three years at Jane Street And he'd hired a bunch of people out of there and they're notoriously opaque and won't talk to journalists I thought ah, maybe I get to Jane Street throw him yeah Money in politics because he was handing out money left, right and center to American politicians. And the crypto world, which he was very skeptical of, and so he wasn't a religionist, so he wasn't defensive about it. So I remember just saying to him, I don't know where this is going to end up, can I
Starting point is 00:05:17 just hang out and watch? He was the world's richest person under the age of 30, according to Forbes Magazine. It wasn't a big brainwave on my part. It was kind of surprising he let me do it, especially in view of what was going to happen. So I just follow my nose. And my nose isn't like, oh, what's going to be the next big story?
Starting point is 00:05:35 My nose is more, what's not dull? And sometimes that works out. Sometimes that ends up being what you should be paying attention to. Yeah, so as you hung out with Sam Bankman Fried more You must have become aware of this whole effective altruism Thing seems to play a huge role in his life and how he thought you're exactly right. It kind of blew my mind I'd never heard of this
Starting point is 00:06:00 The idea that that he had gone to Wall Street with absolutely no interest in money, his parents were both law professors at Stanford. They're like non-materialist people. They don't care about stuff. They don't care about money. You know, in another generation, he'd have been like a high school physics teacher or maybe a college professor. He was not the social type who goes to Wall Street. And he'd forced himself into it because he had discovered, before he discovered Wall Street, this movement called effective altruism that just resonated with him. Yeah. It was seduced into it by the philosophers.
Starting point is 00:06:35 And they proselytized. And they actually, Wilma Gaskell told me, he said, you know, it turned out that the ideal market for these ideas were nerdy math science people at American universities, people who were attracted to this kind of quantitative solution of how to lead your life, maximize the number of lives you save. I mean, just as an aside, I've been writing a column about the Trump administration's decision to cut foreign aid. And it's slightly controversial exactly how much foreign aid they've cut and exactly what
Starting point is 00:07:03 they're cutting and they're conflicting statements, but plausibly Sensible independent analysts reckon that decision is going to kill a million people a year Yep, and it's mostly people with HIV or there were there a few other things. There were some vaccinations There's tuberculosis, but you have HIV You have very effective medication as long as you take the medication You're fine the moment you stop taking the medication the virus starts to come back and pretty soon you're gonna die mmm, and I'm writing about this and somehow I can't quite get as
Starting point is 00:07:35 outraged about those million people as if they were being rounded up and Executed for being illegal immigrants or if they'd started some war that was likely to kill a million people. It just feels different somehow. Dead is dead, right? Part of the reason it feels different is it's a number, right? Especially the bigger the number, it just becomes a number. You would feel much more outraged if you saw just a single child die. We don't respond to data, we respond to anecdote. Yeah. Unless you're Sam Beckman Fried.
Starting point is 00:08:04 Unless you're Sam Beckman Fried. You do respond to data, we respond to anecdote. Yeah, unless you're Sam Beckman Fried. Unless you're Sam Beckman Fried. You do respond to the data. You know, there's a jumble of my previous books that are sort of popping into my head when he starts talking this way. One is Moneyball. Oakland A's succeeded because they ignored the anecdote and looked to the data. And two was The Undoing Project.
Starting point is 00:08:22 Yeah. Kahneman and Tversky were all about the way our minds mislead us when we trust our intuitive sense of things rather than the data. It was interesting that this was wholly persuasive to this kind of person. And not just Sam Beckman Friedman, but there was a whole crowd. I mean, it was cult-like.
Starting point is 00:08:41 As McCaskill said, they tend to be on kind of on the spectrum. They tended to be male. They tend to be math or physics. They tend to be socially awkward. They were people for whom emotion was a weak guide to action. Whatever you or I feel that leads us to do things, they didn't feel so much. But they could talk themselves into using logic and numbers as a guide to action in the way that emotion is a guide to action for us. Are the numbers a better guide to action than emotions? Depends on what you're doing.
Starting point is 00:09:14 Yeah. I mean, let's go back to the moneyball case. Most of the time in baseball, the statistical information that you are way better off being guided by than your nose. Every now and then that's not true. But most of the time in life, you don't really have the numbers. Like how are you going to decide who you're going to marry or where you're going to live or even where you're going to go to college or even where you're going to have dinner
Starting point is 00:09:36 tonight. How do you reduce that to a statistical problem? We are kind of forced by the dearth of statistical information to constantly rely on our intuitive judgment. And so that it leaves us kind of sleepy to the possibility that there is some statistical information that could inform our judgment. Often when there is though, yes, I think it is better. Kahneman and Tversky showed just it's amazing how even when you're looking at people we would all think of as experts, doctors, that algorithms outperform the experts when rendering diagnoses,
Starting point is 00:10:06 that kind of thing. Yeah. One of the things that struck me exploring the story for Cautionary Tales was when Effective Altruism took this, what seems to be a little bit of a weird turn. So on the one hand, you've got the people who are like, okay, I've got $1,000 and I want to save the maximum number of lives. They crunch the numbers and they go, yeah, it's probably bed nets impregnated with anti-mosquito pesticides.
Starting point is 00:10:29 If you give out a lot of bed nets, you could save a life for whatever, $1,000, $3,000. So it's cheap. It's cheap. And that's better than maybe spending the money on vaccinations, even though vaccinations are good. And vaccinations are better than spending the money on work to improve governance. And improved governance is better than giving it to a donkey sanctuary
Starting point is 00:10:49 because all of the donkeys are living in luxury. They're all fine. So that all makes sense. Up to that point, you go, the effective altruists are on it. This is great. I'm with them too. So I feel the same way.
Starting point is 00:10:59 I'm with them too. They just make me, at that point, they just make me feel guilty for how I'm living my life Yeah, and and then They start having these conversations like well You know there could be a lot of people in the world in the future The future could go on for a very long time. We could be talking about trillions of people We could be talking about intergalactic civilizations or human civilization lasting a million years But maybe AI ruins all that.
Starting point is 00:11:26 So maybe if we have a workshop about AI safety, maybe there's only a one in a billion chance that it will do anything. But if it does something, it could save a trillion lives. So a one in a billion chance of saving a trillion lives, that's a thousand lives. That's really good. And so it's totally worth spending $50,000
Starting point is 00:11:44 on our AI workshop. At which point you start to, like, where did it go wrong? Or are they right? Am I wrong? So I had exactly the same feeling. It jumped the shark around 2012. And Toby Orr writes this book about existential risk, where he's putting numbers on the likelihood of various events wiping out humanity. And it's AI and pandemics and meteor strikes and nuclear war and so on and so forth. Very hard to put numbers on that. But of course you can't do the calculations without the numbers so they just accept roughly
Starting point is 00:12:14 good numbers and it's just not at all clear to me the numbers are okay at all. The math becomes much shakier. That's one thing that happens. The numbers become much bigger. You're not just talking about saving lives in Africa right now. You're talking about all future humanity, infinite people. And so the sums of money required to address these problems also becomes much, much bigger. A thousand dollars will be useful in Africa right now.
Starting point is 00:12:43 A thousand dollars right now to prevent AI from eating us all one day is nothing. Sam Beckman frees this down and does a back of the envelope calculation. He thinks he needs at least $100 billion to make a dent in this or that problem. So all of a sudden it gets very grandiose. And it's completely, even if they were never entirely interested in the emotional resonance of saving the life of a small child in Africa today, there was at least some emotional content there. And all of a sudden it vanishes. It's purely an abstract math problem.
Starting point is 00:13:20 And then, of course, as you just said, what can't you justify when the numbers are that big, the number of people you're going to save that big? What can't you justify doing to attack the problem? It seems perfectly rational to do whatever you need to do to make $100 billion to eliminate or at least defray this existential risk. Yeah, this is the end of the world we're talking about, the end of human civilization. You should be able to cut some corners at that point. So what you're doing also, if you think about it,
Starting point is 00:13:46 is you're puffing yourself up. You're a superhero now. You're now gonna save humanity. You may only be increasing the odds that humanity is saved by a few percentile, but nevertheless, you're in the realm of saving humanity. So it creates a grandiosity to the whole thing. And well, as I understand it,
Starting point is 00:14:04 Sam Bankman-Feed would have said, if I'm giving myself a 5% chance of saving the world, then I'm 5% of Superman, right? There we go. So this is how he was thinking when I first met him. And I thought it was bonkers, but I thought it was interesting. It's not wrong. You know, it's not obviously wrong.
Starting point is 00:14:21 It's probably obviously right that there are these extinction level events that could happen. The question is, like, is it bonkers to think you could do much about them? There's something that goes on when you're doing things that have such a remote probability of having any effect. It's very different from buying a bed net and saving a kid. The degree of uncertainty gets so high. And so when you look at how what happened to the EA movement around Sam, they had their arguments about this.
Starting point is 00:14:53 But the people who moved with Sam all bought into, you know, we're saving humanity. And they left behind a collection of original EA's who sort of disapproved and who continued to buy bed nets for kids in Africa. Matthew McNeese Do you think it is fair to blame effective altruism for what Sam Bankman Fried did and, you know, the fraud and the and him going to prison and all of that? To what extent can you track it all back to effective altruism? Alan Bates Can we assign fractional blame?
Starting point is 00:15:21 Matthew McNeese I think that's totally in the spirit of the whole thing. Alan Bates Okay, let's say in the spirit of Sam Bankman Fried, let's do little shares of blame. So I would say the little sliver of blame that effective altruism gets is in how it was sold to the Sam Bankman Frieds of the world. You are this nerd with a high sense of your own self-importance, but it hasn't been appreciated by the world, who is coming of age in a world where, very weirdly, you can get rich very quickly because your particular gifts are all of a sudden in demand at Jane Street and Hudson River Trading and Citadel
Starting point is 00:15:56 and all these quantitative trading shops and also tech firms. So all of a sudden, you're the most monetizable 22 year old on the planet for that maybe a professional athlete and If you unleash your money-making skills for the good of this movement, you're a superhero They're given a kind of license to behave in unorthodox ways Maybe it never occurred to will that it's somebody with over I guess I could commit a massive fraud and as long as I save the planet, that's fine. It possibly never occurred to the philosophers that anybody would think to do that.
Starting point is 00:16:30 It's possible. That's a funny thing you say that. I think that's possible. So I think if Sam Bankman-Free is a creature of anything, effective altruism is really important, but so is Jane Street, the high frequency trading world. They live to game systems. You're in there to figure out how to game markets. And people who are really good at playing the game are the people who win.
Starting point is 00:16:50 And Sam was really good at playing the game. Obviously, you're not supposed to break the law, but it puts you in a certain frame of mind when all you're doing is gaming markets day after day after day. You're looking for little edges. And if I were to assign another sliver of blame, kind of the grown-up world of finance, it was obvious to anybody who walked into the FTX that there was a basic conflict of interest. He had this private trading firm that he'd started first, out of which grew this public
Starting point is 00:17:20 crypto exchange, and that his private trading firm traded on the crypto exchange and it was in a little hut next door to it in the Bahamas. So some blame there, but of course also blame to Sam. What is it in Sam that leads him to do this? He danced his way out of lots of problems before. I think he thought he was so smart that he could kind of figure this out. And in fairness to him, I think he was overwhelmed. If you'd seen their business, your first reaction would be, this is pure chaos.
Starting point is 00:17:49 There's no organization chart. No one knows who the job actually is. The corporate psychiatrist is the person who knows the most about the business because he's the only one everybody talks to. You know, it's like one thing like that after another. What it didn't feel like and still doesn't feel like to me, and I get in trouble when I say things like this, is malice. What he wasn't and isn't is like this natural crook. He doesn't have cruelty in him. Oddly, he doesn't have a lot of dishonesty in him. He
Starting point is 00:18:17 has some. But up to the moment that it's discovered that he has the money in the wrong place, that he's used his customers' customers money do all kinds of things He didn't have permission to do there's not really a trace in Sam Bankman's Fried's life of Criminal behavior or corrupt behavior didn't cheat on tests didn't cheat at golf never played golf. You know you go back to his childhood There's no sign of anything like this. Yeah, and so it's not like this is who he is It makes it even a weirder event. Do you think he was a kind person or is a kind person?
Starting point is 00:18:48 Yes. I mean, you know, qualify it, but he behaved very well to the little people around him who helped him. He would get worked up and scream every now and then, but I think because he was on the receiving end of it when he was a kid a lot, I think he was highly sensitive to suffering. I think suffering bothered him. And he was also highly averse to conflict, like he was really uncomfortable getting into conflict.
Starting point is 00:19:19 He'd go and hide rather than fight. Every now and then he'd get upset and everybody around him would get scared. But it was very private. It was like watching a little volcano go off. So yeah, I think he is basically the kind of person and he to this day, I'm still in touch with him. He writes a prison diary and he sends it to me. He's pretty careful about his interactions with other people. He's in no conflict. People kind of like him because he kind of watches out for them a bit. I mean what he he did was really wrong, but when you know this personality, it's even stranger he did it.
Starting point is 00:19:49 Interesting. Very interesting. Well, Michael, stay with us. I'm going to give you questions from our loyal listeners, particularly on the subject of kindness and altruism, and we shall hear them after the break. We're back. I'm talking to Michael Lewis, the author of Going Infinite, The Rise and Fall of a New Tycoon, a book which tells the story of Sam Bankman Fried, the world's most infamous
Starting point is 00:20:20 effective altruists. And Michael, we've asked our listeners to send in their questions about altruism. Are you game for answering a few? Sure. So here's one from, I think from Nochem, apologies if I've mispronounced your name. They write, as an admirer of Mr. Lewis's books since Liar's Poker, when I heard he was answering questions about altruism, I had to ask this. According to the latest data I could find, the most important cause for effective altruists
Starting point is 00:20:48 after poverty and health is AI risk. The real challenges to the adoption of AI, but to put it so high on the list of causes to donate to seems misguided at best. Can you explain how AI overtook such obvious harms as global warming or the public benefits of the spread of free markets, free speech or democracy as the greatest good and effective altruists can do?
Starting point is 00:21:12 All the best, Nokum. This is one of those really great questions that I probably can't answer satisfactorily, but it is a great question. I think in the minds of Sam Beckman Fried and the people around him, they thought of these existential risks in two ways. How salient are they? How likely are they to cause problems to actually happen? And how tractable are they?
Starting point is 00:21:34 What can we actually do about this? And in Sam Beckman Fried's mind when he did that calculation, it wasn't AI that bubbled to the surface, it was pandemics. He threw much more money into pandemic prevention than AI. Even though he thought AI was the greater risk, he couldn't figure out what to do about it. Would this have been pre-COVID when he was throwing money into pandemic prevention? It was pre-COVID when he was thinking about it.
Starting point is 00:21:57 And when he started to have real money, we were in the thick of COVID. It was like, if a really even more deadly one comes along, how can we be better at defending ourselves against it, or detecting it and preventing it in the first place, that kind of thing. The one move he made in AI prevention, other than funding people with small sums of money who had interesting ideas, but he bought a big chunk of Anthropic, which was the kind of saloon de refusé for open AI people who thought that open AI wasn't paying enough attention to the risks. And that, funny enough, ended up being worth many, many billions of dollars to his creditors.
Starting point is 00:22:31 So why were they so alive to AI as an existential risk? Because your question is right. It's not obvious. It's not obvious to me. Let me flip it around and say, why are they not worried about climate change as an existential risk, for example? The answer to that is they think that climate change, no matter how bad it gets, they'll still be human beings alive. Yeah, and they're probably right. I mean, it's not going to turn the earth into Venus, right?
Starting point is 00:22:55 It could make things really uncomfortable. We could really regret we didn't do more earlier. But the idea that it ends the human race seems unlikely. So they're looking for things that will actually like end it absolutely end human civilization and there I guess you're looking at nuclear weapons and an AI are more plausible. Maybe that's right I had the capacity to do the job completely in a way even pandemic wouldn't I don't even think nuclear weapons are that plausible? So they could play out a scenario where there were no longer people because of AI. It throws a light on this, the difference between a long-term effective altruist and the rest of us.
Starting point is 00:23:32 The rest of us would go, hey, if there was a nuclear war and 99% of the human race was killed, and the rest of us were reduced to the Stone Age and had to build from scratch, that sounds incredibly bad. Whereas for an effective altruist, they're like, well, if you think really long term, on a two million year time horizon, it's not the worst thing. It just shows you where you can go
Starting point is 00:23:53 when you're untethered from normal human feeling and convention. That's the real thing about the effective altruist is that at bottom, it's a status movement. It's a group of people who feel underappreciated, slightly ostracized, who know in some ways they're smarter than their peers. Essentially, the starting point is everybody else is stupid,
Starting point is 00:24:13 they don't know. And it takes them to weird places. Let me throw in a thought from listener Matt. I'm gonna paraphrase, but it's relevant, I think, to what you're just saying, Michael. So Matt conjures up the idea that people who eat meat are, in a way, very good for pigs and chickens and cows. In the same way that people who want ducks are very good for ducks, because you created
Starting point is 00:24:38 the need for these animals to exist. Yeah. If a Martian came down and looked and said, you know, what are the dominant species on earth, then there's a good case that some of the dominant species are the species that humans eat. And the reason they're dominant, there are so many of them, have taken over all these different ecosystems is because that's what the humans want. Anyway, he draws a parallel between that perspective and effective altruism.
Starting point is 00:25:05 Is it too harsh to think of effective altruists as benign butchers ready to sacrifice real people today for the sake of a notional gain to humankind over eons? So there's no end to the suffering you can tolerate today as long as you're maximizing the quantity of humanity in the long run. I'd be willing to play this game with him, except the effective altruists themselves did not acknowledge they were willing to inflict suffering on people today. They just thought they were shifting their attention from people today to people in the future.
Starting point is 00:25:34 They were shifting their neglect, if you want to think about it the other way. So they weren't actively thinking of themselves as inflicting suffering on present humans. Although Sam did. Although Sam did. Although Sam did. Although the logic of their argument would lead them to do it if necessary, yes. Interesting. Question from Victor. Victor asks, why is no one talking about
Starting point is 00:25:56 why we need altruism in the first place? What are we doing wrong as societies that some people are left behind and there's a need for altruism? So I suppose this gets to the idea of if you had a benign government or if you had the right rules or the invisible hand of Adam Smith, but if we somehow organize society better no one would have to be kind or altruistic because all these problems would be solved anyway. This is a fair question too.
Starting point is 00:26:22 Without human sympathy, let's not call it altruism. Let's just call it kind of a basic sympathy for our fellow creatures. Without that, why would you bother to organize society in a kinder way? Like that is what's at the bottom of attempts to organize society so that you don't need it. It's a funny situation in a way because when I think of altruism, the kind of selflessness, every time you scratch the surface, you find some human motive for what they're doing. Selfish might be too strong, but it's not exactly selfless.
Starting point is 00:26:54 So when you ask this question, my mind goes in a completely different direction and my mind goes in the direction of do we actually even have altruism? Is there such a thing? I think it's more complicated than that. The minute you were sitting in a room with Sam Bankman Fried and his fellow effective altruists at FTX, you realize they were engaged in a kind of competition with other effective altruists. Who's going to be the biggest deal in the effective altruist community?
Starting point is 00:27:19 It's like, who's going to save the most future lives? They never put it quite so crudely, but it was in the air. Maybe that's good, you know, you've got this kind of killer competitive instincts and there could be people killing each other with machetes or it could be Wall Street traders trying to make the most money. The beauty of early Sam Beckman Fried and his crowd was, so we have this machine called Wall Street. It has gotten better and better at extracting rents.
Starting point is 00:27:45 The people who are the peak predators on Wall Street now are the high-frequency traders. The Citadels, the Jane Streets, the Jump Trading, the Hudson River Trading, Virtu, all these places, generating more money for individuals than have ever been generated on Wall Street before. All of a sudden, the kind of person who's good at that has this religion about giving the money away. I thought for a brief moment, there was this kind of Robin Hood thing that might go on. And in fact, Jane Street, which is the leader of the pack, Jane Street started to worry
Starting point is 00:28:20 that too many of the people that they were recruiting were effective altruists. And the problem with the effective altruists is they couldn't control them in the way you control a normal person who just wants a fourth house and a third yacht. They didn't have the same materialist needs. They were doing it for this kind of quasi-religious reason, and it made them much harder to manage. But I thought, what a great problem. Would it be cool if instead of reforming Wall Street because we'll never do it, that it sort of weirdly reformed itself because the kind of person who got into the position
Starting point is 00:28:51 of making the most money felt like it was his religion to give it away? That would have been cool. Yeah, it would have been. So a quick question from our listener, Richard, which I think is tied to that point. It's an etiquette question really. He says, if you give money to charity, should you tell everybody about it or should you just do it quietly? On the one hand, you should do it quietly because it's undignified and it's not about
Starting point is 00:29:15 you, it's about the charity. But on the other hand, if you tell your friends you're giving money to charity, then maybe that will encourage them to give as well. That's an interesting way to put it. I don't think people who put their names on buildings are doing it because they want to encourage other people to put their names on buildings. It's self-advertisement. I was raised to be very quiet about this sort of thing.
Starting point is 00:29:35 You give money if you always list it as anonymous. I've had organizations ask, we put our name on it, and I said, and when they want that, I let them do it. My answer to the etiquette question is, let the recipient decide whether they want your name on it or not. What's better for the recipient? And do whatever is better for them. Thank you, Michael.
Starting point is 00:29:57 Brilliant answers. Hold on with us and let's talk about your new book, Who is Government, after the break. We're back. I'm here with my fellow Pushkin podcaster, the brilliant writer, Michael Lewis. Michael, you very kindly agreed to come on the podcast and talk about Sam Bankman Fried and talk about going infinite. and then I got a message saying Oh, by the way, Michael has a new book out. You should probably ask him some questions about that I have to admit my heart sank a little bit. I thought I don't really have time to read a Another book in preparation for this podcast and and of course The moment I opened the first page I was immediately drawn in and I loved the book
Starting point is 00:30:44 The moment I opened the first page, I was immediately drawn in and I loved the book. So congratulations, but tell us about who is government. So I wrote a book about our set in the first Trump administration called The Fifth Risk, where I wandered around the administration, the executive branch, and got essentially an education from the various departments that Trump himself refused to get. So how the agriculture department worked, how the energy department worked, what went on inside these places. And the longer I spent there, the more taken I was with the actual characters in government. Whatever the stereotype of the bureaucrat is in the American mind, they violated it.
Starting point is 00:31:17 There were these breathtakingly devoted public servants who were experts in all kinds of arcane and in some cases spintingly frightening fields who were doing the work that kept the society together. And I thought, you know, there really was a project coming back and just doing profiles of these people. And it wasn't so much because I saw Doge coming. I'm surprised Doge came. It was more that I thought just generally the conversation around American government
Starting point is 00:31:47 had gotten so dumb because people didn't really appreciate what the government did and who the people were who did it. So I recruited six writers, Dave Eggers, John Lanchester, Geraldine Brooks, Casey Sepp, Kamau Bell, and Sarah Vowell. None of them really conventional journalists. Kamau Bell is a standup comedian. Dave and Geraldine Mills are mostly novelists, same as John. And just dropped them into the government, said, look, find a story.
Starting point is 00:32:12 And I did two of them. And the idea was just like inoculate the American public against these really stupid critiques of their government. I mean, these things ran one by one each week in the eight weeks running up to the election. And then we put them together because they got so much attention. The thing that really got me was actually the quality of the material. Like the first story is about a guy named Chris Mark, who figured out how to prevent the roofs of coal mines falling in on the heads of coal miners.
Starting point is 00:32:42 And you think, well, that's an arcane problem. And how big a deal could that be? 50,000 American coal miners killed by roof falls in the last century. And who knows how many more around the world. And there was just an imperfect science in how to keep the roof of a coal mine up. And how this person comes to his expertise, why he does it, how he does it, is literate. It was just like this stuff of novels. You find that over and over again in the government.
Starting point is 00:33:09 The story about Christopher Mark, we don't want too many spoilers, but his father is kind of among other things an expert in why Gothic cathedrals don't fall down. And he rebels against the father. Yeah, they don't get on at all. They don't get on. He leaves home and refuses to get a college education. He goes and joins the working class and then essentially reprises his father's career underground. And when I tell him that he gets angry at me, no, it has nothing to do with my father.
Starting point is 00:33:34 Yes, my father figured out how the roofs of Gothic cathedrals don't fall down, but that has nothing to do with how the roofs of coal mines don't fall down. Yeah, and of course, it's the same problem. It's the same. There's some little differences, but yes. From any kind of perspective outside of his own, it's the same problem. And so that there were these psychological portraits to draw that were just fun. I don't know if you think about what you do. I think about what I do is like gold mining.
Starting point is 00:33:59 I'm a prospector. I wander around the mountains of California, kicking up dirt under my feet, looking for a place where I can sink my pick and maybe find some gold. And it's kind of random when I find it. There's not a great science to the finding of the gold. You get good at sort of analyzing the landscape. You get a sense of where this might be and where it might not be, but it's still a lot of luck involved. I felt that with the Fifth Risk, I found this unbelievably rich vein of horror.
Starting point is 00:34:30 I could only scoop out a fraction of it because there was so much of it, and I was working with my hands and I had a bucket on my back. And there was all this stuff left behind, and I thought it would all be gone by the time I came back to it. And in fact, it's like no one would bother to go find the mind. No one's interested. There's this line by Geraldine Brooks, one of the writers in Who Is Government. She's profiling a guy who is a jujitsu instructor, tennis coach, fights terrorists and pedophiles,
Starting point is 00:35:02 but is a qualified accountant and works for the Inland Revenue Service. And she says, you know, if this was a novel, this would be malpractice, right? You can't just make this sort of stuff up. But because it happens to be true and this guy is a real person, I'm actually allowed to have him as a character in this story. And it tells you something. This person who is like the profit center of the United States government because he's busting up cyber crime rings and raking in their bitcoin and sticking it in the treasury.
Starting point is 00:35:28 He's generated billions with his team, billions of dollars of free money for the government. And the Trump administration has disabled him. They fired half his unit and he can't do what he did before. That our population isn't just completely outraged by this is incredible to me. But they wouldn't know to be outraged unless you knew the story. And you only know the story if you read Geraldine's piece because it's not in the news otherwise. It's astonishing. And he busted Peter Feilwings. There's one point where somebody from Hamas tweets and says,
Starting point is 00:35:57 oh, you can donate to the Revolutionary Cause. Send your Bitcoin to this address. And he redirects it, basically hacks their bank account. I have to say the details of exactly how this happened are lost on me, but anybody who donated Bitcoin, it ended up going to victims of state sponsored terrorism. And anybody who clicked on the Hamas logo got directed to Rick Astley singing, never gonna give you up.
Starting point is 00:36:19 So who says that bureaucrats don't have a sense of humor? Oh, no, they do. What they don't have, by and large, is a sense. No, no. No, no. No, no. No, no. No, no. No, no. No, no. No, no. No, no.
Starting point is 00:36:30 No, no. No, no. No, no. No, no. No, no. No, no. No, no. No, no.
Starting point is 00:36:38 No, no. No, no. No, no. No, no. No, no. No, no. No, no. No, no. No, no. No, what it is, these jobs self-select for people who really like doing big important things but don't care much about credit or money.
Starting point is 00:36:49 It's hard to believe that such people still exist in American life. Everybody else seems to be looking for fame and fortune. These are sort of like the opposite of reality TV stars. They got interested in a problem. They've worried the problem to death for 30 years. It's had enormous consequences and they don't expect anybody to pay attention. And then you get nervous when people do or at least the bureaucracy around them gets nervous. But that's right. That's really actually an important point.
Starting point is 00:37:12 That was the other thing all the writers noticed. And this is maybe by way of a way to explain like why there's this inefficiency in information about this. Is that the bureaucracy has gotten so used to all attention being bad attention. So a journalist shows up and says, I'd just like to know what you're doing and why you're doing it. And they assume this is going to end with this like a congressional hearing and me getting fired and humiliated and prevented from doing anything for the rest of my life. Whereas in fact, it's like, no, I just want to know what you're doing, because it's amazing
Starting point is 00:37:43 what you're doing. I want to tell everyone what you're doing. That's right. In a good way. They themselves are not as wary because they aren't doing anything bad. It's the political people above them are worried that this story will somehow make the White House look bad. And they can't imagine how it would make the White House look good.
Starting point is 00:37:59 Yeah. And there is a story to be told, I think, also about how the poor reputation of government bureaucracy just makes their life harder, irrespective of the politics itself. I think you profiled the woman who studies rare diseases. Yes. So she doesn't really study rare diseases. She's had a bunch of them. Her name's Heather Stone.
Starting point is 00:38:21 And let me just tell you how I found her, because that sort of explains what her role is in the world. So back when I was working on my COVID book, The Premonition, one of the characters was a researcher named Joe DeRisi, who's like, he was a superhero in bio research. And Joe, he was working on COVID, but at the same time, he had discovered what he thought might be a treatment for a rare brain-eating amoeba called Balamuthia. I'd never heard of this thing. Nobody had ever heard of this thing.
Starting point is 00:38:48 It was discovered in like the 1990s in the San Diego Zoo, but subsequently- Matthew Feeney Now, I'm now having nightmares about it, so thank you for telling me. It's terrifying. Matthew Feeney You should be. Yeah, because there's some tens of thousands of cases each year in the United States of people dying of unidentified encephalitis. It's like something's going on in their brain and they never figure it out. They die.
Starting point is 00:39:11 It's just encephalitis. There was a woman whose brain was being eaten by something they couldn't figure out what rolls into the UCSF emergency room. Joe gets involved. Eventually she dies. He gets involved too late, but he's able to identify the bug that's in her brain. It's this Balamuthia mandrillis, it's called. It's a brain-eating amoeba.
Starting point is 00:39:32 He then very cleverly, he says, well, what could you do to treat this? In his lab, he has his graduate students bombard Balamuthia with every drug that's been approved either in Europe or the United States and find that one drug a UTI drug used in Europe called nitroxylene kills the Balamuthia Clearly it doesn't kill people people are taking it for UTIs. So like why not try this and Lo and behold, even though it's rare pretty shortly thereafter someone else rolls into the emergency room, has the same symptoms, they find his Balamuthi in his brain, they give him the nitroxylene and he survives.
Starting point is 00:40:11 Yeah. I say to him, I said, man, that's great. Now everybody will know that there's at least something you can treat it with. And he says, nope, doesn't work that way. He said, maybe we'll be able to publish a scientific paper, maybe doctors will notice it. But there's every possibility that if you roll into a hospital with Balamuthia anywhere else in America,
Starting point is 00:40:29 that they won't even know we did this work. There's two problems, right? So one problem is you didn't run a randomized trial, so you can't be sure, maybe it was a flu. Correct. It's still interesting that you had a theory and you gave him the drug and he got better, so maybe. But then your second problem is,
Starting point is 00:40:46 how do you tell people that maybe? Yes, but with these rare diseases, you're never gonna have a randomized trial. And when they're fatal, if you roll in, Tim, with Balamuthia in your brain, would you rather them throw some nitroxylene in you or not? If they don't, we know what's going to happen if they don't. So your brain is going to get eaten.
Starting point is 00:41:08 That's bad. So why not? And the other side of this is that because it's a rare disease, the pharmaceutical industry has no interest in it. It's like there's no money in it. There's normally not enough people going to have this happen to them. By the way, the way you apparently get it is by ingesting dirt. So be careful with dirt when you're gardening.
Starting point is 00:41:27 Don't put your hands to your mouth. Anyway, so Joe says to me, this is a natural place for like government to intervene, to at least collate, because people like me are doing this stuff all the time. And there's a woman in the Food and Drug Administration named Heather Stone, who is kind of all by herself basically, decided to tackle this problem. She's created an app that try to gather every instance of a rare disease being treated around the world, how it was treated and what the outcome was.
Starting point is 00:41:56 There's hope for people who have rare diseases. But the government does not have the energy anymore to get behind Heather Stone's creation. They're unwilling to really promote it. She's a woman on her own, like going to medical conferences, trying to persuade doctors to plug in their rare disease treatment, their cases into her app. It's going nowhere. What happens, the story in the end is about a little girl, a six-year-old girl in Arkansas who rolls in and takes them forever to find that she's got Balamuthia, but she has Balamuthia.
Starting point is 00:42:27 The completely screwed up way in which her life is saved, the little girl in Arkansas's parents Google around and find that Heatherstone has been thanked by Joe DeRisi for helping him get his hands on nitroxaline. And they get personally in touch with her, and she makes sure that they get the drug and her life is saved. But at the very same time, just like 40 miles from Jodorisi's office in California, another little girl got Balamuthia and never heard of the cure and died. And it's sort of like a parable of good and bad government.
Starting point is 00:43:03 What kind of government do you want to have? A government that is emboldened and strengthened to actually follow through on this really It's sort of like a parable of good and bad government. Like, what kind of government do you want to have? A government that is emboldened and strengthened to actually follow through on this really good idea of how to deal with an intractable problem, or a government that actually has kind of lost its spirit and its energy. Because she created the app.
Starting point is 00:43:18 Cure ID, it's called. It should be useful. It should be something people pay attention to. It should be something that the government throws. It should be something that the government throws its credibility behind, but then the government has less credibility. And it's a story of the frustrating of the impulses of people who previously might have done
Starting point is 00:43:38 really great things in the government. So it's a sad story. It's a sad story masquerading as a happy story. But a kind of amazing story. It's a sad story masquerading as a happy story, but a kind of amazing story. It's absolutely amazing There's this thing That kind of springs from between the lines of the whole book These people feel like people Who have figured out?
Starting point is 00:43:58 The way to lead a meaningful life and this brings us back to effective altruism One of the keys to a meaningful life is to find ways outside of yourself. Find ways to live for things other than yourself. You know, Sam Bankman Fried groped towards this in his own weird way. But these people put their finger on it right away. There are problems I can solve and it will help others and never mind how much I'm paid or whether I'm acknowledged for it. And I'm not wired to do that. I need more attention than I should. But they are.
Starting point is 00:44:30 And we should be just grateful for them instead of heaping scorn and derision upon them. I've been talking to Michael Lewis. Michael is the author of Going Infinite and a new book with co-authors, Who is Government? Michael, it's been great to talk to you, thank you. Always fun, Tim. And you can listen to Michael's podcast against the rules wherever you get your podcasts. As for me, I will be back next week with another cautionary tale. For a full list of our sources, see the show notes at timharford.com Cautionary Tales is written by me, Tim Harford, with Andrew Wright, Alice Fiennes and Ryan Dilley. It's produced by Georgia Mills and Marilyn Rust.
Starting point is 00:45:18 The sound design and original music are the work of Pascal Wise. Additional sound design is by Carlos Sanjuan at Brain Audio. Ben Nadaph Haferi edited the scripts. The show features the voice talents of Melanie Gutridge, Stella Harford, Oliver Hembra, Sarah Jupp, Marceya Monroe, Jamal Westman and Rufus Wright. The show also wouldn't have been possible without the work of Jacob Weisberg, Greta Cohn, Sarah Nix, Eric Sandler, Carrie Brody, Christina Sullivan, Keira Posey and Owen Miller. Co-Optionary Tales is a production of Pushkin Industries. It's recorded at Wardour Studios
Starting point is 00:45:58 in London by Tom Berry. If you like the show, please remember to share, rate and review. It really makes a difference to us. And if you want to hear the show ad free, sign up to Pushkin Plus on the show page on Apple podcasts or at pushkin.fm slash plus. This is an iHeart Podcast.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.