Hidden Brain - Guessing Games

Episode Date: June 27, 2017

Pundits and prognosticators make predictions all the time: about everything from elections, to sports, to global affairs. This week on Hidden Brain, we explore why they're often wrong, and how we can ...all do it better.

Transcript
Discussion (0)
Starting point is 00:00:00 This is Hidden Brain, I'm Shankar Vedanta. We're surrounded by people who tell us they know what's going to happen in the future. A lot of people have no idea that Trump is headed for a historic defeat. Bear Sturz is fine! Don't move your money from there! That's just being silly! These predictions have a few things in common. The commentators have complete confidence in themselves. We as the audience love to hear them make a complicated world seem simple. And finally, no one ever pays a serious price for being wrong.
Starting point is 00:00:39 Donald Trump wins the presidency. The country is win the Super Bowl! Brady has his ship! Trump wins the presidency. Their stirs in the bargain bin sold to rival JP Morgan Chase for just two dollars a share. On today's show, our fascination with making predictions and why we may need a revolution in the way we make them. If you play the game the way it really should be played, the forecasting game, you're much more subject to embarrassment. We look at how some people actually are better than others at predicting what's going to happen in the future. Ironically, these aren't the people you usually find on television
Starting point is 00:01:17 blow-vading about what's going to happen next week. They're ordinary people who happen to know a very important secret. Predicting the future isn't about being unusually smart or especially knowledgeable. It's about understanding the pitfalls and the way we think and practicing better habits of mind. Phil Tetlock is a psychologist at the University of Pennsylvania. Over several decades, he's shown that the predictions of so-called experts are often no better than what he calls, dart throwing chimpanzees. After spending years criticizing forecasts and forecasters, Phil decided to look at people
Starting point is 00:01:59 who were very good at predicting the future. In his book, Super Forecasting, The Art and Science of Prediction, Phil explores how we can learn from these people to become better forecasters ourselves. Phil, welcome to Hidden Brain. So Phil, lots of people watch television at night, and millions of people feel like throwing things at their television set each evening as they listen to pundits and prognosticators, explain the day's news, and predict what's going to happen next. Of all the people in the country, you probably have more cause than most to hurl your coffee cup at the television set because starting in 1984, you conducted a study that analyzed
Starting point is 00:02:39 the predictions of experts in various fields. What did you find? Well, we found that pundits didn't know as much about the future as they thought they did. But it might be useful before we start throwing things at the poor pundits on the TV screen to consider their predicament. They're under pressure to say something interesting. So they resort to interesting linguistic gambits. They say things like, well, I think there's a distinct possibility that Putin's next move will be on a stonia. Now, that's a wonderful phrase, distinct possibility.
Starting point is 00:03:17 It's wonderfully elastic because if Putin does move into a stonia, they can say, hey, I told you, there was a distinct possibility it was going to do that. And if he doesn't, they can say, hey, I told you, there was a distinct possibility. It was going to do that. And if he doesn't, they can say, hey, I just said it was possible. So they're very well positioned. Now, if you play the game the way it really should be played, the forecasting game, and you use actual probability.
Starting point is 00:03:36 So you play it the way Nate Silver plays it. And you wind up with, say, 70% probability that Hillary will win the election a few days before the election in November 16. You're much more subject to embarrassment. If he said there's a distinct possibility that Hillary will win, he could have been very safely covered. Because when you ask people to translate distinct possibility of the numbers, it means
Starting point is 00:03:58 anything from about 20% to about 80%. I want to talk about how and why people often arrive at the wrong forecast. In your book, you tell the story of a doctor in 1956 who went in to see a dermatologist because he saw spots on the back of his hand. Tell me the story of Archie Cochran. Well, Archie Cochran was one of the early pioneers of evidence-based medicine. He was a great believer in running experiments, randomized control trials, and he was very skeptical of his fellow physicians. He was a great believer in running experiments, randomized control trials, and he was very skeptical
Starting point is 00:04:25 of his fellow physicians. He thought that they were systematically overconfident in much the same way that my data suggests that political pundits are systematically overconfident. So he was a skeptic by nature. Yet when he did get this cancer diagnosis from an eminent specialist. He didn't wait for the biopsy results to come in.
Starting point is 00:04:49 He just allowed them to go right into surgery immediately before the biopsy, which was a shame because it turned out that he did not have emoligrancy. We use that as an example of the bait and switch heuristic. So Archie Cochran didn't know whether he had cancer, but he was looking in the eye of a very renowned specialist who looked exactly like the sort of person who would know. So he took a really hard question, do I have cancer? And he substituted a somewhat easier question.
Starting point is 00:05:22 And that is, does this guy I'm talking to right here look like the sort of person who would know if I have cancer? And the answer to that was a resounding yes. So he took the answer to the easier question. He assumed it applied to the harder question and he forged ahead and got massive surgery that proved to be unnecessary when the biopsy results came in.
Starting point is 00:05:46 Archie Cochran didn't make that mistake again. He went on to have a very distinguished career where he developed a series of techniques to help doctors guard against the pitfalls of following their intuitions. He said, look, you might be very smart, you might know a lot, but there's something you have to do
Starting point is 00:06:02 before you act on your intuitions. This did not make him popular with his fellow doctors. Right. And Cochran now is honored, but I think during his lifetime, a lot of his colleagues, Tom, is a pain in the butt. He would run experiments that would repeatedly show that the intuitions of his colleagues were off base. There was a debate, for example, among cardiologists about how long heart attack patients should stay in the hospital after the heart attack, and whether the mortality would be higher if they went home or if they stayed in the hospital. And the conventional
Starting point is 00:06:35 wisdom among cardiologists was that it was better for them to stay in the hospital. But Cochran said, look, we have to test our intuitions. People would be happier, it would be less expensive, and so forth, if people were resting at home. And they thought that would be a terribly unethical experiment to run, because it could kill people, right? If it showed that the people getting the home rest were actually dying at a higher rate. But Cochrane persisted.
Starting point is 00:06:57 He wasn't a guy to take no for an answer, and they ran the experiment. And he also had a mischievous streak. So he presented the results to those two colleagues, and he flipped them around. He showed them a chart that indicated that the heart attack patients who went home were dying at a higher rate, just as the conventional wisdom predicted. And the doctors were outraged.
Starting point is 00:07:19 They said, Archie, we told you you're a fool. Look what you've done. You've caused great harm. You've got to stop this experiment immediately. And then he paused and he mischievously smiled. caused great harm, you've got to stop this experiment immediately. And then he paused and he misdivously smiled and he said, oh, I got the data turned around. It turns out that the people who are going home are living longer. The truth is, making predictions is difficult, but many biases also get in the way of making
Starting point is 00:07:40 accurate forecasts. When we make a prediction and it turns out wrong, most of us don't remember that we predicted something different. In fact, the hindsight bias prompts us to believe we'd gotten it right all along. We also hail people who make predictions that turn out right, whether in this talk market or politics or sports, but that keeps us from seeing the role of luck. Many people who get the right answer are just lucky, and some people who get's wrong are just unlucky.
Starting point is 00:08:08 Over time, the laws of probability mean luck can only take you so far. One reliable way to check if someone's success at predictions is driven by skill or by luck, is to have them make lots of predictions and see how they turn out over time. A few years ago, the federal government launched such an experiment. They conducted a forecasting tournament where thousands of people logged into computers to make thousands of predictions.
Starting point is 00:08:33 As time passed, the forecast could be checked for accuracy. It was a remarkable thing for the professional career government officials to do, to sponsor a project that had the potential to be embarrassing. We were one of five academic research teams that were competing to pull together the best methods of making probability estimates of events that national security professionals cared about. What kind of questions were they asking? All over the map quite literally. security professionals cared about. What kind of questions were they asking?
Starting point is 00:09:05 All over the map, quite literally. So there would be questions about violent clashes in the east or south China Sea. There would be questions about the Syrian civil war, about Russian-Ukrainian relations, about the Iranian nuclear program, Columbia Narco-Trafikers, literally all over the map. So you found, at some time, that people who were not necessarily trained experts in a specific domain were actually able to perform as well, or maybe even better than people who were the experts? Yeah, that's essentially right.
Starting point is 00:09:41 David Ignatius broke the story, I think, in 2013 in the Washington Post about how the best forecasters in the Good Judgment Project are outperforming professional intelligence analysts who had access to classified information, and we're working on an internal prediction tournament. How could this be? Are these people smarter than the rest of us? More knowledgeable? When we come back, I'll talk to Phil about what makes a super forecaster a super forecaster. Stay with us. If you were asked to pick someone to answer a difficult question about the
Starting point is 00:10:22 economy or foreign affairs, you might turn to an Oxford-educated public intellectual who writes a column for a very important newspaper. You probably wouldn't turn to a retiree in Nebraska who spends his time bird watching. But Filtaht Lock says, maybe you should. He was the opposite of Tom Friedman in the Superforecasting book. Tom Friedman, of course, being an eminent New York Times columnist, a very, very elegant writer, best-selling author, well-known for his explanations. But nobody has any idea how good a forecastry is.
Starting point is 00:10:57 And Bill Flack is an anonymous, retired irrigation specialist working in Nebraska, working out of the public library, or out of his home, and doing a fabulous job, making probability estimates in the good judgment project in the intelligence community forecasting tournament. So you helped discover Bill Flatt. Was he part of your team that you entered in this forecasting tournament? He was indeed. He was one of the very best. The very best forecasters in the tournament were called out each year.
Starting point is 00:11:26 The top 2% and we put them into elite teams called super forecasting teams. And we let them work together. And they did just a fabulous job. Super forecasters like Bill Flack turn out to have some things in common. Tell me about the kinds of philosophies they have and the kinds of thinking styles that you seem to find in common among many of these super-forkasters. I would say the most distinctive attribute of the super-forkasters
Starting point is 00:11:56 is their curiosity and their willingness to give the idea a try. And when I say the idea, I mean the idea that forecasting is And when I say the idea, I mean the idea that forecasting is a skill that can be cultivated and is worth cultivating. Because it doesn't matter how intelligent you are, or how knowledgeable you are, if you believe that it's
Starting point is 00:12:16 essentially impossible to get better at these kinds of tasks, you're never going to try, and it's never going to happen. It's as simple as that. So a big necessary condition for moving forward is having the attitude that this is a skill that can be cultivated and is worth cultivating. SuperForkasters tend to gather information and update their beliefs in a very particular way. Filthat Lock points to Aaron Brown, the chief risk officer of the hedge fund AQR. Before he was a big shot in finance, he was a big shot in the world of poker. He was a world-class poker player. And we quote him as saying that you can tell the difference between a world-class poker player
Starting point is 00:12:56 and a talented amateur. Because the world-class player knows the difference between a 60-40 bet and a 40-60 bet. In any pauses and said, oh, maybe like 55, 45, 45, 55. Distinguishing more degrees of maybe is an important skill. Why is that? Well, the very best forecasters are well calibrated. So when they say events are 80% likely, those events happen about 80% of the time.
Starting point is 00:13:22 When they say things are 90% likely, they happen about 90% of the time. So it makes a difference how frequently you update your forecasts. If you don't update your forecasts reasonably frequently, you're going to fall out of phase with events. And that means often making adjustments that are relatively small. You suggest that forecasters should do something that doesn't seem to be very intuitive. Instead of looking at the particulars of an individual case, you say forecaster should zoom out and figure out how often something has happened historically. So Daniel Conhamen is probably one of the greatest psychologists of the last hundred years,
Starting point is 00:13:58 and he calls that the outside view. And he says, people rarely take the outside view when they do forecasting. They normally start from the inside and they work out, but there's a big advantage to you as a forecaster from starting with the outside view and working in. Take another example, let's say you're out of wedding and you're sitting next to somebody who has the bad taste to ask you how likely to think it is this couple is going to stay married? And you look at the person, it's bad taste and all that, you see how happy the couple is and you can see it's a joyous occasion, you say, I can't imagine these people who are so happy together getting divorced,
Starting point is 00:14:34 I think maybe a 5% chance they're going to get divorced. Now, if you ask that question of a super forecaster, they'd say, well, what's the, let's look at the sociodemographics of the couple and let's see, what's the base rate of divorce within this socio-demographic group? Let's say it's 35 or 40 percent over the next 10 years. Okay, I think it's about a 40 percent chance to get divorced in the next 10 years. Now, that's not the end of the forecasting process. That's just the beginning. The real value, though, of doing it this way, of starting from the outside and working in, is it puts you in the ballpark of plausibility right away. 40% is a much more plausible number than 5%. Now then you could start adjusting the 40%. So if you discover things about the couple that suggest they really are
Starting point is 00:15:15 deeply bonded to each other and they've known each other a long time and they really understand each other and they've done great things for each other, you're going to lower your probability. If you discover that the husband is a sociopathic philanderer. You're going to lower your probability, if you discovered that the husband is a sociopathic philanthropist, you're going to raise the probability. Those are inside view sorts of pieces of data that would cause you to adjust. Or you might just see that having a small fight and say, well, OK, I'm going to move from 40 to 41%.
Starting point is 00:15:40 And that's one of the interesting things of a super forecast is they do a lot of updating in response to relatively small events. And most of the news, one of the interesting things of a super forecast is they do a lot of updating and response to relatively small events. And most of the news, most of the time, is what statisticians would call low diagnosticity news. It doesn't change things dramatically, but it does change things a little bit. And appreciating how the news gradually builds up toward one conclusion or another is a very valuable skill.
Starting point is 00:16:03 I'm wondering if one reason many of us start with the inside view rather than the outside view, is that just at an emotional level, that's how our minds think, that you see a couple, and you put yourself in the shoes of that couple, and you try and imagine what's happening in their lives, and we think in stories, and we imagine what life must be like for that couple, and we're trying to see
Starting point is 00:16:23 how that story will turn out, and we're trying to follow the narrative where it leads rather than do this much more, you know, abstract remote process of saying, let me start with the rough estimate of how often something happens. There's something about that in some ways that requires us to step outside this narrative frame that we often use to understand the world. I couldn't put it better. I think that's right. We're quite readily seduced by stories. Another example might be, it comes from research and this is again related to Daniel Coniman and his work on the conjunction fallacy. It's another source of error and forecasting. Let's say I ask you, how likely is it that in the next 10 years there'll be a flood in North America that
Starting point is 00:17:06 kills more than 1,000 people and ask you to make an estimate on that. Let's say I ask another person to make the estimate. How likely is it that there'll be a flood in California that will be caused by an earthquake cracking a dam leading to a massive outflow of water? Now, when I put the two questions together like that, it's pretty obvious that a flood
Starting point is 00:17:26 anywhere in North America, due to any cause, has got to be more likely than a flood in California caused by an earthquake cracking a dam. The California event is obviously a subset of the more general North American flood thing. But people don't see it that way. The California earthquake dam story is more like a story. It's like a story. You can actually put it together in a more meaningful way, whereas a flood anywhere in North America is kind of abstract and vague.
Starting point is 00:17:53 So people can transport themselves into the world, and you can imagine it's like a movie, like a Hollywood movie playing out. And they can see it happening. Yes, I can see that happening. And in fact, I can see it as you're speaking right now, Phil. And that pumps up the probability and that screws up your forecasting track record. So again, think of forecasting as a skill that can be improved with practice. When you're making a prediction, start with the base rate, the outside in view. Beware of the risks
Starting point is 00:18:23 of storytelling. Finally, amateurs make three kinds of predictions. Yes, no, and maybe. Professionals have many gradations of maybe, and the attached specific probability estimates to their predictions, allowing them to go back and learn where they went wrong. But even if you do all these things, I asked Phil how you can be really sure that predictions that turn out correct are because of good technique. You know, there's an old trick that they play in business schools to talk about the role of luck where they divide the class into pairs and they say, you know, have a coin toss between each person and then the winner of each of those coin tosses competes against another winner
Starting point is 00:19:03 and after 12 rounds there's one person who has who has declared the winner and of course that person in a business sense might seem to be extraordinarily good but really all that's happened is they've happened to win 12 coin tosses in a row they've just been very lucky. How do you distinguish between people who are lucky and people who are actually very good? Well that is indeed the $64,000 question and it comes up in finance too. I mean there are some finance professors out there who would argue that the really famous super investors were in Buffett or Ray Dalio and people like that. Aren't some sense like coins that come up heads 20 or 30 or 40 times in a row.
Starting point is 00:19:42 We have a lot of people competing in financial markets, and they're making a lot of predictions over long stretches of time. So you're going to expect some streaks. And when we get really streaky performance, we declare we found a genius. So the skeptics would say, well, Phil Tetlock is doing essentially the same thing here. He's anointing people who are essentially lucky.
Starting point is 00:20:02 So we built in a lot of statistical checks, but you can never be 100% sure. One of the other critiques of Phil's work is that the kinds of questions that super forecasters are answering are not really the questions that people want answered. Most of us are interested in the big questions. Who's going to win the current standoff
Starting point is 00:20:18 between the United States and Russia? Super forecasters tend to answer much more narrow questions. Is Russia going to invade Ukraine in the next 18 months? I think that's a very fair criticism of the first generation of forecasting tournaments that we put all of our effort into improving forecasting accuracy. Now, I'm not going to say the questions that people were trying to answer were trivial because they were by no means trivial. As a way, no means trivial, whether Russia was going to invade the Ukraine or the Syrian civil
Starting point is 00:20:47 war was going to drag out as horribly long as it has and so forth. I mean, the questions were really important things. But were they about the things that matter the most from a public policy perspective? Could we have made the questions more relevant to deep policy questions? I think the answer is yes. Science proceeds incrementally. You do one thing at a time. So what these first generation of forecasting tournaments generated is that there is a skill that is worth cultivating and can be cultivated.
Starting point is 00:21:20 Okay, that's step one. Now I think the next generation of tournaments should be focusing on exactly what you're saying here. I think we should be focusing as much on the insightfulness of the questions as the accuracy of the answers. I'm wondering if some questions just are too hard to forecast. We were having this conversation in 1917 rather than 2017 and someone would tell you, you know, 40% of the country right now is in farming and in 100 years from now, 2% of the country would be farming.
Starting point is 00:21:51 It would be reasonable to conclude that a good chunk of the country would be out of work that our unemployment rate would be 35%-40%. And that would be a perfectly reasonable conclusion to draw. And of course, they just would be no way to fully understand how that would not have come to pass. And I'm wondering in some ways, is the point of super forecasting to break questions down so that they're so simple, so that they're answering questions with accuracy, but not necessarily questions that are of importance. old joke of the guy who loses a coin on the street and he searches for the coin under the street lamp because that's where it's brightest, not where he actually dropped the coin. Is that what we're doing with super forecasting?
Starting point is 00:22:32 Is it possible that we are asking the kinds of questions that are the easiest to answer, the easiest to which we can attach these probability estimates? But the moment you go a little bit further out, we essentially are just guessing. Well, you've raised a host of interesting questions, and I would be the first to agree with you that forecasting tournaments can degenerate into games of trivial pursuit. And I'd also agree very much with your premise
Starting point is 00:22:56 that forecasting accuracy falls virtually to chance, to the dark throwing chimpanzee, when you try to forecast more than about 10 years out on geopolitical and geoeconomic questions of consequence. So both of those things are true. There's a limit on the forecasting horizon. And it's true that forecasting tournaments, if you don't pick the questions carefully,
Starting point is 00:23:18 so that they add up to something important, they can become trivial. So both are true. I don't think that they mean we shouldn't be doing forecasting tournaments. I think that we should be doing forecasting tournaments as well. I'm going to ask you one final question. And this is also I think a potential critique of super forecasting, but it comes in the form of a forecast that I'm going to make. The reason I think many of us make forecasts or look to prognosticators and pundits to make forecasts is that it gives us a feeling like we have a handle on the future.
Starting point is 00:23:48 It gives us a sense of reassurance. And this is why liberals like to watch the pundits on MSNBC and conservatives like to watch the pundits on Fox. A more cautious style that sort of says, the chance that Donald Trump is going to be impeached is 11.3% or the chance that you're going to die from cancer is 65.3%. These estimates run up against a very powerful psychological impulse we have for certainty that we actually want someone to hold our hand and tell us you're not going to die. We don't want a probability estimate.
Starting point is 00:24:21 We want actually an assurance that things are going to turn out the way we hope. So here's my last question for you. If someone advises people to do something that runs against their emotional need for well-being and reassurance, I'm going to forecast that that advice, however well-intention, however accurate, is likely not going to be followed by most people. What do you make of my forecast field? Well, I think there's a lot of truth to what you say. I think people, when people think about the future, they have a lot of goals, and they want to affirm their loyalty to their ideological tribe.
Starting point is 00:24:58 They want to feel good about their past commitments. They want to reinforce their preconceptions. So those are all social and psychological goals that people have when they do forecasting. And forecasting tournaments are very unusual worlds. We create a world in which only one thing matters. It's pure accuracy. So at somewhat analogous to the sort of world
Starting point is 00:25:21 is created in financial markets, or London bookies, or Las Vegas bookies, all that matters is the accuracy of the odds created in financial markets or London bookies or Las Vegas bookies. All that matters is the accuracy of the odds. I would say this. I would say people would be better off if they were more honest with themselves about the functions that their beliefs serve. Do I believe this because it helps me get along with my friends or my boss, helps me fit in, helps me feel good about myself or do I believe this because it really is the best synthesis of the best available evidence. If you're playing
Starting point is 00:25:49 on a forecasting tournament it's only the latter thing that matters, but you're right when people sit down in their living room and they're watching their favorite pundits, they're cheering for their team. It's a different kind of psychology. They're playing a different kind of game. So all I'm saying is you're better off if you're honest with yourself about what game you're playing. Psychologist Phil Tatlock is the author of Superforecasting, The Art and Science of Prediction. Phil, thank you for joining me today on Hidden Brain.
Starting point is 00:26:17 My pleasure. MUSIC This week's episode was produced by Raina Cohen, and edited by Tara Boyle. Our team includes Jenny Schmidt, Maggie Penman, René Clarre and Parth Shah. Our unsung hero this week is Portia Robinson-Migas from our marketing team. Portia helped us brainstorm many ideas in the early days of hidden brain and has remained a steadfast friend. Thanks Portia. You can find us on Facebook, Twitter, and Instagram
Starting point is 00:26:46 and listen for my stories each week on your local public radio station. I'm Shankar Vedantam, and this is NPR. you

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.