Planet Money - Did two honesty researchers fabricate their data?

Episode Date: July 28, 2023

Dan Ariely and Francesca Gino are two of the biggest stars in behavioral science. Both have conducted blockbuster research into how to make people more honest, research we've highlighted on Planet Mon...ey. The two worked together on a paper about how to "nudge" people to be more honest on things like forms or tax returns. Their trick: move the location where people attest that they have filled in a form honestly from the bottom of the form to the top.But recently, questions have arisen about whether the data Ariely and Gino relied on in their famous paper about honesty were fabricated — whether their research into honesty was itself built on lies. The blog Data Colada went looking for clues in the cells of the studies' Excel spreadsheets, the shapes of their data distributions, and even the fonts that were used.The Hartford, an insurance company that collaborated with Ariely on one implicated study, told NPR this week in a statement that it could confirm that the data it had provided for that study had been altered had been altered after they gave it to Ariely, but prior to the research's publication: "It is clear the data was manipulated inappropriately and supplemented by synthesized or fabricated data." Ariely denies that he was responsible for the falsified data. "Getting the data file was the extent of my involvement with the data," he told NPR.Help support Planet Money and get bonus episodes by subscribing to Planet Money+ in Apple Podcasts or at plus.npr.org/planetmoney.Learn more about sponsor message choices: podcastchoices.com/adchoicesNPR Privacy Policy

Transcript
Discussion (0)
Starting point is 00:00:00 This is Planet Money from NPR. About a decade and a half ago, this one idea started showing up everywhere. The nudge. The idea of the nudge came from behavioral economics. Basically said, if you really understood how people think, their psychology, you could make a huge difference in the world just by doing these little tweaks. Just by nudging people in the right direction. Like if you want people to use less electricity, use social pressure.
Starting point is 00:00:31 You could send them a letter saying, hey, you're using way more electricity than your neighbors. Or organ donation. How do we get people to sign up to be organ donors? Turns out the secret has to do with the form at the DMV. This is one of the most famous behavioral economists around, Dan Ariely. He's giving a TED Talk that's gotten more than 10 million views. And yeah, he's right. Some European countries require people to opt out of being organ donors instead of opting in when they're getting their driver's licenses.
Starting point is 00:01:00 And making people opt out, that gets a lot more people on the organ donor list. Ariely, here in a different TED Talk, he's given about a dozen. So what have we learned from this about cheating? He's talking about some of his own research, focused on how you can get people to be more honest, with simple reminders. A lot of people can cheat. They cheat just by a little bit. When we remind people about their morality, they cheat less. Ariely has had some blockbuster research into the subject. Why people cheat and how to make them more honest.
Starting point is 00:01:33 In one study, he asked people to recall the Ten Commandments. You know, like, thou shalt not this, thou shalt not that. And then he had them play this game where they would be tempted to cheat. The moment people thought about trying to recall the Ten Commandments, they stopped cheating. In fact, another one of Ariely's big deal papers was about this one simple trick that would make people way more likely to tell the truth when they're filling out forms or paperwork.
Starting point is 00:01:58 For this one, he collaborated with a bunch of researchers, including another huge star of behavioral science, Francesca Gino. Here's Gino talking about that trick in 2015 on Planet Money. We should sign at the top of the form. That's it. Instead of signing at the bottom, we should sign at the top. Exactly. It's that simple. Makes sense. When you fill out a form and then at the very end you're asked to swear that everything you wrote was true, was it close enough? But if you swear to tell the truth ahead of time, your brain is primed for honesty. Gino and Ariely, they became these academic superstars for all this nudge research,
Starting point is 00:02:39 especially for their papers on how to get people to be more honest. And nudges? Well, nudges became all the rage. In the UK, the government created an entire organization. It's called the Behavioral Insights Team, or as they were nicknamed, the Nudge Unit. That's Michael Sanders. Back in 2014, he was the Nudge Unit's chief scientist. It was his job to test out nudges all around the world. Yeah, and Michael still remembers the moment this famous sign-on-the-top idea was brought before his team. He was at his office in London, and one of Gino and Ariely's co-authors came in to give a presentation on the research. What was the mood like in the room? I think, you know, we're British, so we don't have electrified rooms. People don't buzz like they do in America.
Starting point is 00:03:28 But I exchanged glances with a couple of senior colleagues. It's like, we could do this in this area. We could do this in this area. There's so many opportunities for this. This sign on the top idea, it was the quintessential nudge. So clever and so easy to implement and so surprisingly powerful. It's just so simple. It is. Many, many lies are simple. Allegedly.
Starting point is 00:03:54 That quintessential nudge idea was clever and was powerful. And also, it was a lie. Hello and welcome to Planet Money. I'm Nick Fountain. And I'm Jeff Guo. Dan Ariely and Francesca Gino have played this central role in taking nudge theory mainstream. And they've done really well for themselves. Gino's this big shot professor at Harvard Business School with all these speaking gigs and book deals. Ariely's got this fancy job at Duke with multiple best-selling books. One of them is even being turned into a TV show by NBC.
Starting point is 00:04:29 But now their famous sign on the top paper we were talking about before, that paper is at the center of a massive scandal. And these two superstars, famous for showing how to nudge people to be more honest, they are being accused of dishonesty themselves. Today on the show, the truth about dishonesty, dishonesty about truth, and the spreadsheet detective who cracked the case wide open. It was a lot of work to condense more than a century of contract law into a single 20-minute episode. There was an article that you sent me that had like an old English font at the top of it.
Starting point is 00:05:23 And I was immediately like, I'm not reading this one, Jeff. I'm sorry. You gotta be more selective. It's called research. How we did ours for our recent story on the rise of Internet contracts and how they got so sneaky. That is in our latest bonus episode for Planet Money Plus listeners, whose support helps make this show possible. Back in 2014, before two titans of behavioral science were engulfed in scandal, and back when nudges were still becoming the hotness, Michael Sanders was the chief scientist at the British government's nudge unit. It was his job to take some of these nudges and test them out in the field. You guys were sort of like special economics agents bringing behavioral econ to the far-flung corners of the world. That was sort of your mission? I mean, that sounds very
Starting point is 00:06:12 grand, doesn't it? Yes, I'd say that was our mission. Which is why when the government of Guatemala had a problem getting its citizens to pay their taxes, Michael was like, this is perfect for us. Guatemala has among the, if not the lowest level of tax compliance of any country across Latin America. And so the opportunity for up, the opportunity to improve things was just enormous in that context. You felt like you could make a real difference there? Yeah. The Guatemalan government was losing hundreds of millions of dollars every
Starting point is 00:06:45 year to tax evasion, money that could go to building roads and schools and whatever else. Maybe Michael, armed with all his little nudges, could help fix that. So Michael and a small team flew to Guatemala City to meet with tax authorities there and ask them how they collect taxes. And they said, well, one of the ways is called the value-added tax, which is kind of like a sales tax. Every month, businesses are supposed to fill out this form where they declare how much they owe in VAT. And most people fill in the form, well, most people who do fill in the form say, oh, I don't owe any. I don't owe any sales tax. And that's just not really credible, right?
Starting point is 00:07:25 I feel like in places like the UK and in Europe, the VAT is just part of the sales price, right? I mean, I'm not going to take lectures from Americans about how you should run a tax system. Michael, as you might expect, immediately thought back to that famous paper by Dan Ariely and Francesca Gino, the one where if you make people promise to be honest before they fill something out, they're much more likely to tell the truth than if they promise after they fill it out. Michael and his colleagues realized they could use this idea on the Guatemalan tax website. Before people entered their taxes, they would first have to pledge that they were going to tell the truth. And Michael tells the tax authorities, there's real research behind this,
Starting point is 00:08:10 done by professors from Harvard and Duke. And also to prove to you that this works, we're going to run an experiment to test it. Some people will get the new honesty pledge at the start and others won't. And what'd they say? They said, let's do it. Let's go. So we were in Guatemala for three days, and we went from, hi, my name's Michael, to we're running this study, and we're going to launch it, and we're going to run it for three months with three million tax returns, and we'll call you back in four months with the results. Four months go by. Michael gets the data set, and he stays up late crunching the numbers. Did the intervention work? It worked not at all. Not at all? Not at all. There are
Starting point is 00:08:52 few things in my life that I am, in a statistical sense, more confident of than that this didn't work. But it should have worked. The original research behind the whole sign on the top of the form thing had been done by a sign on the top of the form thing had been done by a team that included two of the biggest stars in behavioral science. And Michael's thinking, I can't even repeat their research. I can't even copy them right, right?
Starting point is 00:09:16 It's not like I'm trying to come up with something new and innovative and I'm struggling at that part. I just can't even copy them in a way that creates the right results. But here's the thing. Michael wasn't the only one having trouble getting the right results. Around the time he published his paper describing his failure in Guatemala,
Starting point is 00:09:38 other researchers started noticing their experiments were failing too. Eventually, one of Gino and Ariely's co-authors on that original sign on the top paper convinced them that they needed to revisit their findings too. So they tried to redo one of their experiments from the paper. And this time, it straight up didn't work. This was a big deal. One of the most famous nudge studies couldn't be replicated. But that didn't necessarily mean anyone had done anything wrong. This is how science works. Sometimes you just get fluke results. In 2020, a bunch of researchers, including Ariely, Gino, and the other original co-authors, published something that says,
Starting point is 00:10:17 yeah, maybe this sign-on-the-top intervention doesn't actually work. But more importantly, they also published, for the very very first time all the original data, all those Excel spreadsheets that they had used. Which, if you had some suspicions about what was going on here, might be very interesting. I'm Yuri. I'm a professor of behavioral science at Estada Business School in Spain. In addition to teaching behavioral science, Yuri Simonson has a little side project called Data Collada, where he and some colleagues evangelize to other behavioral scientists about how they should conduct and analyze their experiments. But what the Data Collada folks are really famous for are their investigations into research that they think smells a little fishy. Maybe a researcher juiced his regressions a little bit
Starting point is 00:11:05 to get the results he wanted. Or maybe he played fast and loose with his Bayes factors. Those tricky Bayesians. Anyways, Yuri in particular loves digging through people's spreadsheets for clues on how exactly they were put together and why. It's like a puzzle that nobody knows the solution to, and then you solve it. I mean, it's like true crime.
Starting point is 00:11:25 That's why true crime is so interesting to people and trying to figure it out, right? And it makes for a good podcast, too. Hopefully. Because of Data Collada's reputation, some tipsters reached out to Yuri and said, hey, you know how the original sign on the top paper had three different experiments in it? You might want to take a closer look at experiment number three, the one about car insurance. Yeah, the car insurance experiment. You know how your car insurance company occasionally asks you what the mileage is on your odometer? That is a sneaky way
Starting point is 00:11:56 of asking how much you've been driving because car insurers want to charge people who drive more, more. So drivers have an incentive to lie, which is where this sign-on-the-top intervention comes in. One of the superstar researchers, Dan Ariely, worked with a car insurer. Ariely said he asked the company to send out different versions of the same form to around 13,000 customers, requesting them to update the odometer readings
Starting point is 00:12:23 the company had on file for them. Half the people got forms with honesty pledges right at the top, and half of them got them on the bottom. And according to the paper, the ones who signed at the top, they were way more honest. But the tipsters who reached out to Yuri said, there's lots of weird stuff going on in the data. Like, even though when people originally reported their odometer readings, they often rounded to like the nearest thousand or 500, they did not do that when they updated their results. And the tipster said, look at the fonts. Why are half the cells in this column in Calibri and the other half are in Cambria? We were puzzled. Like, why would this happen?
Starting point is 00:13:00 Cambria. We were puzzled. Like, why would this happen? Yuri's like, fine, fine, fine. But immediately he hones in on this one thing. This was like piece of evidence seven. And I was like, I don't care about piece of evidence one through six. Like, this is it. Piece of evidence number seven, the distribution of miles driven.
Starting point is 00:13:26 of miles driven. Okay, so usually if you took a bunch of people, what you'd expect is that most of them are driving like a normal number of miles in a year, right? Like say 12,000 miles. And a few people might be driving way less, only like a thousand miles a year. And maybe a few people might be driving way more, maybe like 50,000 miles. But overall, the data would look like a bell curve with a big hump of people in the middle. But when Uri looked at the distribution of miles driven in Dan Ariely's experiment, he did not see a bell curve at all. He saw basically a flat line.
Starting point is 00:13:57 It just had an impossible shape. Like there were just as many people who drove 1,000 miles as there were who drove 2,000 miles, as there were who drove 10,000 miles, as there were 40,000 miles, all the way to 50,000 miles. And that's just crazy. You're saying anyone who works with data would immediately be like, there's no way. Like I shared that with people. People started laughing hysterically looking at the screen. It was just like a literal reaction. So it was just self-evidently faked. This data set was faked.
Starting point is 00:14:28 And as if it could not get any more obvious, even though there were lots of people who drove nearly 50,000 miles, it all cuts off right there. Not a single customer apparently drove more than 50,000 miles. When he saw that, Yuri immediately knew the exact spreadsheet command whoever faked the data had used. Round between, so give me a random number in between zero and 50,000. And if you do that, you get exactly this spreadsheet. So it's kind of remarkable. So you know how they did it. Yeah. Yeah. We figured out every step of the way, like how the spreadsheet was generated.
Starting point is 00:15:06 Yuri had done it. He cracked the case. It's just so, I've never seen anything so blatant in my life. It's just incredible. As far as smoking gun evidence goes, this for you is like, this is it. Yeah, nothing will ever match this. But there was another unsolved mystery here. Who was responsible for this fake data?
Starting point is 00:15:25 Yeah, what I wanted to look for was things that would point to who did it. And there was not a great way for him to figure that out. Before Yuri and his colleagues published what they had found, they reached out to the authors of that famous paper to give them a chance to respond. Usually when they do this, he says, it can get kind of contentious. But this time, everyone was pleasant. That was very unusual because we were saying, hey, you have some fake data there. And they were like, oh, we sure do.
Starting point is 00:15:50 Sorry. And so usually we get, I mean, we were not expecting this. So almost immediately, Dan Ariely replied and said, just to be clear, I was the only person responsible for this in this team. This meaning Ariely was responsible for getting the data from the insurance company. But Dan Ariely was not taking responsibility for forging the data. He basically blamed the insurance company. He said, quote, the data were collected, entered, merged and anonymized by the company and then sent to me. This was the data file that was
Starting point is 00:16:23 used for the analysis and then shared publicly. So Yuri knew the data had been forged, but without access to the original data from the insurance company, there was no way to prove who did the forging. The case was at a dead end. But not for long. In another experiment in that same paper, Yuri would find another smoking gun. And this time, there were fingerprints.
Starting point is 00:16:52 Also, after the break, we uncover a little evidence ourselves. So here's where we're at. After Yuri and Data Kalada pointed out that the car insurance experiment in that original Sun on the Top paper had fake data in it, the authors of the paper retracted it. If you look up the paper online, right there on the top, there's now a warning in red that basically says this paper is no good. But it's not like anyone lost their jobs over it. Now, Yuri and his team had actually gotten another tip that there might be something wrong with the work of one of the paper's co-authors, Francesca Gino. And they start looking at a lot of her research, including one of the experiments she worked on for the sign on the top paper. So they start digging in. Maybe this time they can do more than just show that the data was obviously wrong. Maybe they can find definitive proof of how the data had been manipulated.
Starting point is 00:17:57 So the experiment in question asked people to solve some math problems. They were told they would get paid a certain amount of money for each one they got right. They did their math and then they got to say on this official-looking form how many problems they'd gotten right. The ones who signed on the top of the form, the paper said, were way more honest. But unlike with the odometer data, there was nothing obviously wrong with the math problem data. Uri looked at it. He looked at it again. And then he thought back to this one secret little quirk about the spreadsheet program Excel. Excel remembers. Excel remembers.
Starting point is 00:18:32 And when I first heard about this, I could not believe this was true. But if you use formulas in Excel, and everybody uses formulas in Excel, Excel essentially keeps track of where you move stuff and it records all your moves in a secret file called CalcChain. CalcChain is something that I think 99.9% of people have never heard of before. But Yuri, of course, is part of the 0.1% who has heard of it. He loves it. Yeah, he wants to figure out it's very addictive. Like you get a data set and you start just checking the CalcChain file, you know, like immediately. So Yuri pulls up the calc chain file for experiment number one and realizes that a bunch of those spreadsheet cells, they have been moved by somebody.
Starting point is 00:19:17 And all of those cells that have been moved, every single one happens to strengthen the finding that signing on the top increases honesty. Like, how convenient, right? Yes. And this time, Yuri and his colleagues at Data Colada could do more than prove that the data was fishy. They could prove that somebody involved in the research had gone in and actually butts with the data. And that really sort of wraps it up in a very compelling way
Starting point is 00:19:46 because you have like confirmatory evidence from a completely different source of information. And by all accounts, this was Francesca Gino's experiment. Run under her supervision, the results analyzed by her. So just to reiterate, two experiments by two different superstars of the field, two researchers who are famous for showing how to make people more honest, their own studies about honesty are based on lies, on fraudulent data.
Starting point is 00:20:17 Like, this is bananas. And it's even bigger than that. Yuri's team has identified three more suspicious experiments done by Gino, pointing out what they call, quote, evidence of fraud. She's on administrative leave from Harvard, and two more of her papers have already been retracted. Dan Ariely's story is also getting more complicated. Remember, back in 2021, Ariely more or less blamed the insurance company for the fabricated data that had appeared in his odometer study. He said, quote, the data were collected, entered, merged, and anonymized by the company and then sent to me. This was the data file that was used for the analysis and then shared publicly.
Starting point is 00:20:58 Now, for years, the insurance company, which is called the Hartford, hasn't said much of anything about all this. But a couple of days ago, we convinced the company to finally go on the record. The company told us in a statement that they'd pulled the original data set they sent to Dan Ariely. And it looked dramatically different from the published data from Ariely's experiment. The company said they'd only given Ariely data for about 3,700 insurance policies. they'd only given Ariely data for about 3,700 insurance policies. But in Ariely's paper, he claimed he had data for more than 13,000 policies. That's a difference of almost 10,000.
Starting point is 00:21:35 In their statement, the company told us, quote, Though some of the data in the published study data is originally sourced from our data, it is clear that the data was manipulated inappropriately and supplemented by synthesized or fabricated data. The company basically confirmed everything that Yuri and the Data Colada people had said about Ariely's numbers. And the company made it clear that whatever had gone wrong with the data set had gone wrong after they had already given it to Ariely. We shared parts of the insurance company statement with Ariely, and he responded in an emailed statement.
Starting point is 00:22:09 Quote, As I said two years ago, I was responsible for the relationship with the insurance company that provided the data for the paper. I got the data file from the insurance company in about 2007, and I can't tell now who had access to it. Getting the data file was the extent of my involvement with the data. We also reached out to Francesca Gino, but she didn't make herself available to be interviewed for this episode, nor did representatives from Harvard or Duke. The co-authors of the original
Starting point is 00:22:33 paper declined to be interviewed. But the rest of behavioral science, they are commenting aplenty. So this is a disgrace. This is Michael Sanders again. He's the researcher who tried to bring the sign on the top thing to Guatemala, who left thinking that he had been the one who somehow messed up. Now he feels like the whole field was duped, not just him. Sometimes it does make me feel like we are stupid when you see Dan Ariely, who famously says, everybody lies a little bit. And Francesco Gino says, here's my book on how you can succeed at work if you don't follow the rules. And it's like you walk in with an eyepatch and a tricorn hat and a cutlass. Then 100 pages in the book later, we're like, oh my God, they were a pirate.
Starting point is 00:23:28 I never saw it coming. But Michael says the problem with all these fake studies is bigger than the feelings of a bunch of academics like him. In Guatemala, there was this big investment. Hundreds of thousands of people were put through this experiment that was never going to work. All of that time and money could have been spent on something different, a different intervention with a better chance of working if we'd known, right?
Starting point is 00:23:50 This is not a petty academic squabble about he said, she said. This is like, this has real impact in the real world. So I'm pissed off about that. I'm also pissed off by how stupid it is. So when you see like, eminent Harvard professor has committed research fraud, what you want is really sophisticated, clever ways of cheating. Not, oh, I thought I'd copy and paste these observations because they were higher and that'll make more result come together this is this is dumb and that annoys me but it's also it's also a source of terror because that means there'll be people out there
Starting point is 00:24:34 who have cheated in a really sophisticated way who we haven't caught and who we may never catch where michael is that's where his whole field is right now. Because of the questions around Dan Ariely and Francesca Gino's work, this cloud of uncertainty now kind of hangs over all of behavioral science. Like, what else might not be true? Ariely and Gino's research offered this really tempting story. That if you're just clever enough, you could come up with all these powerful nudges. One simple trick after another that could solve all the world's
Starting point is 00:25:12 problems. But now, that story is starting to look like just that. A story. If you know of anything sketchy going on in your neck of the woods, let us know. We're at planetmoneyatnpr.org. I'm at nfountainatnpr.org. We're also on all the socials media. This episode was produced by Emma Peasley with help from Willa Rubin. It was edited by Keith Romer and fact-checked by Sierra Juarez. It was engineered by Robert Rodriguez. Alex Goldmark is our executive producer. I'm Jeff Guo. And a special thanks to our funder, the Alfred P. Sloan Foundation, for helping to support this podcast.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.