Planet Money - Runaway Recommendation Engine

Episode Date: May 26, 2021

Recommendation systems have changed how we choose what we want. But are they choosing what we want? | Subscribe to our weekly newsletter here.Learn more about sponsor message choices: podcastchoices.c...om/adchoicesNPR Privacy Policy

Transcript
Discussion (0)
Starting point is 00:00:00 This is Planet Money from NPR. There's a person who I'm pretty sure helped make the modern internet so addictive. But it was kind of an accident, and I'm not even sure he knows it yet. His name is Doug Terry. I'm a distinguished scientist at Amazon. Oh wow, a distinguished scientist. Should we call you Distinguished Doug? Except it's actually Distinguished Dr. Doug is what people call me. Distinguished Dr. Doug is an engineer, an inventor.
Starting point is 00:00:34 What's the thing that you're proudest of that you've invented ever? The thing that I'm... Unless it's a state secret. Well, there's this thing called the domain name system. All these little names that you type in these days with dots on it. You know, Amazon.com. You did that? That was my PhD dissertation was designing that.
Starting point is 00:00:57 Yeah. In 1990, Doug was a researcher at this utopia called Xerox PARC, this Disneyland of computer inventions that would give the world such hits as the personal computer, Ethernet, laser jet printing. Doug says their inventions were often just solutions to things that annoyed them. Right, which for Doug was his email. Back then, email didn't really have rules yet or etiquette. Doug's inbox every day was just full of chain letters and meeting scheduling threads and message board posts and reply-alls and people just spamming each other with random links.
Starting point is 00:01:33 It was chaos. And so people were getting hundreds of email messages per day, which by today's standards maybe isn't that much. But at the time, it seemed like a lot to be spending an hour of my day reading email. A whole hour. Right, I know, right? Let's trade. I want to go back to an hour of email a day. Seriously.
Starting point is 00:01:54 Doug's inbox, like everybody's inbox, just displayed emails in the order they were received. And you could do some simple kinds of filtering like send all emails from that super annoying co-worker directly to the trash. But there wasn't a really good way to do other, more subtle kinds of filtering. Filtering that required a little more judgment. And that's where I came up with this idea of saying, you know, let's figure out how to make humans part of a system.
Starting point is 00:02:21 Taking human likes and dislikes and turning them into a recommendation system. That was Doug and his team's invention. It was just for them and some of their colleagues to clean up their inboxes, but their solution would eventually come to define the modern internet. Doug's invention, this thing that he created to manage his inbox, was the first real recommendation engine. And today, the entire world runs on recommendation engines. Billions of dollars a year are spent and lost based on what is and isn't recommended to us. They determine everything that shows up on our Instagram feeds, our TikTok for you pages, our Twitter timelines. They determine what we watch or listen to, maybe which politicians we vote for. They tell us what we like and what we hate.
Starting point is 00:03:13 Hello, and welcome to Planet Money. I'm Kevin Roos, New York Times tech columnist. And I'm Mary Childs. And Kevin, you wrote this fantastic book, Future Proof, Nine Rules for Humans in the Age of Automation, which is how I learned about Doug and the world of recommendation algorithms. Your book is so good, Kevin. I love it. I read it every day. Thank you for recommending it. Today on the show, how Doug's mission to build a better email inbox kind of created the world as we know it. And not only changed what we watch on TV and what music we listen to, but really changed us, all of us, at a really fundamental level.
Starting point is 00:03:54 Distinguished Dr. Doug and his team are about to do this world-changing thing. They're going to start with fixing their email inboxes. with fixing their email inboxes. And in doing so, they will teach machines how to fake making value judgments, how to curate, how to determine importance. Because the problem, as he saw it, wasn't actually that people were getting too many emails. Like, that was bad. But the bigger problem was that all the emails looked the same. An email message from my boss saying, I need a response in five minutes is much more important than, you know, a message forwarded from a friend of mine saying, did you see the baseball game last night? Doug's inbox was organized by chronology, but he doesn't want
Starting point is 00:04:35 chronology. He wants priority. So he starts to teach his inbox what was important. Emails from one person to just one other person? Important. Send it to the top of the inbox. Emails to a bunch of people? Eh, bottom of the list. Maybe straight to the trash. You need sort of like the bouncer at the velvet rope of your inbox saying like, you can come in, but you look like trouble. Well, yeah, but not only the bouncer, but a personal bouncer, right? Someone who knows me personally and knows what, you know, the people I want to let in and the people I don't want to let in. So he and a group of coworkers build this filtering program and his inbox actually gets a little cleaner. His boss's emails are surfacing properly. Their next step is to add another layer from a totally different approach. One that draws
Starting point is 00:05:21 on the infinitely complicated and ever-changing human user. How do you talk to your bouncer and say, hey, man, not so many today? Or like, listen, you let somebody in last week who was a real scoundrel. I don't want that to happen again. I introduced like it and hate it buttons. So now in Facebook, you always see like it buttons or like buttons and so on. But the system I built was the first one that had a like it button. And it also had a hate it button. Wait, you invented the like button? You invented the like button and the domain name?
Starting point is 00:05:53 We are in the presence of greatness. So Doug is happily hitting like and hate on his emails. And it's working. His inbox is even better. So he invites some of his coworkers to try out their little inbox filter too. And they start hitting like and hate on their emails too, which leads Doug and his colleagues
Starting point is 00:06:13 to their next real big insight. Doug didn't need to rate every single email because his colleagues were doing it too with essentially the same emails. This right here is the insight that will change the internet forever. That Doug's colleagues' likes and hates could act as a filter on his inbox. He called this collaborative filtering. The whole point of collaborative filtering is I would say I liked or hated something,
Starting point is 00:06:40 and then that would help other people. Because that's the community aspect. Then other people could say's the community aspect. Then other people could say, oh, show me articles sent to this news group on baseball that Doug liked or don't send me anything that Doug didn't like. Doug and his colleagues built an email system that sorted people's inboxes.
Starting point is 00:06:58 It recommended the emails they should look at first by using this collaborative filtering idea. And it kind of worked. He says his email time got cut down from an hour a day down to 30 minutes a day. It improved his life and the lives and inboxes of his Xerox PARC co-workers. And because it was at Xerox PARC, which is where the architecture of the computer age was drafted, their collaborative filtering invention becomes part of the canon, a tool for every other
Starting point is 00:07:27 engineer who comes after. And that's what the next part of this show is about, taking Doug's invention and scaling it to billions of people, attaching it to superpowered artificial intelligence and using it to bring order to this chaotic world of online information. The modern era of collaborative filtering started 16 years after Doug and his team's invention in 2006. It started because this big company, Netflix, had a problem-tunity. If people couldn't find something they wanted to rent, they were more likely to cancel their subscription. If Netflix hooked them with a really good recommendation right away, they were happier. They would consume more. So anything that made those recommendations more precise and appealing to users, that was money. Netflix at the time was using a descendant of Doug's collaborative filtering idea as the
Starting point is 00:08:16 basis for their recommendation system, which drove 60% of rentals. But its recommendation system had plateaued. It wasn't improving. Here's Robert Bell, Bob. He is also a distinguished computer scientist, talking about Netflix's problem. And so they had some ideas for how to improve those recommendations, but they weren't quite sure how to do it. And so they came up with the idea of a contest. This is from a talk Bob gave about entering this Netflix contest. And to show that they were really serious about this, the idea was to release data to people, and they offered a million-dollar prize.
Starting point is 00:08:51 A million dollars. They called it, very creatively, the Netflix Prize. And to win, you had to improve Netflix's recommendations by at least 10%, to make better predictions than Netflix had about what people on its service would watch and like. And to help, Netflix published all of its data,
Starting point is 00:09:13 of all customers' anonymized ratings of movies and TV shows and whatever else. More than 100 million ratings in all, from 480,000 customers. A hundred million ratings? That was a ton of data, way more than anyone had had access to before. Big enough that engineers could actually play with it, which made it the most exciting thing to happen in data science since ever. Almost 50,000 teams downloaded the data, including Bob. He downloads this data and he starts digging it. And the surprise for everybody, I think, is that the most rated movie during this period was Miss Congeniality. That was rated by almost half of the users.
Starting point is 00:09:54 Bob and his team set out to make a better recommendation system. He starts with the obvious part of collaborative filtering, the part Netflix was already doing. Basically, matching up people with similar tastes. Like, oh, you liked Miss Congeniality. Well, we know that people who like Miss Congeniality often tend to like Legally Blonde, so maybe you should watch that. But this system couldn't take into account how weird people are and how weird movies are. Movies are very complex things, and humans are even more complex. And so what we see in a movie that we like or dislike is very hard to characterize in just a handful of factors. And so in some sense, in order to model all that, you almost need an
Starting point is 00:10:38 unlimited number of factors that you might consider. The original collaborative filtering required users to tell the algorithm, hey, I like what this guy likes, thumbs up or thumbs down. Bob and his team took this idea further. They realized that machines could figure out people's tastes on their own by grouping movies together and teasing out what they had in common, based on factors that you couldn't just see or guess. Right, like, I no longer had to say,
Starting point is 00:11:04 hey, I liked The Pelican Brief and Aaron Brockovich. Please find me other movies to watch. Now the machine could actually sniff out on its own these hidden threads between the movies I'd watched. Like, maybe it's suspenseful legal thrillers. Or maybe it's depth of character development. Or maybe it's something that only the machine sees. Bob and his team looked at data that was explicit, like ratings, and implicit. Like the data point that helped them the most was whether you had rated something at all.
Starting point is 00:11:33 If we're trying to figure out whether you'd like a science fiction movie, if you've rated other science fiction movies, even if you didn't rate them real high, that sort of says that you at least have some interest in science fiction. Using that factor got them to a 5% improvement over what Netflix was already doing, which meant Bob and his team were halfway to the prize already. From there, they added layer and layer and layer of different factors until they did it. They crossed the 10% improvement threshold. Bob and his team won the Netflix prize and the million dollars.
Starting point is 00:12:05 People who work on recommendation algorithms told me that this Netflix prize ended up being this huge moment in their field. That this thrillingly large data set that got all these brilliant engineers excited about building recommendation engines basically started a recommendation revolution. And this is when lots of other tech companies started realizing that recommendations could be used for so much more than just movies. They could help us figure out which music to listen to, which clothes to wear, which restaurants to patronize, maybe even which news was important.
Starting point is 00:12:37 And for a long time, people were mostly focused on building these cool, new, exciting algorithms. Scientists were stoked to see what they could do, and most users were like, oh, great, I really am interested in this. Thank you. But there's obviously another side to this, which is when do these suggestions stop being helpful nudges towards something I would get to eventually and start becoming programming? That's after the break. programming. That's after the break.
Starting point is 00:13:13 In the world of recommendation algorithms, there's this team of researchers who've been studying the code and the systems and their effects on us to find out if, in addition to guiding us to our preferences, these programs can actually change our preferences. guiding us to our preferences, these programs can actually change our preferences. My name is Jingjing Zhang, and in my work, I look at how online recommendations affect our decision making. So you would really think that Jingjing would be better than the average person at not getting sucked into a vortex of addictive stuff. In my personal life, if I go to YouTube and then search for certain videos, they will have related videos. I thought, oh, this must be highly relevant to what and then search for certain videos, they will have related videos.
Starting point is 00:13:45 I thought, oh, this must be highly relevant to what I'm looking for now. So I will consume more and more and then spend more time than I expected on YouTube or some other websites. I can definitely relate. And to figure out if recommendation systems are changing us, Jingjing and her team created a series of experiments using college students, basically fiddling with recommendations and seeing how those recommendations affected the students' behavior. And in one of those experiments, they were trying to gauge just how much of an effect these algorithms have on us, how powerful the power of robot suggestion is. So they chose something consumable but perfectly subjective, So they chose something consumable but perfectly subjective. Music.
Starting point is 00:14:35 What we did is that we took the top 100 songs from this annual billboard list, and then we provide different recommendations for these songs. The recommendations, which were on a five-star system, were manipulated, tweaked up or down. But the researchers told the students that the ratings were perfectly tailored to them. The students had to listen to the whole song and then say if they wanted to buy it, and if so, how much they would pay. And this is a great way to design an experiment, by the way, because it relies on something that economists call revealed preference theory, which basically holds that if you want to figure out what someone actually likes, look at what they actually pay for. And like Jingjing and her team are pretty sure that forcing students to actually listen to these songs would negate the impact of the recommendations. The students would form their own opinions and ignore the star ratings. You're not going to let an algorithm boss you around, right? What we found is that on average for each one star manipulation, that's going to increase the willingness to pay per price by 7% to 17%.
Starting point is 00:15:29 The students offered significantly more money for higher rated songs, even when those ratings were totally manipulated. Jingjing tested this and retested this, and the results were clear. When a machine tells us that we're going to like something, we trust the machine more than ourselves. And like, look, recommendations aren't all bad. Sometimes they're great. They save us time. They help us avoid decision fatigue. Sometimes I just don't want to like manually curate my own playlists of vibey electronic music. But here's what I worry about. These recommendation systems are getting so good that if we aren't vigilant, we're just going to end up drifting toward
Starting point is 00:16:10 whatever the machine tells us we like. This isn't just a problem of human psychology. It's also a computer science problem. Jingjing says it becomes a feedback loop. Those little drifts add up. Over time, this will make the system less effective, less accurate, and provide less diverse recommendations. Eventually, I know this longitudinal impact on the system will make the system provide similar items to everybody, like regardless of personal
Starting point is 00:16:39 test. Which, knowing the history, knowing how all of this got started, this is so far from the original goal. We built, or, okay, Doug built, these machines to save us time and help us find what we want. Now, instead of Doug's friends and colleagues all helpfully filtering each other's emails, it's just these giant, opaque algorithms owned by companies that want to make a profit off our behavior, recommending things that will keep us scrolling and clicking and watching forever and ever. Do you ever feel like you contributed to something that is not entirely good for society? I hope not. So up until this point, thanks, Kevin. I hadn't really felt that way. And to be fair to Doug and his team, this is not actually what they built.
Starting point is 00:17:34 They built an algorithm designed to read their minds. There's a huge difference between an algorithm built to read your mind and an algorithm designed to lead your mind, to take you to some new destination that you maybe wouldn't have arrived at yourself. Right. Like there's always been advertising, billboards and magazine spreads and TV ads. And you could see those things and decide like, no, I don't like those shoes. Not for me. But now we've got these machines telling us like, no, no, no, you do like those shoes. They are perfectly selected for you. Trust us on this one, because we know you better than you know yourself. And sometimes that might be true, but it's not always true. And so now the question for all of us is,
Starting point is 00:18:16 which of our preferences are actually ours and which were put there by a machine? and which were put there by a machine. If you know anyone who inadvertently changed the course of history and there's a great story behind it, let us know. Send us an email, planetmoneyatnpr.org. We are also on all the social media things influencing your choices. This week on TikTok, we explore the idea that you can only be friends with about 150 people. Today's show was produced by Dan Gurma with help from Emma Peasley, mastered by Gilly Moon, and edited by Nick Fountain.
Starting point is 00:18:51 Planet Money's supervising producer is Alex Goldmark. Thank you also to Jesse Boxstead and Camille Roth. Here's a decision I am not influencing you to make. Kevin has that new book out. This is basically a chapter from it. Future proof, nine rules for, Kevin, what is it again? Nine rules for humans in the age of automation. Humans in the age of automation. Stunning, beautiful. Resist that machine drift. Thank you for that very human, non-algorithmic recommendation. I have not been manipulated. Okay, Kevin, say your name. I'm Kevin Ruse.
Starting point is 00:19:24 And I'm Mary Childs. This is NPR. Thanks for listening. It's been a hundred years since the massacre that wiped out a thriving black neighborhood in Tulsa, Oklahoma. No one has been compensated for the homes, for the businesses that they lost. They returned to the Greenwood District and there was absolutely nothing. Listen on the Code Switch podcast from NPR. And a special thanks to our funder, the Alfred P. Sloan Foundation,
Starting point is 00:19:56 for helping to support this podcast.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.