The Ezra Klein Show - We Didn’t Ask for This Internet

Episode Date: February 6, 2026

Ragebait, sponcon, A.I. slop — the internet of 2026 makes a lot of us nostalgic for the internet of 10 or 15 years ago.What exactly went wrong here? How did the early promise of the internet get so ...twisted? And what exactly is wrong here? What kinds of policies could actually make our digital lives meaningfully better?Cory Doctorow and Tim Wu have two different theories of the case, which I thought would be interesting to put in conversation together. Doctorow is a science fiction writer, an activist with the Electronic Frontier Foundation and the author of “Enshittification: Why Everything Suddenly Got Worse and What to Do About It.” Wu is a law professor who worked on technology policy in the Biden White House; his latest book is “The Age of Extraction: How Tech Platforms Conquered the Economy and Threaten Our Future Prosperity.”In this conversation, we discuss their different frameworks, and how they connect to all kinds of issues that plague the modern internet: the feeling that we’re being manipulated; the deranging of our politics; the squeezing of small businesses and creators; the deluge of spam and fraud; the constant surveillance and privacy risks; the quiet rise of algorithmic pricing; and the dehumanization of work. And they lay out the policies that they think would go furthest in making all these different aspects of our digital lives better.Mentioned:Enshittification by Cory DoctorowThe Age of Extraction by Tim Wu“Fighting Enshittification” by Josh RichmanBook Recommendations:Small Is Beautiful by E. F. SchumacherManipulation by Cass R. SunsteinThe Rise and Fall of the Great Powers by Paul KennedyCareless People by Sarah Wynn-WilliamsLittle Bosses Everywhere by Bridget ReadJules, Penny & the Rooster by Daniel PinkwaterThoughts? Guest suggestions? Email us at ezrakleinshow@nytimes.com.You can find transcripts (posted midday) and more episodes of “The Ezra Klein Show” at nytimes.com/ezra-klein-podcast, and you can find Ezra on Twitter @ezraklein. Book recommendations from all our guests are listed at https://www.nytimes.com/article/ezra-klein-show-book-recs.This episode of “The Ezra Klein Show” was produced by Annie Galvin. Fact-checking by Will Peischel. Our senior engineer is Jeff Geld, with additional mixing by Isaac Jones and Aman Sahota. Our executive producer is Claire Gordon. The show’s production team also includes Marie Cascione, Rollin Hu, Kristin Lin, Emma Kehlbeck, Jack McCordick, Michelle Harris, Marina King and Jan Kobal. Original music by Pat McCusker. Audience strategy by Kristina Samulewski and Shannon Busta. The director of New York Times Opinion Audio is Annie-Rose Strasser. And special thanks to Natasha Scott. Subscribe today at nytimes.com/podcasts or on Apple Podcasts and Spotify. You can also subscribe via your favorite podcast app here https://www.nytimes.com/activate-access/audio?source=podcatcher. For more podcasts and narrated articles, download The New York Times app at nytimes.com/app. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.

Transcript
Discussion (0)
Starting point is 00:00:03 When was the last year that the internet felt good to you? I think everybody's different answers to this. Mine, I think, go fairly far back, maybe to the heyday of blogging, at least before the moment when Twitter and Facebook went algorithmic. But whatever your answer to it is, I have not found many people who think, 2026, right now this internet, with all of its anger and its outrage and its AI-slop, this is. is what we were promised. This is living at the technological peak. But even if there is this growing consensus, something went wrong with the internet somewhere and that it is driving our society somewhere we don't want it to go, there's not really a consensus of what to do about it. What to do about these giant platforms increasingly spammed up with ads and sponsored results, boosting
Starting point is 00:01:25 content that will keep us hooked and angry, isolating and dividing us and deranging our our politics, so making a few billionaires ever richer, held up by an army of low-wage workers and warehouses and on delivery bikes. Something has gone so wrong. But what do we do about it? My guest today have two theories of the case. Cori Doctero is a longtime blogger, an activist with the Electronic Frontier Foundation, and a science fiction writer. His new book is In Shittification, why everything suddenly got worse and what to do about it. Tim Wu worked as a special assistant to President Biden for technology and competition policy. He's a professor at Columbia Law School
Starting point is 00:02:04 and author of influential books on technology, including his latest to the age of extraction, how tech platforms conquered the economy and threatened our future prosperity, and shittification and extraction. Those are the ideas I wanted to put in play together here, and to also think about what solutions they might present. As always, my email as Rekline Show at n.ytimes.com.
Starting point is 00:02:37 Tim Wu, Corey Doctoro. Welcome to the show. Thank you very much. Great to be here. So I just learned that you both went to elementary school together. Yeah, that's true. In suburban Toronto, a weird little school with like 80 kids. It was kindergarten to eighth grade in one classroom.
Starting point is 00:02:54 Older kids taught the younger kids. We more or less were left to go feral and design our own curriculum. They chucked us out of the school on Wednesday afternoons to take our subway pass and find somewhere fun in Toronto to go do stuff. It was great. Is there anything about that school that would lead people to becoming sworn enemies of our tech overlords? Well, we love tech at the time.
Starting point is 00:03:14 I mean, we were early in on Apple II's, and frankly, that's where it all started in the way. You know, both of our books have this kind of pining for a lost age, and I think some of it is this early era of computing when we were just like so excited, so optimistic, everything was just going to be so amazing. And that, to me, a little bit was fifth grade or grade five, as we say, programming the Apple II. Can I slightly problematize that? So we were both also science fiction readers back then. And so, you know, 1981, First William Gibson story has been out for a couple of years.
Starting point is 00:03:47 I was pretty alive to the dystopian possibilities of computers at the time. So I wouldn't call myself optimistic. I would call myself hopeful and excited, but not purely optimistic. And I would also like to say that like, John Hodgman, nostalgia is a toxic impulse. And when I think about what I like about those days, it's not that I want to recover those days. It's more that I kind of dispute that the only thing an era in which people had lots of control over their computers could have turned into is one in which the computers had lots of control over them, that there is probably something else that we could have done. When you're spending time on the internet, what feels bad to you about it? So what I would do is contrast what happens when things aren't great now with how I felt about what happened when things weren't great before.
Starting point is 00:04:32 So I think when I was a larvum on the early internet and I saw things that sucked, I would think someone that's, I would think someone that's, going to fix this, and maybe it could be me. And now, when I see bad things on the internet, I'm like, this is by design, and it cannot be fixed because you would be violating the rules if you even tried. Tim, how about you? I feel it's like a tool I cannot trust. You know, I feel like the tools I like in my life, like a hammer, you know, I swing it, it, does something predictable. The internet seems like it's serving two masters. You know, I search for something. I get a much stuff I don't really want and I don't really know what I'm getting. I want to write like one email or check one thing. I end up in some strange rabbit hole and like three hours go by and I don't
Starting point is 00:05:16 know what happened. So I feel like I'm constantly at risk of being manipulated or taken from and I don't trust the tools to do what they say they're going to do. And I feel that makes using it kind of like living in a fun house. So I want to make sure I give voice to somebody who is not in the the show at the moment because this has a it's going to have the flavor of the prophet elijah has entered the chat yeah right yeah three three uh middle-aged guys who think the internet went wrong somewhere along the way uh when i was working on this episode of my producer one of the interesting tensions beyond the scenes was she doesn't think the internet is bad she thinks tictock is uh she said a perfect platform she's young kids and feels amazon is a godsend for a young parent obviously there are many people like
Starting point is 00:06:05 this who are using these platforms freely of their own volition happily. So what do you say to somebody who says, what do you fucking all talking about? Yeah, I mean, I guess I'll start. I think that, well, the middle-aged thing was they used to be better, which is I don't want to fall into that sort of situation. I just think the deal is not what it could be. And I think that maybe as a consumer who sort of lightly uses this, the internet is still useful. But if people, I mean, I have children too. And I think it's hard to deny that social media has been tough on kids and has had all kinds of negative effects on that. And that really started accelerating over the last 15 years or so. I think we have a highly polarized political structure, which is made worse by social media.
Starting point is 00:06:51 I think we have a problem with inequality, which has gotten worse and worse, and is accentuated by the fact that the margins are just so thin for independent business. And I also think this vision that it would be this equalizer, leveler, this technology that made a lot of people rich, not just a few people rich, that it was more or less reasonable and a lucrative thing to do to start your own business, that it would sort of change some of the challenges of inequality and class structure in the United States. Now, maybe those were very high hopes. But this is the key concept in my book. And I think key to understanding the economics of our time is the importance of class. platforms, which are any space or any institution that brings together buyers or sellers, speakers or listeners. Every civilization has had platforms. I was in Rome a few weeks ago, and you go to the Roman Forum, and there it is. It's all together. The buyers, the sellers, they have the courts, they have where people gave their speeches. You know, they're kind of the core of every civilization. And at some level why I wrote this book is I was interested in this question of what our fundamental platforms look like and how that reflects on the civilization we are building
Starting point is 00:08:07 because I do think they have a large impact. I think that's kind of undeniable. I think that things have gotten worse in many dimensions, and I guess it relates to my view of the state of the country as well. I think we've been in a better place in other periods of American history, and I think the Internet's not the only cause, but I think it's part of it. If I were having this conversation with your producer and we had some time to talk about it.
Starting point is 00:08:30 I would probably walk them through a couple of the undisputed ways in which some people have found the internet get worse for them. So Tim has talked a little about margins for small businesses. There's also people who are performers who found that the take
Starting point is 00:08:44 that's being sucked out of their pay packet every month is going up and up from the platforms. There's people who would not like to be snatched by ICE snatch squads who installed ice block on their iPhone only to have Tim Cook decide that ICE officers were a member of a protected class and remove that app. And now you can't install that app
Starting point is 00:09:02 because the iPhone only lets you install official apps. And I'd say, like, just because this hasn't hit you, I think unless you have a theory about why you are favored by these platforms, then you should at least be worried that this would come. And I would follow up by saying, like, let's not fall into the trap of vulgar Thatcherism. You know, Thatcher's motto was, there is no alternative.
Starting point is 00:09:25 And I think tech bosses would like you to believe that, too, that if you're enjoying having a conversation on Facebook with your friends, which I stipulate lots of people do. I think that's absolutely the case, and we should value and celebrate that, that you just have to accept that there is no way to have a conversation with your friends that Mark Zuckerberg isn't listening in on. And that to ask for otherwise than that would be like asking for water that's not wet. It's just like not possible. And what I'm militating for is not you don't like that thing you like. It's like, I like that you like the thing you like. I want to make it good, and I also want to guard it against getting worse, because just because it hasn't happened to you yet, it would be naive to think that it would never come for you.
Starting point is 00:10:08 So your books are two frameworks for understanding what I would call corporate capture of the internet. The way we went from the dream of a decentralized user-controlled internet to something that a small number of corporations really run and have enormous power over. And Tim, the term you focus on is extract. Corey, the term you focus on is insidification. So I'd like you just both to define those terms for me. What is extraction, Tim? What is inshidification, Corey?
Starting point is 00:10:37 So extraction is actually a technical economic turn that refers the ability of any entity or any firm with market power, monopoly power, to take wealth or other resources far in excess of the value of the good being, not only the value being provided, but also its cost provided. That's the technical definition. You know, outside of tech, it might be a pharmaceutical company. They have, you know, there's a rare disease. They have the only treatment for it. And, you know, maybe they're extracting as much as they
Starting point is 00:11:12 can, you know, $100,000 a year is about the usual. They pin those kind of diseases at. And that's the definition. And I think the idea of it comes from a sense, something I get from teaching at business school sometime, is that American business has, in my view, moved increasingly to focus its efforts on trying to find points of extraction as a business model, as opposed to, say, improving the product or lowering the price. You know, try to find the pain points where your customers really have no choice and then take as much you can, kind of like in a poker game when you go all in because you got the good hand. Now, there's always been a little bit of that in business, or maybe a lot, like in the gilded age. But the question is, what is the ratio and how much of business is providing
Starting point is 00:11:57 good services for good prices, you know, making a profit that's fine, and how much is just that different thing of extraction? So, Tim, I want to, before I move on to Corey, zoom out on something you said there, because a lot of that definition seem to turn on how you define value. And I mean, a lot of economists would say price is a method of discovering value. If you have a pharmaceutical, people are willing to pay $70,000 for it. That means they value it. $70,000, even if you think that is extractive. So how do you know when a price, when a profit, is actually extractive versus when we're just seeing that people value that product very highly and bully on the producer for creating something people value so highly? Yeah. So if someone,
Starting point is 00:12:43 for example, has no choice, but they are desperate, let's say, for water, and someone is able to sell them, you know, a bottle of water because they're dying for $100,000 or something like that. Yes, that person does value it at that level. But we would like to have an economy, and every economist would agree with this, is that an economy full of nothing but maximize monopoly prices where people are in a position to extract, a firm is inefficient for two reasons. One is too much money gets spent on that water versus, you know, other things like maybe pursuing an education. And second, that the entity that holds that much power actually has an impulse to reduce supply,
Starting point is 00:13:24 reduce output, and therefore produce less of the stuff so that they can extract the higher price. I mean, this is just classic monopoly economics, I guess I'm getting into. Everyone inside themselves has something they are willing to pay. But that doesn't mean it's a good society when you're constantly paying the maximum you are willing to pay in every situation.
Starting point is 00:13:43 For many of these platforms for a Facebook, for a TikTok, we're not paying for them. So when you say they are extractive, what are they extracting and from whom? So when you use Facebook, you are constantly being mined for your time, attention, and data in a way that is extraordinarily valuable
Starting point is 00:14:02 and that, you know, yielded something like $67 billion in profit last year. You know, things that feel free, is it free when you suddenly spend hours wandering around random things you didn't intend to? Is it free when you end up stuff that you didn't really want and wonder why you get it later? Is it free when you feel
Starting point is 00:14:22 that you've, you know, had your vulnerabilities exploited? I would say none of that's free. You're poorer, both in your own consciousness and in terms of what attention and you're controlled over your life and you're poorer probably in misspent money. I also wanted to react to something that you were sort of fainting at, Ezra, which is this idea of reveal preferences, which you often hear in these discussions, right? That if you let Facebook spy on you, no matter what you say about how you feel about Facebook spying on you, you have a reveal preference. And Tim used the word power when he responded to that. And I think that, you know, if you ask the neoclassicals, they'll say, well, we like models. And it's hard to model qualitative aspects like power.
Starting point is 00:15:02 So we just leave them out of the model and hope that it's not an important factor. And this is how you get these incredibly bizarre conclusions, like if you sell your kidney to make the rent, you have a reveal preference for having one kidney. But what we actually know when we give people choices, when the state intervenes or when there's countervailing power, is that often you get a different revealed preference. You know, when Apple gave Facebook users the power to tick a box and opt out of Facebook spying, 96% of Apple users tick that box. So the argument that Facebook users don't mind being spied on, I think, has blown out of the water when you actually give them a way to express preferences. And I assume the other four percent were like either drum.
Starting point is 00:15:43 or Facebook employees or drunk Facebook employees, which makes sense because I would be drunk all the time if I worked at Facebook. But I think it's hard to deny that people really don't want to be spied on if they can avoid being spied on. I think that's a good setup to in shittification. Yeah. In shittification, it's really a label I hung on both an observation
Starting point is 00:16:02 about a characteristic pattern of how platforms go bad, but I think much more importantly, why they're going bad now, because we didn't invent greed in the middle of the last decade. So something has changed. My thesis is that some exogenous factors are changed. So the pattern of platform decay is that platforms are first good to their end users while locking them in.
Starting point is 00:16:21 That's stage one. And once they know that the users have a hard time departing when they face a collective action problem or when they have high switching costs, you can make things worse for the end users safe in the knowledge that they are unlikely to depart in order to lure in business customers by offering them a good deal. And so far so good. I think a lot of people would echo that, but they would stop there. They would say, oh, you're not paying for the product, so you're the product.
Starting point is 00:16:43 So this is about luring in users and then getting in business customers will pay for it. But that's not where it stops because the business customers are also getting screwed. Because the business customers get locked in, and this power that the platforms end up with over their business customers is then expressed in stage three, where they extract from those business customers as well. They dial down the value left behind in the platform to the kind of minimum, like homeopathic residue. needed to keep the users locked to the platform, the businesses locked to the users, and everything else is split up among the executives and the shareholders. And that's when the platforms are a pile of shit.
Starting point is 00:17:20 But the more important part, as I say, is why this is happening now. Broadly, my thesis is that platforms used to face consequences when they did things that were bad for their stakeholders. And those consequences came in four forms. They had to worry about competitors, but we let them buy those. They had to worry about regulators, but when a sector is boiled down to a cartel, they find it very easy to agree on what they're going to do and make their preferences felt because they have a lot of money because they're not competing with one another and they capture their
Starting point is 00:17:49 regulators. They had to worry about their workers because tech workers were in very scarce supply and they were very valuable and they often really cared about their users and they could really say, no, I'm not going to insidify that thing. I miss my mother's funeral to ship on time and make it stick because there was no one else to hire if they quit and they were bringing a lot of value to the firm. But of course, tech workers famously thought that they they were temporarily embarrassed founders, and they didn't unionize. They didn't think they were workers.
Starting point is 00:18:16 So when the power of scarcity evaporated, they had not replaced it with the power of solidarity. And so now you have 500,000 tech layoffs in three years, and tech workers can't hold the line. And then finally, there was new market entry. There were new companies that could exploit something that I think is exceptional about tech. I'm not a tech exceptionalist broadly, but I'm an exceptionalist about this,
Starting point is 00:18:34 which is that every program in your computer that is adverse to your interests can be neutralized, with a program that is beneficial to your interests. And that means that when you create a program that is deliberately bad, you invite new market entrance to make one that's good, right? If you lock up the printer so it won't take generic ink, you just invite someone to not only get into the generic ink business, but get into the alternative printer firmware business,
Starting point is 00:19:00 which eventually could just be the, I'm going to sell you your next printer business. But what we've done over 20-plus years is monotonically expand IP law until we've made most forms of reverse engineering and modification without manufacture permission illegal, a felony. My friend Jay Freeman calls it felony contempt of business model. And as a result, you don't have to worry about market entry with this incredible, slippery, dynamic character of technology. And when you unshackle firms from these four forces of discipline, when they don't have to worry about competitors or regulators or their workforce or new market entry through interoperability,
Starting point is 00:19:35 the same CEOs goes to the same giant switch on the wall on the C-suite marked in shittification, and they yank it as hard as they can as they've done every day that they've shown up for work. And instead of being gummed up, it has been lubricated by an insidogenic policy environment that allows it to go from zero to 100 with one pull. And that's how we end up where we are today. All right, I want to bring these out of theory, though, Corey, I applaud how well structured that was on the fly. And have you both walk through this with an example that you use in your books? Sure.
Starting point is 00:20:10 And Corey, I want to start with you. Walk me through how you see an gentification is having played out on Facebook itself, not all of meta, but Facebook, where it started when it was adding value to users in the early days to where you feel it has gone now. Tell me your Facebook story. So Facebook really, it's Big Bang, it's 2006. That's when they open the platform to anyone, not just people with a dot EDU address from an American college. And Mark Zuckerberg needs to attract users. And his problem is that they're all using a platform called MySpace.
Starting point is 00:20:41 So he pitches those users and he says, look, I know you enjoy hanging out with your friend on MySpace, but nobody should want to use a surveillance-driven social media platform. Come to Facebook and we'll never spy on you. We'll just show you the things that you ask to see. So that's stage one, but part of stage one remember is that there's a lock-in. It's just the collective action problem, right? You love your friends, but they're a pain in the ass. And if the six people in your group chat can't agree on what bar to go to this Friday,
Starting point is 00:21:07 you're never going to agree on when it's time to leave Facebook or where to go next, especially if some of you are there because that's where the people with the same rare diseases you are hanging out. And if some of you are there because that's where the people in the country you emigrated from are hanging out. And some of you are there because that's where your customers or your audiences, or just that's how you organize the carpool for the kids of the league. And so we are locked in. And so that ushers in stage two, making things worse for end users to make things better for business customers. So think about advertisers.
Starting point is 00:21:36 Advertisers are told, you know, do you remember we told these robs that we weren't going to spy on them? Obviously, that was a lie. We spy on them from asshole to appetite. Give us pennies, and we will target ads to them with exquisite fidelity. So the advertisers pile in, publishers pile in two. They become locked to the platform. They become very dependent on it. And in stage three, advertisers find that ad prices have gone way up.
Starting point is 00:21:59 Ad targeting fidelity has fallen through the floor. ad fraud has exploded to levels that are almost incomprehensible. Publishers famously now have to put their whole article there, not just an excerpt, and Wobatide the publisher that has a link back to their website because Facebook's downranking off-platform links is potentially malicious. And so they don't have any way to monetize that except through Facebook's own system. And we've got a feed that's been, you know, basically denuded of the things we've asked to see. It has the minimum calculated to keep us there.
Starting point is 00:22:28 And this equilibrium is what Facebook, wants, but it's very brittle because the difference between I hate Facebook and I can't seem to stop coming here and I hate Facebook and I never coming back, it can be disrupted by something as simple as a live stream mass shooting. And then users, both for the exits, the street gets nervous, the stock price starts to wobble. The founders panic, although being technical people, they call it pivoting. And, you know, one day Mark Zuckerberg, like, arises from his sarcophagus and says, hearken unto me, brothers and sisters, for I've had a vision, I know I told you that the future would of arguing with your most racist uncle
Starting point is 00:23:03 using this primitive text interface that I invented so I could non-consensually rate the fuckability of Harvard undergraduates, but actually I'm going to transform you and everyone you love into a legless, sexless, low polygon, heavily surveilled cartoon character so that I can imprison you in a virtual world I stole from a 25-year-old comedic,
Starting point is 00:23:19 dystopian cyberpunk novel that I call the Metaverse. And that's the final stage. That's the giant pile of shit. All right. Corey, you got a good rant there, my man. I give you real... I could be a rapper if he decided to get into that line of business. The world is crying out for a middle-aged technology critic rapper. All right, let me ask you at least one question here, so I'm not just too taken in by your charisma, which is to say, I think that the counter argument somebody would offer is, I think two things.
Starting point is 00:23:49 One is for all the pivots, all the scams, by the way. I mean, I was a publisher during the era of the Facebook fire host to publishers and the era of pivot to video when Facebook videos were getting. these astonishing view counts. And one, they kept first all the money. They promised everybody, you know, come get this huge scale. We're giving all this traffic. You can build a business here. There was no business to build there at any significant scale.
Starting point is 00:24:12 And two, it turned out that the video view counts were fraudulent. Yeah. And so a huge amount of the news industry, among other things, pivoted to video, and it was based on lies. And there was a recent Rerder's report that Facebook was actually charging advertisers more for these things that they knew were scam. 10% of their ad revenue is ads for scams by their own internal accounting. So I am really not here to defend Facebook as an actor.
Starting point is 00:24:38 But one of the crazy things in midst all of this, a thing you really focused on there was moving from showing us what we had asked to see to showing us what I would say Facebook wants us to see. There was just the FTC versus meta case. Tim was, of course, involved in that. And one of the statistics that came out during it is that only 7% of time spent on Instagram is spent on things people you follow are showing you. Similarly on Facebook itself, it's under 20%.
Starting point is 00:25:04 I forget the exact number, but it's very low. They have moved under competition from TikTok, specifically, although not only, to these AI-driven algorithmic feeds showing you not what you have asked to see, but what they find will keep you there. And what they are finding is that it will, in fact, keep you there and people are coming back to it
Starting point is 00:25:23 and they spend more time on Instagram when you turn the feed into this algorithmic feed. This is the whole review. old preference thing that you were talking about earlier. My personal experience of Instagram when I go on it now is one reason I try to go on it less is that I can actually feel how much more compelling it is I like it less, but the feeling of getting pulled into something is much stronger. And so I think if you had Mark Zuckerberg arisen from his sarcophagus. I was going to say office because I'm a more polite person here. He would say, we did this under competitive pressure. TikTok was
Starting point is 00:25:58 eating our lunch, we stole a bunch of things from TikTok and now we're doing better. We also stole a bunch of things from Snapchat and now we're doing better because in fact we are under a lot of competition and we are incredibly good at responding to that competition in ways that our user base responds to. This is not in shittification. This is the magic of competition itself and you know that because actually look at our profit margin and look at how much we've changed. So let me say that I don't think competition is a good unto itself and I think it is absolutely possible to compete to become the world's most efficient human rights violator. The reason I like competition is because it makes firms into a rabble instead of a cartel. So in 2022, two teenagers
Starting point is 00:26:41 reverse engineered Instagram, and they made an app called OG app. And the way OG app worked is you gave it your login and password. It pretended to be you and logged into Instagram. It grabbed the session key. It grabbed everything in your Instagram feed. It discarded the ads. It discarded the suggestions. It discarded all of the stuff that wasn't a chronological feed of the people you followed that they posted recently. Facebook or Meta sent a letter to Apple and Google who obliged them by removing the app because there's honor among thieves. So if you want to find out what people actually prefer, you have to have a market in which people who disagree with the consensus that people are kind of gut flora for immortal colony organisms we call limited liability corporations and that
Starting point is 00:27:24 they are entitled to dignity and moral consideration as beings unto themselves, those people have to be offering some of the alternatives to find out what they want. But because under modern IP law, something called the Digital Millennium Copyright Act, it is a felony to modify the app without permission. When Meta sent the letter to Apple and Google, they agreed that they would side with meta. And because you can't modify those platforms to accept apps that haven't run through the store, that was the end of the road for OGA. But I think this is a little bit of a narrowed example. As somebody who gets a huge number of press releases for all these pro-social apps that are
Starting point is 00:28:04 built to compete with Instagram and TikTok and all of them, apps that are meant to respect your attention, apps are meant to be virtuous in a way these apps are not, and watches one after another, after another, basically go nowhere, get out-competed. The point I'm making is the example you're giving, they were able to basically say there was a term of service violation. Maybe they should not be allowed to do that, but there are lots of things that emerge and are meant to be better or different or something.
Starting point is 00:28:30 And people do. This is where I want to make sure my producer is a voice. There are people who just absolutely like TikTok. There are people who like Instagram. They know there are other things out there and they're not clamoring for a competitor or an alternative. I think suggesting that there is no capacity to switch is going a little far.
Starting point is 00:28:48 No, I'm not saying there's no capacity to switch. I'm saying the higher the switching costs are, the lower the likelihood that people will leave. You know, when we had pop-up ads in our browsers and real pop-up ads, the Paleolithic pop-up ad that was a whole new browser window that spawned 1 pixel squared, auto-played audio, ran away from your cursor. The way that we got rid of that was it was legal to modify browsers to have pop-up blockers. More than 50% of us have installed an ad blocker in our browser.
Starting point is 00:29:16 Doc Searle calls it the largest consumer boycott in human history. And as a result, there is some moderation. upon the invasiveness of what a browser does to you that is in marked contrast with apps because reverse engineering app because it's not an open platform is illegal under American copyright law. It violates Section 1201 of the Digital Millennium Copyright Act.
Starting point is 00:29:37 And so when we talk about how these platforms have competed their way into toxicity, we're excluding a form of competition that we have made illegal. For example, ad blockers, for example, privacy blockers, for example, things that discard algorithmic suggestions and so on. taking those off the table means that the only competitors you get
Starting point is 00:29:54 are firms that are capable of doing a sort of wholess, bolus replacement to convince you that, no, you don't want to use Instagram anymore, you want to use TikTok instead, as opposed to you'd like to use TikTok or Instagram, rather, but in a slightly different way that defends your interests against the firm's interests.
Starting point is 00:30:11 And we mustn't ever forget that within digital technology and living memory, we had a mode of competition that we prohibited that often served as a very rapid response to specifically the thing you're worried about here. I have a friend Andrea Downing who has the gene for breast cancer,
Starting point is 00:30:32 and she's part of a breast cancer pre-viver group that was courted by Facebook in the early 2010s, and they moved there, and this group is hugely consequential to them, because if you have the breast cancer gene, you are deciding whether to have your breasts remove, your ovaries removed, the women in your life, your daughters, your sisters, your mothers,
Starting point is 00:30:47 they're dying or sick, and you're making kids, care decisions. This group is hugely important. And Andrea discovered that you could enumerate the full membership of any Facebook group, whether or not you were a member of it. This is hugely important to her friends there. She reported it to Facebook. Facebook said that's a feature, not a bug. We're going to keep it. They sued. It was non-consensually settled when the FTC settled all the privacy claims, and they are still there because they cannot overcome the collective action problem that it takes to leave. Now, they will eventually. When Facebook is terrible enough, that community will shatter, and maybe it will never reform.
Starting point is 00:31:25 That is not a good outcome. All right, Tim, I want to go to your story here. One of the core tales you tell in your book is about Amazon. So walk me through the process of moving a platform from a kind of healthy, constructive platform to become an extractive platform through your kind of story of what happened with Amazon. So back in the 90s, when people were thinking of, about what is going to be great about the internet, there was this great, really serious promise
Starting point is 00:32:15 that it was going to make a lot of people rich. It was going to distribute wealth in new ways. And when people talked about how that was going to happen, a lot of it is like, well, you know, everyone's going to be able to have their own store, sell stuff, and people are going to make a lot of money that way. So that was a big promise, and I think it was an important promise.
Starting point is 00:32:34 Back in the day, it was mostly eBay that people were talking about. So then you have Amazon. Amazon is, you may remember, remember was once upon the time a bookstore. I do remember that, actually. It's how old I am. And, you know, their basic idea was to be bigger and we'll sell more stuff. At some point, they opened the marketplace, the Amazon marketplace, which was different because it was a platform. In other words, it was a place that people could come and sell their stuff. At first, it was used books. Then it spread into other markets. And they realized a few things. One is that fulfillment
Starting point is 00:33:07 would be very important. eBay in the old days, the sellers had to wrap it themselves and send it off. So that wasn't a very scalable model. And they had a good search engine. Amazon invested hard in search. And it worked.
Starting point is 00:33:21 And more and more sellers came. More and more buyers came. And so the Amazon marketplace took over eBay and became very successful. And at that point, I would say, you know, maybe around 2010
Starting point is 00:33:33 or something like that, was fulfilling what I think, you know, would call it the dream of the internet age, which is a lot of people will be able to go on this place, start their thing, make a lot of money. It coincides with the rise of the blog and small online magazines, you know, that whole era that we are talking about. During that period, Amazon's take was below 20 percent. It kind of depends how you count. But, you know, somewhere between 15, 20 percent.
Starting point is 00:34:00 Their take of what a small business is selling. Of the sales, yeah. Yeah. So if you sold like $100, they take $20. I mean, it depends a little there were some storage fees and so on. So, you know, it was a good place to make money. What changed, I think, was once Amazon had confidence that it had its sellers and it had its buyers more or less locked up. This is basically over the 2010s. They bought a couple companies that were potential threats to them, diapers.com, for example. It might seem ridiculous. Diapers, you know, could have been a kind of a way in to threaten them. Why don't you tell the diapers.com story for a minute?
Starting point is 00:34:39 It's a kind of famous story in Amazon, but I think it's worth telling. So there was the platform launched to be an alternative to Amazon, and their thought was, you know, every parent needs diapers delivered quickly. So why don't we make that at the beginning? In the same way Amazon started with books. And then Amazon saw this, thought it was kind of threatening, and in the strategy of the day, just bought them. Of course, the founder's pretty happy.
Starting point is 00:35:03 and Amazon managed basically to capture this market. And that's when I think it turned to the extraction phase. In the last 10 years, Amazon's strategy has just basically been for its marketplace to turn the screws and increase the fees, change the margins, so that many sellers are paying over 50% or more, basically the same as brick-and-mortar businesses. And Amazon prices are rarely any lower. actually have done a lot to try to prevent anyone pricing lower. And I think the one thing I would focus on is there, what they call advertising, which may be familiar to you as sort of the sponsored
Starting point is 00:35:45 results that you get when you're searching. So what's going on there is that sellers are bidding against each other, bidding down their own margins to get a higher up in the search results. And that little trick, that sort of one weird trick, has become this extraordinary cash cow. It's more more profitable than Amazon Web Services, which is sort of surprising. Last year, it was 56 billion. Just paying Amazon for higher rankings in their search results was $56 billion. Yes, $56 billion. It's looking like it's going to be over $70 billion.
Starting point is 00:36:20 Corey, when I'm searching on Amazon, and I see that Amazon's choice looks like a little prize, like that product won a competition, where a bunch of editors chose it as best one, what am I looking at there? So that is broadly part of this thing Tim was discussing, where they're piling on junk fees for the right to be at the top of the results. Sometimes it's more subtle where if you're not paying for prime and paying for fulfillment by Amazon and paying for all these other things, you aren't eligible. And the more of these you buy, the greater chance you have of being chosen.
Starting point is 00:36:50 But is that, are they literally paying to be Amazon's top choice? I mean, as a dumb consumer, maybe I look at that and I think, oh, this is some algorithmic combination of, is it the best seller, what are its reviews, etc.? So you're right that it is algorithmic, but the algorithmic inputs are not grounded primarily in things like quality or customer satisfaction. They're grounded in how many different ways you've made your business dependent on Amazon in such a way that every dollar you make is having more and more of that dollar extracted by Amazon. There's some good empirical work on this from Maria Mazukato and Tim O'Reilly, where they calculate.
Starting point is 00:37:29 that the first result on an Amazon search engine results page, on average is 17% more expensive than the best match for your search. So that's what you're saying is basically the Amazon top choice is the worst choice. So this really feels to me like a place where, to use Corey's word, things and shitified, that when I go around the internet now, when I play something in a Spotify playlist or click on a song I like and move to the radio version of Spotify, or when I search something on Google or when I search something on Amazon, these used to be very valuable services to me
Starting point is 00:38:04 to search for something on Amazon and see rankings weighted by how popular product is, how high the reviews are, right? I took the weighting of the search as to some degree of signal of quality. Certainly Google the whole idea was that what comes first in search was built on page rank
Starting point is 00:38:20 and it was going to be quality. Spotify had an algorithmic dimension to it, but it was supposed to be showing me things that people like me would like to listen to. And now there is so much sponsored content in every one of these results. And it is so unclear what is what and who is paying for what and why I'm getting this song or that result. You know, one reason I ended up on these platforms is because I trusted these results. And now I trust nothing.
Starting point is 00:38:49 Yeah, I mean, it's going back to the definition of extraction. I mean, we are kind of paying $70 billion collectively to make search. worse. So when does this move from this is just their business model? And if you want to find something else, like go buy something on Walmart, go buy something on Target, go buy something at Best Buy, you can do all those. I've done all those. I just ordered a blender from Coles versus we've moved to extraction and we should see it as a public policy problem. Yeah, I think that's a really great question. There's a kind of question we've faced, I think repeatedly in history when you start to have a a business model start to settle down, you see less real disruptive competition possible.
Starting point is 00:39:36 And I think at some level, once a market has settled, at some point you've got to call a limit. And we do that in many other markets. And Amazon is still a great way to find a lot of product. It's the world's largest marketplace. But I would say they're running themselves like an unregulated monopoly. Both of you spend a lot of time on the number of small acquisitions that these companies make. And maybe many of them get shut down or they aquire the top people, but there are also things that might have grown into something bigger or else. On the other side, sometimes it really is a case that a big player buying something smaller, they can scale it up into something new like Google bought. I mean, this was actually a fairly big acquisition,
Starting point is 00:40:21 but Waymo and kind of amazing. they seem to have made driverless cars work. And I think access to Google's compute and other things was not insignificant in that. And you can look at other cases where these companies buying something small, they're able to build it into something that ends up being a great option
Starting point is 00:40:39 in Microsoft Office or in Google Docs or whatever it might be. So how do you think about, on the one hand, the ways in which, I mean, I think it's what Google's made more than a thousand acquisitions over some number of years, How do you think about the ways in which that harms competition,
Starting point is 00:40:57 but also, you know, I've known founders who get acquired and are excited to get acquired because they think it will give them scale and the capacity to compete in a way they wouldn't versus Google just trying to do it itself? I think the antitrust level is one thing, but the sort of anti-competitive versus pro-scale level is like a much bigger challenge to the way Silicon Valley now works. And I'm curious to hear you talk through the pros and the cons of that.
Starting point is 00:41:21 Yeah, I think that's a really great question. You know, Joseph Schumpeter back in 1911 wrote a book about entrepreneurs, basically, and he said, you know, these very unusual people who are willing to go out and start to, you know, take these kind of risks, they have some vision, they do this kind of thing. And he thought they were essential to economic growth, that they were these kind of unusual, almost like superheroes, and would do these things and go out and take these chances. The United States economy in general has thrived because it has a lot of those kind of individuals. and they can start things. And I think we've erred too far in having all the brains under one roof. You know, it's starting to remind me of AT&T in the 60s or IBM, where they sort of became much more centralized about innovation.
Starting point is 00:42:11 And big ideas would never be developed. It became kind of group-thinking, I think, when the Justice Department did a deal, sued AT&T, tried to break them up. And they forced them AT&T to stay out of computing forever and also license all of their patents, including the transistor patent, and all kinds of people started quitting their jobs and say, I'm going to start a semiconductor firm. And there lies the origins of U.S. semiconductors and also, frankly, U.S. computing without AT&T. So I think we have done much better
Starting point is 00:42:40 with divided technological leadership. I frankly think that, you know, LLMs might never have started without OpenAI being an alternative force, because they're obviously threatening to Google's business model. Although don't, in a way, you have to give Google some credit on LLM specifically. You were talking about transistors a minute ago, but Google does the fundamental research in Transformers and releases it publicly and creates a many ways of the industry. But doesn't do anything with it internally until there's a competitor that threatens them. It's just striking how good of an actor they were for a period on AI specifically, right? Like treating it like they had a Bell Labs. I agree with that. It actually is a lot like Bell Labs in the sense that Bell Labs kept inventing stuff.
Starting point is 00:43:20 I mean, Bell Labs collected these, you know, a lot of amazing people and then never let things come to market. The Internet being probably the best example of it. Yeah, I think when you look at these companies in their acquisitions, what you see is that these companies very quickly suffer from what both Brandeis and Tim called the Curse of Bigness, that they find it very hard to bring an actual product to market that they invent in-house. When you look at Google, they've had like one really successful consumer-facing product launch, and that was in the previous millennium. and almost everything they made in this millennium failed. It either didn't launch or after it launched, they shut it down. Whereas their giant successes, their video stack, their ad tech stack, documents, collaboration, maps, navigation, server management, mobile, right? These are companies they acquired from someone else and operationalized.
Starting point is 00:44:08 And I'm an ex-ops guy. I'm a recovering sysadmin. So I'm not going to say that that's nothing. It is a skill unto itself, the careful work to make things work and make them resilient and scale them. But the idea that that has to happen under one roof, I think, is a false binary. I mean, one of the things Google did, arguably far more efficiently than they hired innovators is they hired operations people. And those are the people who really do the yeoman service at Google, because the innovators, the product managers, never get to launch.
Starting point is 00:44:38 They only get to buy other people's products and refine them. You know, it comes down to what you think of is the track record, I guess, of monopolize innovation. And it has some here. And it has some hits, but I'm saying a much more mixed model, I think historically is a lot stronger. If you look at the entire track record of U.S. innovation, I think monopoly innovation, you know, leads you towards AT&T, Boeing, General Motors kind of model, as opposed to what the best of Silicon Valley has been. And meanwhile, I think you mentioned aqua hires for people who aren't unfortunate enough to be steeped in the business of Silicon Valley. And aqua hires, when a company is purchased, not for the product it makes,
Starting point is 00:45:16 but because the team who made it have proved they can make a product, and then they shut down the product, and they hire the team. And aqua hires are, I think, a leading indicator of pathology in tech and investment. An aqua hire is basically a post-grad project
Starting point is 00:45:33 where venture capitalists sink some money into you pretending that you're going to make a product. It's a science fair demo in the hopes that the company will buy you, and in lieu of a hiring bonus, will give you stock, and in lieu of a finder's fee, will give them stock.
Starting point is 00:45:47 But no one's trying to actually capitalize a product or a business. I think any time you see a preponderance of aqua hires in your economy, that should tell you that you need to sit down and figure out how to rejigger the incentives because your economy is sick. Corey, we've been talking here about these markets as really having two players in them, which is, well, maybe three. We've been talking about users, sellers, and platforms. But something the yearbook focuses quite a bit on is a fourth, which we need to talk about to, which is labor.
Starting point is 00:46:18 There are huge numbers of people working for these companies, huge a number of people delivering Amazon packages and Walmart packages. And one thing that both of you focus on is the way in which as these companies become bigger and more dominant, their labor practices can become, I don't know if in shittification is the term you would use there, but shittier or more extra. active. Can you talk a bit about that side of it? What has happened to the labor practices? Yeah, I mean, we could talk about the other tech workers, right? The majority of tech workers drive for Uber or for Amazon or work in a warehouse, and they certainly don't get like free kombucha and massages and a surgeon who'll freeze their eggs so they can work through their fertile years. They're in a factory in China with suicide nets around it. But an example that kind of pulls this all together, how you get monopoly, regulatory capture, the degradation of labor
Starting point is 00:47:11 with technology that relies on blocks on interoperability, I think we could do no better than to talk about nurses. And I'm going to be making reference here to the work of Vina Dubell. It's a legal scholar coined a very important term, algorithmic wage discrimination. In America, hospitals preferentially hire nurses through apps, and they do so as contractors. So hiring contractors means that you can avoid the unionization of nurses.
Starting point is 00:47:38 And when a nurse signs on to get a shift, through one of these apps, the app is able to buy the nurse's credit history. And the reason for that is that the U.S. government has not passed a new federal consumer privacy law since 1988 when Ronald Reagan signed a law that made it illegal for video store clerks to disclose your VHS rental habits. Every other form of privacy invasion of your consumer rights is lawful under federal law. And so among the things that data brokers will sell anyone who shows up with a credit card is how much credit card debt is any other person carrying and how delinquent is it. And based on that, the nurses are charged a kind of desperation premium. The more debt they're
Starting point is 00:48:24 carrying, the more overdue that debt is the lower the wage that they're offered on the grounds that nurses who are facing economic privation and desperation will accept a lower wage to do the same job. Now, this is not a novel insight. Paying more desperate workers less money is a thing that you can find in like Tennessee Ernie Ford songs about 19th century coal bosses. But the difference is that if you're a 19th century coal boss who wants to figure out how much the lowest wage each coal miner you're hiring is willing to take, you have to have an army of Pinkertons that are figuring out the economic situation of every coal miner, and you have to have another army of guys in green eye shades who are making annotations to the ledger where you're calculating
Starting point is 00:49:10 their pay packet. It's just not practical. So automation makes this possible, and you have this vicious cycle where the poorer a nurse is, the poorer they become, the lower the wage they're offered. And as they accumulate more consumer debt, their wage is continuously eroded. And I think we can all understand like intuitively why this is unfair and why as a nurse you might not want it, but also like, do you really want your catheter inserted by someone who drove Uber till midnight the night before and skip breakfast this morning so they could make rent? This is a thing that makes everyone except one parochial interest worse off. And this is not a free-floating economic proposition. This is the result of specific policy choices taken in living memory by named
Starting point is 00:49:52 individuals who were warned at the time that this would be the likely outcome and who did it anyway. I think this is getting at something we're starting to hear a lot about, which is anger over algorithmic pricing of various kinds. So when I was walking up to do the podcast today, the Kairon on CNN, was about an investigation finding the Instacart was charging many different people, many different prices. And so the price you were seeing on Instacart wasn't the price. It's your price. And I could imagine a neoclassical economist sitting in my seat right now and saying, pricing becomes more efficient when it discriminates. That the market will be more efficient
Starting point is 00:50:32 if it can charge Ezra a higher price for kombucha if I'm getting that delivered because of things it knows about me and my kombucha habits and it charges somebody else a lower price because it knows they value the kombucha less or a nurse a higher price or a higher wage and a lower wage
Starting point is 00:50:48 depending on their situation. That in fact we're just getting better and better and better at finding the market clearing price And this is what economics always wanted. We're finally hitting the utopia of every person having, you know, the market clearing wage and the market clearing price. Why don't you agree with that? Yeah, I mean, the fundamental question is, is that really the kind of world you want to live in? In other words, do you constantly want to live in a place where you are being charged the maximum you would pay for something?
Starting point is 00:51:17 Now, you know, that could rebound to the benefit of people who are very poor. But in economic terms, it's always only about producers taking everything from the market. And I just think it's a very, you know, just moving away from the efficiency potentially of it. I think it makes for a very unpleasant lifestyle to be constantly feeling you're being exploited. And the other thing I'll say is there's also a huge amount of effort people make trying to move what category they're in and, you know, pretend to be poor. So I think it is overrated. And I guess it relies on overly simplistic models of what makes people happy. There's a way in which efficiency housing is an interesting term in economics.
Starting point is 00:52:02 Because in economics is in life, you want things to be somewhat efficient, but too much efficiency becomes truly inhuman. Yeah. I find this even in the very modest example of, like, personal productivity efforts, you know, it's great to have a to-do list. If I really force myself under the scaffolding of a to do list at all times, I feel like I cease to be a human being and become a kind of machine, always just getting things done and responding to the emails. And I think it was important, Tim, when you said it raises the question of what kind of world you want to live in. Because the truth is that I don't want to live in a maximally efficient world. I have other
Starting point is 00:52:43 competing values. You know, the competitive efficient market is good. up to a point and after a point, it becomes something corrosive to human bonds, human solidarity, just in time scheduling makes sense from the perspective of economic efficiency and not if you want healthy families in your society. And I think being able to articulate that question of what kind of world you want to live in, not just what kind of economy works on models, I think, is important and often a lost political art in my view. Yeah, I agree. And I feel there's some intuitive feelings. Like people feel it's unfair. People don't like being ripped off. People hate paying junk fees. The original word for that, by the way, was bullshit fees. But there was inside
Starting point is 00:53:27 government we felt we couldn't have the president say that. So yeah, I think that gets the heart of the matter. I mean, you had also talked about, you know, human attention. And human attention turns out to be quite commercially valuable. But do you really want every second of your time and every space you inhabit to being mined for your attention and its maximum value, even if that contributes to the, I guess, overall GDP of the economy. I mean, I'd like to have some time for my kids and friends in which no one's making any money. And, you know, it's an example of a commodity that is very close to who we are. At the end of your days, what your life was was what you paid attention to. And the idea that you can, with maximum efficiency, mine that at every possible
Starting point is 00:54:15 moment seems to me a recipe for a very bad life. I think one way to frame this rather than around efficiency is around optimization. And I think that we can understand that for a firm, the optimal arrangement is one in which they pay nothing for their inputs and charge everything for their outputs. So optimization, things are optimal from the perspective of the firm. When they can discover who is most desperate and pay them as little as possible, or who is most desperate and charge them as much as possible. But from the perspective of the users and the suppliers, things are optimal when you get paid as much as possible and are charged as little as possible. And so much of kind of the specific neurological injury that arises from getting an economics
Starting point is 00:55:06 degree is organized around never asking the question sort of optimal for whom. I mentioned before that we don't have any privacy law in this country. One of the things that a privacy law would let us do is to become unoptimable. All optimization starts with surveillance, whether it's things like TikTok trying to entice your kids to spending more time than they want to spend there, or whether that's advertisers finding ways to follow you around and hit you up with things that you're desperate for, or whether it's discrimination in hiring or in lending. All of this stuff starts with an unregulated surveillance sector. We have platforms that take our data and then sell it and use it and recycle it
Starting point is 00:55:47 and become sort of the lacota of information where they use the whole surveillance package. And we do nothing to curb that behavior. It is not an incredible imaginative lift to say that we might tell them to stop. I want to pick up on surveillance because when you talk about, about the harms to an economy working in a human way, I think that the new frontiers and how you can surveil workers, I think this is going to become a very big political issue and probably should be already.
Starting point is 00:56:43 I agree. The category that this falls into, it's broadly called bossware, and there's a whole lot of different versions of it. Like if your firm buys Office 365, Microsoft will offer your boss the ability to stack rank divisions within your firm by like how often they move the mouse and how many typos they make and how many words they type. And then this is amazing. They will tell you how you perform against similar
Starting point is 00:57:04 firms in your sector, which is like the most amazing thing I can imagine that Microsoft is finding customers for a sales pitch that says, we will show you sensitive internal information about your competitors. And apparently none of those people are like, wait, doesn't that mean you're going to show my competitors sensitive commercial information about me? So you have this on the kind of broad strokes level. But I have this notion I call the shitty tech. technology adoption curve. If you've got a really terrible idea that involves technology, that's incredibly harmful to the people it's imposed on, you can't start with me. I'm a mouthy, white, middle-class guy with a megaphone, and when I get angry, other people find out about it. You have to
Starting point is 00:57:42 find people without social power. And you grind down the rough edges on their bodies. You start with prisoners. You start with people in mental asylums. You start with refugee. And then you work your way up to kids and then high school kids, blue color workers and pink color workers and the white color workers. And it starts with like the only people who eat dinner under a CCTV are in Supermax. And 20 years later, it's like, no, you were just dumb enough to buy a home camera from like Apple or Google or God help us all Facebook, right? So that is the shitty technology adoption curve. And if you want to know what the future of workers is, you look at the least privileged workers at the bottom. And then you see that technology working its way up. You look at drivers for Amazon.
Starting point is 00:58:19 They have all these sensors pointed at their faces, sensors studded around the van. They're not given a long enough break even to deal with things like period hygiene. And so women who drive for Amazon who go into the back of the van to deal with their periods discover that that's all on camera because that's all being recorded. All of this stuff is subject to both manual and automated analytics. And at one point, Amazon was docking drivers for driving with their mouth open because that might lead to distraction while driving. And so, as you say, it kind of denudes you of all dignity. It really is very grim. And, you know, Tim and I used to ride the Toronto Transit Commission buses to school in the morning when we were going to elementary school. And we loved the drivers who would sing and tell jokes and remember you.
Starting point is 00:59:04 This is the thing that makes working in the world, being in the world great. It's having a human relationship with other humans, not having standardized labor units that have been automated and standardized to the point where they can be swapped out. You know, if you give a cashier a cash register instead of making them add up things on the paper, you could give them the surplus to talk with the customers and have a human relationship with them. Or you could speed them up so that you fire nine-tenths of the cashiers and you take the remainder and you make them work at such an accelerated pace that they can't even make eye contact. Tim, there were things in Corey's description there in his answer there that in my view, we should just make a social decision to outlaw.
Starting point is 00:59:47 Like, I am willing to say politically, I want to vote for the people who think you can't eyeball surveil workers. And if other people want to stand up and say the surveillance of workers' eyeballs is great, like, that's a good values debate to have in a democracy, and I know where I fall on that. Then there are other things, right? I'll build on the cash register example to say that I really struggle with what I think as a public policy measure, one should think about the rise of automated checkout and the way we've seen it. I watch people turned into these managers of machines. They've gone from being somebody who did check out with me and asked me how my day was and asked them how their day was. And now they get called over
Starting point is 01:00:28 because the three apples I put on the weighing machine didn't weigh in correctly. And it seems dehumanizing to them, dehumanizing to me. I also get it. How do you think about weighing that? There's the stuff that is genuinely, like, grim and dystopic and maybe we should just outlaw. And then there is stuff like,
Starting point is 01:00:47 the just generalized automation in which there genuinely can be a consumer surplus from that. Like time is a surplus for me. Things moving faster is a surplus for me. More checkout stations is a surplus for me. And there's a cost on the other side of it. Well, the first thing I'd say is we should be making
Starting point is 01:01:03 more of these kind of decisions about what we really care about and what world we want to inhabit. I mean, one of the things that I think happens is by default, we don't pass any laws or have new ethical codes. I mean, ethics does a lot of work and we just sort of allow a Trump card to new stuff because it's new.
Starting point is 01:01:22 And, you know, I get that you don't want to ban everything new that shows up, but I feel that we have over the last 15 years or so sometimes just taken a position that, you know, the people don't get to vote on this. I mean, a good example is everything to do with children. I don't think there's a lot of people who think it's a great thing to surveil children and have targeted ads for children and try to create addictive technologies for children. And when I worked in government, we tried to pass just basic, even child privacy laws. We couldn't get a vote ever.
Starting point is 01:01:53 And so one of the things that's going on is we're not even deciding these things as society. And that gets to, you know, the problem of Congress not taking votes on popular issues. But I also think this relates to our conversation earlier about competition and when it's good and when it's bad. Because I think for almost any endeavor, there's such a thing as healthy competition, competition and such a thing as toxic competition. We were talking about the attention markets earlier. What is good, healthy competition in the attention markets? It's like making really great movies, new TV shows that people love, podcasts that people want to listen to. Toxic competition was the stuff you're talking about, essentially different forms of manipulation and addiction.
Starting point is 01:02:40 And we've had this kind of like hands off. We cannot try to direct things in a positive direction. I think that has been a giant mistake. So first I would say we have to even try to make the decisions. You know, how would I do the trade-off? I mean, I guess I would start with the most unredeeming toxic stuff and ban that first and then see if we can, you know, I mean, that's maybe easy, but we haven't been able to even do that. And I was sort of shocked when I worked in government that we just could not get a vote on what seemed like stuff. Privacy laws. Like how many Americans, you know, there's tradeoffs of privacy, maybe, you know, of less, data, there's certain things work better, but the basics of sort of anti-surveillance law,
Starting point is 01:03:22 I mean, even national security was really into this stuff. They're like, it's too easy to spy on everybody. And, you know, that's a problem for us as a national security issue. And we just could not get a vote on even the most basic anti-surveillance, which would suggest, like, if you download a dog-walking app, it shouldn't be just tracking you and uploading every kind of information about you, that that should be illegal. I have been very I have been disturbed. We've not been able to do more on surveillance and privacy.
Starting point is 01:03:51 And I've also been struck by how badly what has been done elsewhere seems to have worked out. I find, I call this terms and conditions capitalism where you just move the burden onto the consumers. So Europe has put out some very sweeping rules that have given me the opportunity to individually decide which of the 303 cookies on every website I visit might be good and might be bad. Similarly, nobody's ever, in my view, to a first approximation,
Starting point is 01:04:18 read an iOS terms and conditions update. And I have found that very often it seems to me where policymakers end up after the debate is saying, well, as long as there is a disclosure, then the consumer can decide. But the consumer, in a very rational way, does not want to decide. So it has ended up, I think, in a very dispiriting place, instead of creating a structure in which I'm confident what companies are doing is well-bounded. It is demanded of me a level of
Starting point is 01:04:50 cognitive work I'm not willing to do, and I think nobody else is willing to do, to oversee those companies myself with not really great options if I don't like what they're doing. And so I'm curious how you think about that. No, I couldn't agree more. I feel like if the
Starting point is 01:05:06 byproduct of government action is that you are clicking on more little window, like that is government failure. And I would trace it to, frankly, a lack of courage on the part of government and the regulators or the officials or, you know, to make decisions that are really supposed to help people. It's much easier to say, well, you know, I'm afraid to do something. So I'm going to help them decide. So I agree.
Starting point is 01:05:35 I think the GDPR has actually failed to prevent surveillance. That being the European bill that created all those pop-ups. Yeah. GDPR, the European privacy law, succeeded, you know, in creating a lot of pop-ups and things to mess with, succeeded in making it harder to challenge big tech companies in Europe because they're over-regulated and the little guys have to also go through all this stuff. And so, yes, I think this has been a failure. I think for people to start to believe in government again, it has to help us in situations where we are not strong enough to deal with something much more powerful or something that has a lot more time to think
Starting point is 01:06:12 about it. I mean, it's like we're playing poker against experts. You know, at some point, we need to get backbone and have government on people's side. Now I'm starting to sound like a politician. But I mean it. Like, people say that, but really doing it makes making, you know, helping people when they are powerless or distracted or don't have energy to deal with things. Corey? So, look, I love you both, but I think you're dead wrong about the GDPR. I just, is a factual matter about where it comes from, what it permits, what it prohibits, and why it failed. Because I agree it failed. So you may ask yourself, how is it that GDPR compliance consists of a bunch of cookie compliance dialogues? And the answer to that is that European
Starting point is 01:06:51 federalism allows tax havens to function within the federation. One of the most notorious of those is Ireland. And almost every American tech company, except for Amazon, pretends that it's Irish, so that its profits can float in a state of untaxable grace in the Irish. and because of the nature of the GDPR, enforcement for these bullshit cookie pop-ups, which are the progeny of the big American tech companies, starts in Dublin with the Irish Data Commissioner, who, to a first approximation, does nothing. That sounds bad, but I want to get you to explain the core mechanism you're describing here better
Starting point is 01:07:26 because I actually don't know it because that bill did pass, and then all of a sudden the entire internet filled with these pop-ups. So that's only because the companies went to Ireland, broke the law and said, we're not breaking the law. And if you disagree, you have to ask the Irish Data Commissioner to enforce against us. But a few people, Johnny Ryan with the Irish Civil Liberties Association, Max Schrems with NOYB, this none of your business, this nonprofit, European nonprofit, they've dragged some of those cases to Germany. More importantly, they've got the European Commission to start modifying the way the law works. So you can just, you can tick a box in your
Starting point is 01:08:01 browser preferences, and it can come turned on by default that says, I don't want to be spied on. they're not allowed to ask you. I mean, the answer's just going to be no. And so I think that corporations want you to think that is transcendentally hard to write a good law that bans companies from collecting data on you. And what they mean is it's transcendently hard to police monopolies once they've attained monopoly status because they are more powerful than governments. And if that's their message, then a lot of us would be like, well, we need to do something. We need to turn the cartel into a rabble again, as opposed to, God, I guess governments just have no role in solving this problem. The one place where I do disagree with you having covered a lot of different,
Starting point is 01:08:44 both cartels and rabbles, lobbying Congress, it's not easy to regulate the Association of Community Banks, for instance. When you have something where there are in every single district, like individual leaders of the district who will come and lobby their member of Congress, it's really hard. I'm not saying that monopolies are good because they make it easier to regularly. I'm just saying that it doesn't solve the problem of the government runs on money and influence. Sure, but can we agree on necessary but insufficient? Yeah, so we can we can do that. I want to ask, but I want to build on this and ask Tim about a separate but related question. Tim, you mentioned a second ago, sort of the entertainment industry. And one of the questions about to come up is whether
Starting point is 01:09:27 Netflix should be able to buy all of the assets of, or all the entertainment assets, I should say, of Time Warner. And this is one where I think people who care about the quality of the media we consume seem, for reasons it seem compelling to be very, very worried about having that happen. How would you think about that? And is this a place where we need to be, say, making values judgments that are different than our antitrust judgments? Is this a place where the antitrust laws can suffice. Is everybody just worried about something they don't need to be worried about? How do you see it? Yeah. No, I think this is a place where if the antitrust laws are enforced correctly and fairly, that the acquisition would be blocked. And I'd say that this is not a particularly exotic
Starting point is 01:10:14 situation in the sense that you have the number one premium streaming company wanting to buy the number three or number four. And if you do the numbers under the guidelines, which the government issues that tell people when their mergers are presumptively illegal. The result is that this is a presumptively illegal merger. The reason I do think it's bad is I think that Netflix and Time Warner have frankly, over their history been some of the most innovative, interesting outlets. And often in an oppositional role, you know, this goes way back, but like, you know, Time Warner took a chance on sound film back in the 20s. In the 50s, they took a chance on television, which people thought was, you know, and then prestige television, early thousands with HBO and the Golden Age. So they've taken a lot of
Starting point is 01:11:02 bets. Netflix has done a lot of innovative stuff, really interesting, obviously. And frankly, you want to talk about good tech over the last 20 years. How about, you know, not having to wait until your show comes on? That's a form of efficiency I can agree with. And I think it would be a tragedy to have these two companies who are often so oppositional combined into one. I think culturally it would be a great mushification. At the economic level, just to continue on this, I think it's usually going to be those two companies who are bidding for the most interesting shows. So if you had a new version of a white lotus or something or the wire, who are going to be bidding for it, it's going to be Netflix and Warner, HBO Netflix or others. So, you know, the elimination
Starting point is 01:11:46 of one bidder is just the definition of a loss of useful competition. So yeah, I think it's pretty straightforwardly illegal. I don't think it's that complicated. Cori, you looked like you wanted to jump in on that. No, I think that one of the things we should probably anticipate Time Warner saying in defense of this merger is the same thing that Simon & Schuster and Pengman Random House said in defense of their failed merger that was blocked under the Biden administration. They said, oh, we'll still internally bid against one another within our divisions for the most premium material and that will be exposed to discipline that way. And I love what Stephen King had to say about this when he testified. He said, that's like me and my wife promising to both
Starting point is 01:12:27 bid against each other on the next house we move into. Tim, one thing I was thinking about while I was reading your book was the metaphor of a gardener. The way to think about economic regulation and antitrust and a bunch of the different buckets of solutions we're talking about is like a gardener who is trying to prune certain species and plants from taking over their garden. And the gardener has to make judgments. And, you know, there are some decisions you make as a gardener where you don't want blight getting all over your garden and killing everything. But others are made for aesthetic reasons and others are made because you want to have native species and non-invasive species. And there are all these sort of decisions being made. And having been around conversations of economic regulation
Starting point is 01:13:06 and tech regulation for a long time, I've come to this view that there is a fetish in them for truly neutral rules, that what people always seem to be looking for is a rule that you don't have to apply any judgment on. You can just say, if you get over the sign, everybody knows it's bad. as opposed to actually having to say, we have views about how the economy should work. We have views about how our society should work. We want the interests of small businesses to prosper, and they'll prosper more.
Starting point is 01:13:34 They don't have to give 30 cents of every dollar to Apple or Google, or, you know, if you're selling on the Facebook marketplace, Facebook. And yet, I mean, you've been a policymaker, Tim. I think that there has been in a kind of a, like a defensive crouch, particularly among Democrats. And, you know, Lena Con and others were an exception to this. but a sort of effort to describe everything neutrally
Starting point is 01:13:54 when sometimes you just don't want to be neutral on how fundamental companies and markets in your economy are working. You want to be able to have values that those serve as opposed to your values are subservient to your economy. Yes, I know, I agree with that, and I think it's an astute observation. I think it kind of comes, as I said earlier, from the lack of courage or vision that it reminds me
Starting point is 01:14:19 you said when you were talking about, well, okay, we'll just create a bunch of windows and let everybody decide what options they want for their privacy and hope that works. You know, because it comes from that same impulse that we don't actually want to arrive at a vision of the good society. It's one of the flaws of classic liberalism, frankly, if you get into the political theory. And frankly, the Gardner metaphor is targeted that it's not just like let it all run and see what happens. It is one where you have some idea of what you're what kind of world we want to live in and what kind of society we think is good,
Starting point is 01:14:57 and you have to make decisions based on that. I think we need a vision of what we want and what a good country looks like and a good place to live. The thing we really want to be asking before we ask any of these other questions is, how often are you going to have to answer this question? So lots of people are like, oh, we should just ban hate speech and harassment on platforms. Well, that's hard, not because we shouldn't do it, but because agreeing what hate speech is, agreeing whether a given, an act is hate speech, agreeing whether the platform took sufficient technical countermeasures
Starting point is 01:15:26 to prevent it, is the kind of thing you might spend five years on, and hate speech happens a hundred times a minute on platforms. Meanwhile, if we said, we're going to have a bright line rule that platforms must allow people to leave but continue to communicate with the people they want to hear from, then people who are subjected to hate speech who are currently there, because the only thing worse than being a member of a disfavored and abused minority is being a member or disfavored abused minority who is isolated from your community, those people could leave and go somewhere else. And it's not that we shouldn't continue to work on hate speech in parallel.
Starting point is 01:15:56 But if you think that a rule that takes three years to answer a question is going to solve a problem that happens 100 times a second, you're implicitly committing to full employment for every lawyer in the world to just answer this question. One thing I admire about both of your books is that you spend a lot of time on solutions. And so I don't think we can go through everyone, But let me do it this way. For each of you, and Corey, why don't we start with you?
Starting point is 01:16:22 If you were king for a day, what are the three areas or the three policies you can define it the way you want that you think would make the most difference? Sure. One would be getting rid of this anti-circumvention law in America at Section 1201 of the Digital Law Name Copyright Act and saying that it should be legal to modify things you own to do things that are legal and that it shouldn't be the purview of the manufacturer to stop you for. from doing it. Another one would be to create a muscular federal privacy right with a private right of action so that impact litigators like the Electronic Frontier Foundation as well as aggrieved individuals could bring cases when their privacy laws were violated. And I guess the third would be an interoperability mandate specifically for social media. So it would be a rule, and we've had versions of this, the Access Act, was introduced, I think, three times, various versions.
Starting point is 01:17:14 They're all pretty good. Mark Warner, I think, was the main senator behind them. But a thing that just says that you should be able to leave a social media network and go to another one and continue to receive the messages people sent to you and reply to them the same way you can leave one phone carrier and go to the other. And there's a lot of technical details about what that standard looks like and how you avoid embedding, you know, parochial interests of incumbents and so on. I don't think they're insurmountable. And I think that the tradeoffs are more than worth it. Tim? So I'll say three things. So first, I think we need the confidence to ban the worst and most toxic business models.
Starting point is 01:17:47 that are out there, you know, whether it's exploitation of children, whether, frankly, it's some of this total absolute price discrimination you're talking about, which may technically already be illegal. Number two, I think that it's unquestioning that the platforms have become essential. Commerce, the main tech platforms. I'm not in any ways thinking you can do without them. And so I think we need to understand which of them need to be treated more like utilities and which of them need to be not allowed to discriminate in favor of themselves or as between customers to try to maximize their extraction.
Starting point is 01:18:21 Can I hold you on that one for a minute? Yeah, sure. I always, when I hear this, it makes sense to me. Yes. And then I think to myself, do the people I know who focus on how utilities act and are regulated seem happy with the situation? And the answer is, no, they all think it's a total disaster.
Starting point is 01:18:37 So when you say they should be treated as utilities, but you know, you worked in the Biden administration, you know everybody who works on, say, green energy will tell you that the models and regulatory structures of the utilities is like a huge, huge, huge, huge problem. What specifically do you mean? It's a good question, and I've spent a lot of my life exposed to that. But I think what's important about utility regulation is what it doesn't allow to happen. Like the electric networks, the electric utility regulators are not perfect. On the other hand, if you think about the electric and network, it has been an extraordinary foundation for people to build stuff on. And the reason
Starting point is 01:19:14 they're able to build on it is they don't think the electric network is going to take like half their profits if you invent the computer on top of it. Or they don't think that, for example, the electric network is going to decide that, you know, it likes Samsung toasters instead of like LGE, I don't know, whoever else is toaster, Zenith, Zinth, something like that. So they don't discriminate between manufacturers on the electric network. And so I think we need to understand. and look carefully at which part of the platforms are the most like the electric network or the broadband network where there are essential to the rest of business and therefore need to play by different rules. And some of those main rules, the most obvious are duties of treating everybody
Starting point is 01:19:54 the same so they don't play favorites. And then if you've got it figured out, you get to the question of price regulation. Maybe Amazon's margin be capped at 30% or something like that. And then number three for you? Number three, I think we need a constant, you know, I'm an anti-monopoly kind of guy, constant pressure on the main tech platforms so that they stay, I guess, insecure in their position and aren't able to easily counter new forms of competition. I think you have to take out of the picture the easiest ways of tamping down or eliminating challenges to your monopoly.
Starting point is 01:20:35 I think that's been a really important thing in U.S. tech since AT&T, since IBM, since Microsoft, of keeping the main dominant market players insecure and force them to succeed to improve themselves as opposed to buying off the competitors or excluding them. So that's my third. So before we wrap here, I want to return to something we've sort of been circling, which is what kind of competition do we want to be encouraging among these platforms? Tim, one thing you said earlier was there can be this difference between healthy competition and toxic competition, which if you read a lot of economic commentary from the early 20th century, you hear a lot about that, and I feel like we don't talk about it that much anymore. But this is a place where I've been skeptical of the argument that many problems would be solved by breaking up the big, particularly attentional social media and algorithmic media giants that I don't think Instagram has gotten better under pressure from TikTok. I don't think that we're more ferocious innovation and entrepreneurial work to capture my attention or my children's attention is necessarily good. Maybe the problem isn't that we're not unleashing competition. Maybe the problem is that the entire thing that the companies are trying to do, whether there are two of them or 50 of them, is negative. Yeah, it's a really good point and a good question. I think in the
Starting point is 01:21:58 markets you're talking about, we have a serious failure to wall off, discourage, ban, or ethically consider wrongful the most toxic ways of making money. So there is such a thing as healthy, intentional competition, like making a great movie that keeps the audience enraptured for two hours. You know, producing a great podcast. That is good intentional competition. And frankly, the intentional market, you know, includes all these forms. But we have just allowed the flourishing of negative models. So I think if you had a world in which you had much more limits on what counted and what was frankly legal in terms of manipulating your devices, you would see more positive competition if you broke up some of these companies. I just think the entire marketplace
Starting point is 01:22:49 of social media is cursed by the fact that we haven't gotten rid of the most brutal, toxic, and damaging business models for our country and for our children and for individuals. I think that is a nice place to end. So always our final question. What are the books you'd recommend to the audience? And Tim, why don't we begin with you? Sure, I'd start with E.F. Schumacher's Small is Beautiful. Economics as if people mattered.
Starting point is 01:23:14 And I say that because it targets this question of what kind of world do we want to live in. And, you know, I think our efficiency obsession is taking us in one direction. And I think we should choose a different direction. A second book is more recent. Cass Sunstein wrote a book on manipulation that is underrated and is really good for understanding what we have allowed to happen. It's called manipulation, what it is, how to stop it.
Starting point is 01:23:41 The last book, I guess this is where I got some ideas about tech platforms and the big picture is from Paul Kennedy, Rise and Fall of the Great Powers. I feel everything is on a cycle and every empire has its destiny, golden age, its decline, its stagnation, and fall. And I feel like understanding imperial dynamics is very important to understanding the technological empires of our time. Corey? So my first pick is Sarah Wynne Williams' book, Careless People. And it's a great example
Starting point is 01:24:12 of the Streisand effect that when a company tries to suppress something, it brings it interest. So Wynne Williams, she was a minor diplomat in the New Zealand diplomatic corps. She became quite interested in how Facebook could be a player geopolitically. She started to sort of nudge them to give her a job as like an international governmental relations person. No one was very interested in it, but she just sort of kept at it until she got her dream job. And then the dream turned into a nightmare. My second choice is a book by Bridget Reed. It's a book called Little Bosses Everywhere. And it's an argument that the American Pyramid scheme is kind of the, it's the center of our current rot. And everywhere you look in the MAGA movie,
Starting point is 01:24:52 you find people who have been predated upon by the kinds of scams that are characteristic of this and who've adopted the kind of toxic positivity that comes with it. It is an incredibly illuminating, beautifully researched book. And then the final book is a kid's book by my favorite kid's book author ever, this guy called Daniel Pinkwater. And last year he had a book out from Takion Press called Jules, Penny, and the Rooster. And recapping the plot of this book would take 10 minutes because it is so gonzo and weird. But suffice it to say, it revolves around a young woman and a talking prize dog who find a haunted woods nearby where the young woman is welcomed by a sort of Beauty and the Beast story as a kind of savior, but who wants no part of it.
Starting point is 01:25:41 It's funny. It's madcap. It's gonzo. It's full of heart. It is like everything great about a kid's book. I read my daughter so many Daniel Pinkwater books when she was little. They're so fun to read at bedtime. It's a middle grade's book, and I cannot recommend it highly enough.
Starting point is 01:25:56 Jules, Penny, and the Rooster by the Incredible Daniel Pinkwater. Corey Doctor O and Tim Wu. Thank you very much. Thank you. Thanks, Desra. This episode of Isler Clancho is produced by Annie Galvin, fact-checking by Will Paisal. Our senior audio engineer is Jeff Gelb,
Starting point is 01:26:22 with additional mixing by Ammon Soda. Our executive producer is Claire Gordon. The show's production team also includes Marie Cassione, Marina King, Jack McCordick, Kristen Lynn, Emma Kellebeck, Roland Hu, Michelle Harris, and Jan Kobel. Original music by Dan Powell and Pat McCusker. Audience Strategy by Christina Samaluski and Shannon Busta.
Starting point is 01:26:46 The director of New York Times' opinion audio is Annie Rose Strasser. Special thanks to Natasha Scott.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.