Tech Won't Save Us - The Threat of Data Colonialism w/ Ulises A. Mejias & Nick Couldry

Episode Date: October 24, 2024

Paris Marx is joined by Ulises A. Mejias and Nick Couldry to discuss how Silicon Valley's extractive data collection regime and the power it grants them resembles a much older form of exploitatio...n: colonialism.Ulises A. Mejias is a professor of Communication Studies at SUNY Oswego and Nick Couldry is a professor of Media, Communications and Social Theory at the London School of Economics. They are the co-authors of Data Grab: The New Colonialism of Big Tech and How to Fight Back and among the co-founders of the network Tierra Común.Tech Won’t Save Us offers a critical perspective on tech, its worldview, and wider society with the goal of inspiring people to demand better tech and a better world. Support the show on Patreon.The podcast is made in partnership with The Nation. Production is by Eric Wickham. Transcripts are by Brigitte Pawliw-Fry.Also mentioned in this episode:Read an excerpt of Ulises and Nick’s book.Ulises has helped advance the Non-Aligned Technologies Movement.The World Economic Forum and Accenture published a report on governance of AI.Geoffrey Hinton was one of the winners of the Nobel Prize for Physics. Paris wrote about why we shouldn’t trust his assessment of AI.Google told the UK Labour government it will be left behind in the AI race if it doesn’t do what the company demands.Data centers use 21% of electricity in Ireland, and number that could jump to 31% within the next three years.Home building in West London could be restricted until 2035 because data centers have used up the available energy.Kenya is being drafted into the US’s anti-China tech alliance, which includes building data centers while ignoring the poor working conditions of data labelers and content moderators.Support the show

Transcript
Discussion (0)
Starting point is 00:00:00 What's happening with data and AI is the seizing of the assets of the planet, but in a very particular form, which is the seizing of an asset that hadn't fully been captured before, which is human life. Hello and welcome to Tech Won't Save Us, made in partnership with The Nation magazine. I'm your host, Paris Marks, and this week my guests are Ulysses Mejias and Nick Coultery. Ulysses is a professor of communication studies at SUNY Oswego, and Nick is a professor of media, communications, and social theory at the London School of Economics. They're both the authors of Data Grab, The New Colonialism of Big Tech and How to Fight Back, and also of Costs of Connection. I had the great pleasure of meeting Ulysses and Nick back in May when we were both speaking at Republika in Berlin. And at the time, I had actually already read their
Starting point is 00:00:58 book, Data Grab. So, you know, it wasn't something that just came on my radar at that moment. And it was really great to meet them because I really enjoyed the book personally. I think that we see a lot of comparisons about particular historical frameworks that what's happening with the tech industry fit into. But when I read their explanation about data colonialism and how they see what is happening today as this extension of this colonial relationship, that really resonated with me, particularly the way that they described it throughout the book. And so I knew I wanted to have them on the show. It was just finding the right time with, you know, their schedules and my schedules in order to do it. So I was thrilled that we could finally set this up and discuss the book, discuss data colonialism, what it means and the wider implications of this concept for both how we think about technology, but also, you know, how we use it as users and what this way of developing digital technology and digital
Starting point is 00:01:51 platforms means for human society itself. So in this conversation, we explore many angles of that, right, from what it means on the platforms itself, how these platforms are constructed in this particular way to the more infrastructural angle of it as these things kind of actually become physical, are part of the land and are really transforming parts of our communities and the world itself to further entrench this very harmful and exploitative system that Ulysses and Nick are describing in this book. So as I said, I was really happy to have them both on the show. I think that you're really going to enjoy this conversation. I certainly enjoyed being able to dig into the book with them and ask them more questions
Starting point is 00:02:29 about data colonialism and its impacts. If you enjoy this conversation, make sure to leave a five-star review on your podcast platform of choice. You can also share the show on social media with any friends or colleagues who you think would learn from it. And if you do want to support the work that goes into making Tech Won't Save Us every single week, you know, from making projects like Data Vampires to having really in-depth conversations like this one you're about to listen to, you can join supporters like Michael in St. Louis and Sam in Mexico City by going to patreon.com slash techwontsaveus and becoming a supporter as well. Thanks so much and enjoy this week's conversation.
Starting point is 00:02:58 Ulysses, welcome to Tech Won't Save Us. Thank you very much, Paris. Happy to be here. And Nick, really happy to have you on the show. Yeah, it's great to be here. Thanks a lot, Paris. Awesome. So you both had this great book, Data Grab, which gets into this concept of data colonialism that I wanted to explore with you both. And so to get us started, I was wondering, how would you explain data colonialism to someone who has never encountered this concept before? Well, I think where we'd start is to say, look, we all agree a lot is happening with big data, big tech, and now AI. Something big is going on, and it's about taking data for sure.
Starting point is 00:03:36 Most people explain that through the framework of capitalism, and obviously it is part of capitalism, but is it something deeper? And in thinking about this, we go back to ask the question, what is the absolute core of colonialism? What is that thing without which there would have been no colonialism? There would have been no point in doing colonialism. And that is the land grab. The violence was terrible. The racism is terrible. But without the land grab providing the point for those, they wouldn't have happened. It might have happened anyway for other reasons, but they wouldn't have happened in exactly that way. They were the means to get the land, the land grab. And what we're saying is that not necessarily through the same means at all. We'll certainly
Starting point is 00:04:20 come back to that. There are big differences here. What's happening with data and AI is the seizing of the assets of the planet, but in a very particular form, which is the seizing of an asset that hadn't fully been captured before, which is human life, which for the past 20 years, it's been possible since we live a lot of our lives online to seize in the form of tracing what we do and then trapping that data and exploiting it. That's the core of what we're saying. And that's why we believe it's a really convincing analogy to the core of colonialism. Yeah. And I think, you know, since we're talking about our book, I do want to give our listeners sort of like the official definition that they will find in the book of data colonialism. Nick and I argue that data colonialism is an emerging social order, specifically for the appropriation of human life, as Nick was just saying, through this medium of
Starting point is 00:05:18 digital data, so that data can be continuously extracted for two reasons, mostly for the generation of wealth, for the generation of profit, and also for social control. So that's how we define data colonialism. And we're not against data. We're not saying data is bad. Obviously, we all generate it. We all rely on it for different things like checking the weather, you know, trying to figure out if it's going to rain today. So we're specifically talking about this use of data, this extraction, this appropriation of data for these specific purposes, generation of wealth, which also generates inequality and also for social control. And, you know, we do realize that using the word colonialism can
Starting point is 00:06:06 be tricky. Obviously, it's a word with a long history. Some people might, in fact, may perhaps even resent our use of it in the sense that we might be equating whatever's happening with historical processes that, as we know, were brutal, consisted in many cases of genocide, of extraction of resources. We are very much aware of these differences. In our work, we do acknowledge that historic colonialism is very different from data colonialism as we see it. There are differences in terms of the modes, the scales, the intensity. Colonialism in India was very different than colonialism in my native Mexico, very different than the kinds of colonialism we might be seeing today in Palestine. But we do argue that there is one crucial similarity across all of these different
Starting point is 00:06:59 forms of colonialism, and that is the function. The function is the same, and that function is to extract, to dispossess. So whereas before we were talking about territory and extraction and the appropriation of territory, today we're talking about the extraction and the appropriation of data. The modes and intensities might be different, but that function is still the same. Yeah, I appreciate you both outlining that. And, you know, like I've told you both, I read a lot of books about tech and a lot of books that make different comparisons. But I found that your comparison in reframing this around colonialism to be a really compelling one, right, for what we are seeing today and what we are experiencing
Starting point is 00:07:40 with this tech industry and everything that it is doing and the model that it has developed over the past number of decades in order to, you know, become so commercially successful, right? And so I wanted to dig down into this a bit more with you, right? Because you talk about that land grab and this progression at this point to wanting to seize the understanding of human life or, you know, whatever it is about human life, because we know that capitalism always has to commercialize more and more aspects of our life to continue growing and what have you, right? So you talked about this comparison to colonialism, how the land grab aspect of it and wanting to seize control of, you know, these aspects of human society and human existence is something that it has in common, but the violence is something that is quite different, right? And that we don't see so much with the expansion of these platforms and
Starting point is 00:08:29 with this model of data gathering and data control. So what do you see as the main similarities between, you know, say colonialism of the past and what you're describing here with data colonialism, but also the biggest differences that you see between the two and where it doesn't fit so well? Well, you're right to point out that violence is a big issue here, because we do mostly associate, when we think of colonialism, just brutal violence of that system. And so we want to be careful. We don't want to say there's a one-to-one correlation here. We're not trying to pretend that brutality of slavery, for instance, that happened during colonialism, we are seeing that in the exact same
Starting point is 00:09:11 form today. But we also don't want to pretend that violence is not part of this system because it is in different forms. So for instance, when we think about about algorithmic bias that we encounter, the way in which powerful algorithms and artificial intelligence has been found to discriminate against certain peoples, of course, then we continue to see a different kind of violence, again, propagated through the system. So it's important to keep in mind that there are important differences, but there are also certain continuities. And our work is about sort of excavating and exploring those continuities. The question of violence is a really fundamental one.
Starting point is 00:09:56 So I'm glad you brought that up early because we've talked about this to many, many audiences over the past five or six years. And it is one that for many people, they can't get over. They may have themselves experienced that violence in their family recently, or their whole history of their family is shaped by that violence. And of course, we take that seriously. That's one reason why we are so engaged in the history of colonialism. So why would there be a fundamental difference? Well, if we put it in simple terms, when the first colonization happened and through the early history of it, people were leaving European lands and finding human beings and territories. Sometimes they didn't even know
Starting point is 00:10:37 they existed. They certainly didn't know anything about those people and they probably didn't care. They had no social relations with those people. If you're trying to take something from someone you have no social relations with, there are only two options, brutal violence and lying. And they use both. Now, lying might still be going on in a subtle form. The brutal violence isn't quite so necessary. And there's a very interesting and fundamental reason for that. And it links back to this link between capitalism and colonialism. What's going on with data and AI today is to do with both capitalism and colonialism, the two together. In the book, we call it the double helix. You could never have had capitalism without the huge grabbing of resource from across the planet
Starting point is 00:11:27 that was colonialism. But where does that leave us today, two and a half centuries on? It means that we've spent two and a half centuries as human beings across the planet getting used to what capitalism demands of us, which is we sign terms and conditions. We know if we want that product, we've just got to accept the terms. There's no negotiation about that. That's a lot of violence being used in indirect means, because we're so used to capitalism. Just to give you a simple example of how there really is still something like violence going on, but it takes a radically different form. I was really shocked to find that marketers in the health data sector are tracking you just as much as the marketers in the market for sweets or trainers. Perhaps I don't care too much what they know about my body,
Starting point is 00:12:18 but I really do care that they know about my search data in relation to my life expectancy because that's of huge use to insurers. And this data is being traded. Those marketers are doing exactly the same thing. That's profoundly offensive and to many people, truly violent. And it's going on as if it was totally normal. So there are many forms of disguised violence, but the explicit violence just isn't necessary anymore because we're already plugged in to the system, which is a capitalist system, which started with colonialism. I appreciate you both outlining that. And, you know, I think bringing up that point of violence
Starting point is 00:12:55 is really important, right? As you were both talking about the examples that you were sharing, I was thinking also about things like predictive policing, right? And how we can very much see the use of this data in ways that are very harmful to people. And, you know, that's just one small example. There's so many more. But you talked about the history of colonialism there too. And before we continue to probe this concept of data colonialism, I wanted to return to that as well, because one of the points that you made in the book, which I found really compelling as well, was that, you know, when we think about colonialism, there was an aspect of Western technology that was very much connected to this colonial project, right, and enabling its growth around the world, but also, you know, the expansion of this capitalist
Starting point is 00:13:34 system that was driving so much of this colonialism as well. I wonder if you could expand on that a bit too, because I feel like, you know, when we think about what is happening today with digital technologies and whatnot, it can feel very disconnected from this larger discussion and this larger trajectory that we've been on for several centuries now. Yes, you're absolutely right, Paris. I think, as I mentioned earlier, in our work, we try to trace these continuities, you know, to use the past to help explain what's happening in the present. In order to do that, of course, we need to look at the history of colonialism to trace discontinuities.
Starting point is 00:14:10 You know, since we were just talking about violence, that I think it's important to keep in mind in terms of discontinuities is that a lot of what we're seeing today in terms of data colonialism continues to impact the traditional victims of colonialism disproportionately than the rest of us. So what do I mean by that? People of color, black people, women, and of course, the poor, you know, those three categories, I think, are essential that during colonialism, certain frameworks for rationalizing the oppression of these groups were put in place that continue up to today.
Starting point is 00:14:48 And that data colonialism, of course, also continues. We see this clearly not just in terms of algorithmic bias, but as Nick was just saying, if we start to think about the effects of health in terms of what's happening through the system of data colonialism, how it might impact the health of some people more than others. So yes, some of us might benefit in this system, but we do have to keep in mind that the traditional victims of colonialism continue to pay a heavier price. And you're right, the history of Western technology,
Starting point is 00:15:21 the history of Western science is very much a part of this discussion, because we do have to remember that Western science and Western technologies very much developed hand in hand, let's say, with the history of colonialism, which is not to say all of science is suspect. And we're not trying to argue that Western science is bad. Obviously, there are many benefits to it. There are many discoveries that have benefited humanity. So the point is not that. The point, however, is to acknowledge that a lot of Western science and technology was developed as a means to manage the colonized world.
Starting point is 00:16:01 The challenge, to put it in these terms, the managerial challenge was exactly that, how to manage colonies at a distance, when the center of empire was in Europe, Western Europe, and the colonies were in the rest of the world, what is now known as Latin America, or Africa, or Asia, etc. So this was a management problem. And so Western science and technology developed certain tools through the scientific method, through the application of science to technology, to develop tools of communication, to develop tools of transportation, to develop new tools in terms of military weaponry. All of it sort of geared towards making the management of these colonies possible.
Starting point is 00:16:49 And so I think that helps us reframe a lot of what we're seeing also in terms of these new technologies and artificial intelligence and algorithms, and also the power that some corporations are acquiring. Because let us also remember that colonialism was always a partnership between certain states and certain corporations. So it helps us sort of rethink a lot of these relationships and all of these new innovations in terms of these continuities. If I could just add a little footnote, continuing that to bring us maybe not right up to date, but 20 years in the past, rather than 500 years, what will be the equivalent today to these technologies of managing at a distance, as Ulysses said, which was the core of colonialism? I think it could be something as simple as the technology of Facebook. If you look at the early patents of Facebook, which I've been
Starting point is 00:17:43 doing recently, they literally lay out a technique in super technical language. And of course, using the whole architecture of patents, which is a key part of the scientific infrastructure to create a shadow world where you can't do anything without Facebook tracking it. It's literally impossible. There is nothing there except what Facebook has tracked. They had to build that world from nothing. So yes, this seems radically different from taking Brazil or taking a Caribbean island, because they pre-existed and they had their own utterly integrated history. Building through writing code and backing it in a patent, a territory where you can't do anything except that it's already captured is a very different thing.
Starting point is 00:18:28 But it has a broadly similar impact that you literally can't do anything there that what is managed. It is a way of managing you at a distance from a lab somewhere, even though you're the other side of the planet. And we feel very close, perhaps, to our social media accounts because it's our friends and our family on the other side. But in the middle is a mechanism for managing all of us at a distance. And you can see it laid out in something like the patent. And that's the creation of what we call in the book, a data territory, which is a brilliant concept to make the whole world treatable, just like like a territory because it's laid out right there in front of you and only you the engineer can see it the rest of us we're just inhabitants
Starting point is 00:19:11 of the territory i think that connects us really well to you know kind of the next stage of this right because we've talked a lot about this comparison to colonialism but now we want to get to the data side of this and i'll come back to data territory specifically in just a second i think what you were saying there nick really starts to get us close to this question that I want to ask. And that is, why is data so essential to the understanding of this concept, right? Obviously, we're talking about colonialism and this sort of relationship that comes with it. But why is data so important to both this analysis that you both are doing, but also this structure, this model that these companies have set up?
Starting point is 00:19:49 How does it give them so much power and so much control that makes this so concerning? That's a great question. And often people have said to us, well, surely what about digital? Digital colonialism, isn't that a better word? Problem is, when stuff became digital from being analog, nothing dramatic in itself happened. Yes, it was the start of something, but it wasn't the completion of something. We're all happy we now have digital data. If we didn't, we wouldn't be having this conversation now. So digital is not enough. Some people say, well, it's about
Starting point is 00:20:24 informational capitalism. Information has been flying around for the past three centuries of capitalism. It's not obvious there was such a fundamental break. But there was a fundamental break with the gathering of data about 30, 40 years ago, the beginnings of supply chains, the beginnings of credit card data. These were a radical change in the terms of business across the whole of capitalism. That's where we start to look for the origins. Why is it so important? Because data is a means to, first of all, track everything in a territory, and then secondly,
Starting point is 00:20:59 through tracking to control every possible parameter. And not just a few parameters, every imaginable parameter can be tracked and can be managed. So that creates a possible of much more integrated, multi-dimensional forms of management. Think of the Amazon warehouse floor. If there are still human workers there, and there are some, Every conceivable thing they do will be timed, measured, evaluated for its emotional input. If they have a sense of anger, it will be picked up instantly and tracked and compared with other forms of anger. This is what Marx imagined in terms of capitalist surveillance as management, but taken to a much higher dimension because there are no limits to the dimensions of data. That's why data brings about a fundamental change in the nature
Starting point is 00:21:51 of power relations and a fundamentally deeper and broader asymmetry of power relations between management who holds the data and the workers or the consumers who are held. I think it's important to sort of recognize that data is a form of abstraction. It abstracts our lives. It transforms them into something that can be quantified, that can be collected, that can be analyzed. So, you know, if we go back to Marx and his critique of capitalism, he was basically talking about how things that were outside of the market were brought inside of the market. And that's sort of the whole process of capitalism, how it functions, right? It commodifies things. And so when we looked at this modern phenomenon of digital data,
Starting point is 00:22:43 that's sort of the medium, the channel through which this happens in our lives. Because we live social lives, we interact with families, with friends, we go to work. None of that is necessarily part of a market. But what data did is to precisely transform all of these activities, all of this social richness, convert it into zeros and ones that could now be managed, that could now be analyzed and used even more recently to train large language models. So that's why data is very important, because without data, we wouldn't be able to abstract social life and bring it from outside of the market to inside of the market, where basically it can
Starting point is 00:23:31 be commodified and transformed into wealth. Hey, listeners, I'd like to take a quick moment to shout out another show I think you'd really enjoy. Cited is a podcast about the politics of science, research, and expertise. It's a reference to citation, you know, like an academic citation. Anyways, they have a new season that is definitely worth checking out. It's called The Use and Abuse of Economic Expertise. Economics sometimes feels like a physics, so sturdy, so objective, and so immutable. At least, that is the way that politicians tend to talk about economic research.
Starting point is 00:24:09 Yet behind every clean number or eye-popping graph, there is usually a rather messy story. A story shaped by values, interests, ideologies, and petty bureaucratic politics. In the use and abuse of economic expertise, Cited reveals the hidden stories behind the economic ideas that shape our world. You can find Cited wherever you find your podcasts, or at CitedPodcast.com. That's C-I-T-E-D Podcast.com. in particular, to talk about how this data is used in these ways that are particularly concerning when you think about it through this colonial lens. So I wonder if you could expand on these particular concepts and these particular uses of data and these particular ways of setting up these digital environments that we use in order to maximize the collection of data, but also the use of this data in these ways that you both find very concerning. So what is it about these two concepts that are so important to your broader understanding here? Well, a data territory is an abstract space created by writing code and software that is
Starting point is 00:25:16 so configured that data can be continuously extracted from it by whoever built and controls that territory. That sounds innocent, but it's a way of describing all the platforms we're on, all the apps, everything. And of course, there are good apps and there are good platforms. We're not saying in itself any data territory is intrinsically necessarily bad, but it creates the possibility of power. Because if you wrote the code, if you wrote the software, then you are in control of the tracking of everything. Because if you wrote the code, if you wrote the software, then you are in control of the tracking of everything that goes in there. And that means that you have the possibility of managing what people do in that territory, because you can send signals, you can nudge people, you can give them only five rather than 10 options. So there's no
Starting point is 00:26:01 option to dislike stuff on certain social media. There is always an option to like and so on. Simple things like that. Some of those decisions are good. Some of them are bad. But that's a data relation because we're basically in a modulated form of social interaction, which is very different from everyday life, where behind the scenes, there are certain menus which are limiting the sorts of things we can say. And that is absolutely not like our experience of social life normally, unless we're in a job interview, maybe. And even then, we sometimes take risks and we're not physically stopped from doing so. We may not get the job.
Starting point is 00:26:36 But in a software, there literally is nothing you can do outside the menu. So that is a highly constricted form of data relation, which is therefore the perfect means to impose a certain control form of power that doesn't exist in everyday life and never existed until we spend our life online. So that's why, and it links to the point about abstraction that Ulysses made, which is the other side of it, many things that would be part of our conversation simply can't be. And we're not in control of that not being there because they've been abstracted out. And that abstraction is normally there to benefit those in power. And of course, we're seeing this very much in the political situation at the moment. I heard someone just
Starting point is 00:27:20 yesterday describing two or three days ago at a street action, describing how everything they want to say about the humanitarian crisis in Palestine, they literally can't say on the social media platform of a certain sort. They literally cannot say it. through complex parameters, because the person said to me, I speak, but no one hears it, because it's not in people's feet, because the algorithm controls what it is, how I appear to you. These are radically different forms of social relating, therefore the different forms of power relations, different forms of data relation, and they create this territorial power, which has these profound wider impacts on the way whole areas of society are being structured today. And we'll come on to some examples in a minute, I'm sure.
Starting point is 00:28:08 Yeah, maybe just to exemplify some of these data territories, because of course, you know, we are all familiar with social media, with search engines, with using an AI chatbot. Those are apparent. But I think there are all sorts of data territories that are a little bit more hidden, sort of beneath the surface that we might not realize what is happening in these bounded spaces. So, of course, we have ag tech, agricultural technology. The production of food has become a very data intensified process where companies are using artificial intelligence to design thousands of variation of seeds that can grow in extreme environmental conditions because,
Starting point is 00:28:53 of course, corporations have seriously damaged our environment and now they're going to provide us the seeds that are able to survive in that destroyed environment. But we also have health tech that Nick was just mentioning as another data territory. We have ed tech, what's going on in schools in terms of all of the capturing of data, not just by schools, but schools are partnering with third parties who then conduct all of this appropriation of data
Starting point is 00:29:19 for their own benefits. We have all forms of bossware for tracking workers at home, in their offices, workplace, warehouses, tracking every single movement. We have a war tech as well, which we are seeing evidence of this across all of the war fronts that we see today in Ukraine or Gaza, Lebanon, using data basically to come up with targets of victims, in many cases, civilian victims. So each one of these functions as a data territory. There's a corporation or two, and probably a government or two behind them looking for new ways in which to use data to control that bounded space, again, for generating profit
Starting point is 00:30:06 or for controlling socially what goes on in that space. I think that makes perfect sense, right? And I find both of your descriptions there really compelling in thinking about the concerns that this presents, right? And I feel like on the one hand, you know, what you were describing, Nick, in relation to the limits that creating these platforms, these digital interfaces, you know, kind of bring into being, right? When things that we are used to doing in an old-fashioned traditional way then become replaced with some sort of digitized system that restricts what can actually go on here. I know
Starting point is 00:30:40 that speaking to people and reading about this for quite a while, that this is one of the concerns that comes with so much of the shifting of the interactions that we have, the way that we access services and all this, as it moves digital, it requires this kind of recrafting or reforming of how that service is actually delivered in ways that can really reduce people's ability to engage with it or, you know, to get the service that they actually need from it, because now all of a sudden your choices are limited to what works for however this system is designed. And then I feel like the broader piece of this, you know, when you're thinking about data, I feel like so much of the discussions that we have about how to respond to all this data and the power that comes from this data is always around, do we give people more rights to control
Starting point is 00:31:24 their data? Or are we going to pay people for their data or something like this? And I feel like never part of that discussion is, do we need so much of this data in the first place? Should we be creating all this data? But that doesn't really work for the broader conversation. So you don't hear it so much. Yeah. I mean, we see basically generative AI as a form of prospecting. We can talk about it sort of in terms of the tragedy of the commons, what happened in the early days of the internet. We were all led to believe that we were contributing to these wonderful commons where all of the data we were uploading, we were
Starting point is 00:31:57 sharing, would enrich humanity because we were all going to benefit from it equally. And then, of course, what happened is that these commons got enclosed. The boundaries were defined by corporations. And one day we woke up and we realized this wonderful community we were building online, now all of that data that we contributed belongs to a corporation. And they're putting real limits and applying this data in very specific ways, obviously, to create this wealth. By now, of course, with AI, we've gone from tragedy of the commons to perhaps a farce of the commons, because now the next step is now that we have enclosed the commons, let's take all of this data to create a machine that basically the whole point is that it's going to replace your job. It's going to replace even your creative outputs and the ability to make a living by making music, making art. Now we will need all of that because a large language model has been trained with all of the data that we have been producing for the last
Starting point is 00:33:05 couple of decades. And this machine is going to be able to replace human beings in the production of culture. So that's kind of a scary development in all of this. And if we just take that one stage further to, well, how is this even possible? Some of this even goes beyond our model of the restricted data territory in a way, because we're now finding out, as Ulysses was just saying, that even when we were making a song and it just happens to be online or putting our pictures up there or chatting to people or maybe just doing a little blog about something, all of that is fuel for AI. It can just be taken. And in fact, it needs to be taken. And we were very intrigued to see OpenAI when they gave testimony to the UK House of Lords. They actually came clean. And Sam Altman said in his written testimony, well, our business model depends on taking copyrighted material. Hang on a minute. Now, if you or I take copyrighted material, that's a crime. We're going to go to jail.
Starting point is 00:34:09 We're in big trouble. But when OpenAI with a smile take everything anyone has ever made, that's somehow not a crime. In fact, it's a necessity. And in fact, just today, by a weird coincidence, the World Economic Forum, which for more than a decade has been a cheerleader for everything data colonial, published a report called Governance in the Age of Generative AI, co-authored by someone from Accenture and an academic or someone from WEF. Well, they actually
Starting point is 00:34:41 cheerleaders for this very idea. They know that countries like Singapore, Estonia, and a few others have actually made exceptions to their own copyright law and say, no, it's totally fine to take everything you want, because that's either fair use, which is okay, always has been under copyright law, or we just make a big exception for generative AI in any event. And the words used, and this is an example of the sorts of, we call them in the book, civilizing stories that get told to sweeten the pill or repackage data colonialism, just as historical colonialism was repackaged. The idea is, well, that's necessary for innovation to advance science back to science again, or to create business value. And that must be all right, even though it breaks all the rules. As the New York Times is protesting in the courts at the moment, it breaks all the rules of capitalism, but somehow it's okay. So this is an example, and it's a very interesting one in the past two years, where in a sugared way, and open AI is, ChatGP is fun to use. It's a nice little game. Something, a land grab,
Starting point is 00:35:47 a data grab is going on, but it has to be disguised. And that's where you need the civilizational story that colonialism has been incredibly successful in generating for 500 years. We just have new versions of that same old story. We're taking it from you, but hang on, it's all going to be better. So just trust us. It's all going to be for the progress and betterment of humanity. And we were told that story with colonialism, and we're still being told that story today. In the book, we sort of talk about this narrative of the myth of cheap nature. So in order to colonize the world, of course, colonizers had to imagine the world,
Starting point is 00:36:26 the nature, territory as something cheap. It was abundant. It was just there for the taking. So colonizers could arrive in Australia, for instance, and say, you know, this is terra nullius, no man's land. It doesn't belong to anybody. We can take it. Of course, to transform that territory into wealth, you needed also cheap labor. So you go from cheap nature to cheap labor. And of course, colonialism was a racialized system. So the labor of black people, the labor of people of color, the labor of women was seen as, again, cheap, abundant, just there for the taking. We can just take that labor and use it to create more wealth. So let's follow that progression. We're going from cheap nature to cheap labor. And now,
Starting point is 00:37:11 of course, we have cheap data. That's the latest civilization or story. Cheap data has many of the same characteristics of cheap nature and cheap labor. It is, again, said to be abundant. It is free. And it is just there for the taking. Open AI can say, well, look at all of this data. It doesn't really belong to anybody. Yes, individuals produce it, but by the time it gets collected, we, companies, corporations, need to provide the advanced technology to refine it, just like oil refineries of the past, sugar refineries of the past during colonial times. So now corporations provide this advanced technology to take this refuse, this exhaust, and transform it into something useful. The story goes, the colonial narrative tells us,
Starting point is 00:38:00 again, this is happening, but the good news, the story says, it's for your benefit. It is for the betterment of humanity. Those narratives are so important, right? And the tech industry has shown itself to be so successful at creating them time and again, and perpetuating them through society. Thinking back to the things that you were saying, right, about the internet and this vision that we were sold of the internet as a commons and then how it was commercialized. But these ideas that came out of that era were still weaponized and used by these tech companies to get this kind of broader acceptance. And, you know, you've explained a lot. Both of you have explained a lot about the problems that this data colonialism presents us, right?
Starting point is 00:38:38 And the very real harms that come of, you know, this particular model that these tech companies have created. I think going beyond those civilizing narratives, what do you see as the way that these companies so successfully sell this to the public to not get this wide swath of the public saying, we're not okay with this, we don't want this to happen, and we want some other way of arranging society or the digital world or what have you? Well, I mean, we could take ChatGBT as an example. And it's a really interesting one because again, the news has been good today. We're speaking on the day when Geoffrey Hinton, one of the leading developers of neural nets, got the Nobel Peace Prize in physics. That, I have to admit, took my breath away. That was astonishing. Let's just step back for a moment. Yes, Geoffrey Hinton, without question, is one of the most thoughtful and skillful people in the area of data points, can be processed in a regularized,
Starting point is 00:39:47 automated way more effectively than it could be before. That could cure cancer. It could find the structure of proteins. No one's denied that that is positive in a good scientific context. But he's also the person who came to fame with the ImageNet data set, which he analyzed in 2012, which was actually the basis of the labels that human beings had attached. They'd been repointed through Amazon Mechanical Turk, which is another one of these human labor forms, which is forgotten in the history of artificial intelligence. There's a lot that's not artificial at all. And of course, it tends to be distributed in the global south, where people are more open to exploitation, or parts of the US where people can't get by and they need money through
Starting point is 00:40:35 Amazon Mechanical Turk. So his history is quite interesting. But then if you take ChatGPT in itself, there was a point in the development of neural nets when the most sophisticated people, the people working with Geoffrey Hinton, realized we've hit a brick wall because a machine is just a machine. It doesn't live in the world. It doesn't know really what a car is, even if we have hundreds of people to tell it. And the one thing it doesn't know is what is Nick and Ulysses Paris, what are they talking about right now? Why are they talking? How does this relate to what they just said?
Starting point is 00:41:09 It doesn't know context and human communication is based in context. So the great idea was to access context. How do you do that? Well, you create predictive machines that are sufficiently credible that human beings think it might make sense to interact with them by asking them a question and getting a reasonably credible answer, which is what we do with ChatGBT. And then if we like the answer, it's an answer. If we don't, then we feed that back to ChatGBT and it corrects the model. And next time it gives a slightly better prediction. Vin Cerf, who is one of the founders of the internet, came up with an incredible analogy for this. He called it a salad shooter. You just chop everything up into the maximum number of tiny bits, put it in a massive salad washer. It comes out as long as it's grammatically credible
Starting point is 00:42:00 as something like an answer to the question you just asked, that counts as the truth. So again, this is a weird form of serializing that makes something that is merely a mathematical prediction, okay, based in smart math, into a truth that rely upon sufficiently so that the person who devised it gets the Nobel Prize in physics. I mean, that is a question, right? Why don't people refuse? I actually think most people, once they understand some of the issues and what's involved, they do refuse. But as with the history of colonialism, you know, when people resisted colonialism, it rarely made the news until anti-colonial movements, movements of decolonization got big enough, important enough that people in the center
Starting point is 00:42:45 of empire had to pay attention. So, you know, in some ways that brings us to the point about resistance. I think people are resisting today. I think if you listen to your podcast, Paris, we get to hear a lot of those stories about resistance happening. Now, maybe it doesn't make it to mainstream news. Maybe it will in the future. But I do think, especially with generative AI, I think people are realizing some people are embracing it. I just look around my campus and I think it's a microcosm. Some people are embracing it.
Starting point is 00:43:18 Some people want to incorporate it in the classroom. And others, including students, are realizing, no, I'm not okay with this. In some ways, I think we need like a student bill of rights for the age of generative AI, because some students are saying, I won't use it in my assignments, but can you, the university, also ensure that you won't use it for grading? Professors won't use it for developing their lecture materials. Can the university ensure that you won't use my data that you collect when I visit the website or interact with the learning management systems? Can you ensure that you won't use my data to just feed an AI machine if I am agreeing that I won't use it to cheat on my assignments.
Starting point is 00:44:07 So I think we're coming to that point where more and more people will start to resist and to basically refuse some of these narratives that so far a lot of us have been believing in. Yeah. And I feel like even hearing what you're saying, I'm saying to myself, I need to do an episode on ed tech and the effect that this is having in education. But that's for another conversation. I want to talk to you a bit more about resistance. But first, I want to go back to one other point that you've both been making, right? You know, when we think about these generative AI models, and Nick, you were mentioning in particular, these governments around the world and these tech companies who are demanding that their collection,
Starting point is 00:44:45 their mass collection of data in order to feed these models and train these models be something that they're allowed to do, right? And it brought to mind this story in the UK recently, where Google basically said to the new labor government, you are going to be left behind in the generative AI race if you don't pass a bill that says that our taking and stealing of all this data to train our models is legitimate. But also there was another point to that where they said, and also if you're not building enough data centers to be able to properly compete in this. And, you know, we've been talking about data colonialism and data territories and things like that, but you mentioned how the original colonialism was very much linked to land, right? And grabbing land.
Starting point is 00:45:22 And I feel like there is a very physical aspect to what you're talking about as well, when we think about how these tech companies have been building out these major underwater fiber optic lines and increasingly trying to control those. And of course, the push that we're seeing now to build out these major data centers around the world. I'm wondering what you see as the infrastructural piece of data colonialism as well and that angle of it. I think in many ways, what we're trying to do in our work is to trace a line between the plantation, the factory, and the data center as three discrete moments in the history of colonialism and capitalism. And you're right, Paris, that it might seem like data colonialism is part of the cloud,
Starting point is 00:46:04 is something that happens in this ethereal space. But as you're pointing out, this is very much a conversation also about land and territory. And data centers make that incredibly apparent. So just like plantations were about land, appropriating land, and then developing monocultures so that countries would only be dedicated to producing sugar or minerals or whatever. And then factories also took the land and created these infrastructures for exploiting labor in new ways. We have data centers which occupy space, actual space in our universe, and use a lot of resources that are becoming more and more scarce. So as you have talked about in your podcast again and again, energy use, water use, the fact that we're building data centers in parts of the world where some of these resources are already scarce. And basically, we will need to take those resources away from people who need them to be able to feed these hyper data centers that are very hungry monsters that consume these
Starting point is 00:47:16 resources in the large scale. All of it while not paying taxes, all of it while being extremely secretive about what's happening in some of these areas, all of it while making special deals with governments that many times, most of the time are not, you know, advantageous to those government citizens. So I think that's plenty of reasons to start to, you know, to look at this also as a territorial issue. And, you know, the latest signs from, say, the UK government that you mentioned are not encouraging. Google already had a very close relationship with the British government through the NHS, along with Palantir. And Google was involved in various scandals through the use of NHS data,
Starting point is 00:48:01 which you probably heard about over in North America, and no doubt lay behind the UK government's 2023 pro-innovation, that was a phrase, regulation of tech, which is enthusiastically cited by the World Economic Forum just today, because it's exactly the direction they think everyone should be going in. But on the question of the data centres, one was recently opened or was proposed up in the northeast of England by the new Labour government. I drew breath at that point to think, hang on, haven't you listened to Paris Marx? Haven't you listened to the blogger? Don't you know the danger? Just to give you one homely example, I used to live in West London, which is a relatively prosperous part of London, but it needs houses for sure. And it's very ethnically mixed as close to Heathrow.
Starting point is 00:48:46 There will be no new houses being planned in that area until 2035. Why? Well, because the future supply of electricity, and these things are mapped out in big blocks by the big tech, has already been taken. Who by? Data centers, as reported by the Financial Times. Data centers. One third of Ireland's electricity is going to data centers. Now, you might say first world problem, but actually when that first world problem, and it's a problem, gets pushed down to the global south, which is even less chance of negotiating the terms of trade, you can be sure that when Kenya is told that for its future, it absolutely has to get the number of data centers
Starting point is 00:49:29 the World Bank thinks is consistent with a liberal market economy, there's going to be a lot of problems. In addition to the lack of luxury housing being bought, there's going to be many other costs for water, electricity, and so on. These are problems that will grow. It's not sustainable going in this direction, basically. And no one seems to be saying that, least of all governments. Yeah. The example that you gave in West London is such a shocking one. And you can give so many more examples of that for so many communities around the world, right? Which is where it really becomes just staggering. And I'm happy that you brought up
Starting point is 00:50:03 Kenya there because that was another thing I wanted to ask you both. And we've been talking generally about data colonialism, but do you see a distinction in how this plays out, say, in the global north and in the global south in the broader impacts of these things, you know, and the way that these models and these methods of, you know, data collection and using this data actually play out in different parts of the world? Well, I think there certainly are going to be differences, partly because of the legacy of historic colonialism, which we insist in our book has not gone away. Some people misread our book and say, well, we're saying data colonialism replaces historic colonialism. We're saying the exact opposite. Of course, historic colonialism has not gone away. We have a profoundly unequal
Starting point is 00:50:43 global economy. Racism disfigures every society on the planet linked to historical colonialism. And that means that the field in which data colonialism, the new toolkit of colonialism today will be tried out and then applied and then forced on people is a profoundly unequal. I'm not without hope. I think there's some very interesting things going on in Africa. The African Commission has a major consultation on AI legislation at the moment, which I hope will come up with something radical. Similar radical things have been happening in parts of Latin America. But we shouldn't deny the fact that these are operating against a very difficult situation. The supply of capital, the supply of development capital, the fact that these are operating against a very difficult situation. The supply of capital,
Starting point is 00:51:26 the supply of development capital, the fact that all the apps and the texts that are being used are already existing in the West or are in China, which in our book is the other pole of the new data colonialism. And that's another thing about today's world, which is radically more complex than the world of historical colonialism, when it was just Spain and Portugal initially, that means that it's going to be incredibly difficult to resist this tremendous push, which is so dominant, even in super rich countries like Canada, Ireland, Britain, Germany, and so on. Yes, I think, you know, obviously one thing that colonialism did was to divide the world into metropolises and colonies. And so since then, that distinction has been very real. Today, we speak mostly in terms of global north and global south. And it's true that each one has its own peculiarities and specific problems. So I think we need to acknowledge that, that data colonialism will look, will take very
Starting point is 00:52:30 different forms depending on the part of the world that we're talking about. Having said that, I do think that the other thing that colonialism did that we should keep in mind is that those boundaries also became very porous in the sense that as soon as colonialism started, we also have migration from the colonies to the center of the empires. And so the South became a part of the North and the North became a part of the South. And so we see that also replicated in data colonialism. Obviously, if we're talking about the gig economy in the global north, we can see how exploitation and extraction happens right in the middle south that basically replicate the same models, the platforms, the artificial intelligence models. They're copying them from the global north and applying them to the global south. So again, we have sort of miniature enclaves of the north
Starting point is 00:53:39 also in the south. So these relationships are very complex and very dynamic, and we can expect to continue to see those exchanges in data colonialism as well. Yeah, that makes perfect sense to me. And as a final question, you were talking a bit about resistance earlier, right? And how people might want to push back on this. And I feel like, you know, you also talked about how one pole of this data colonialism that you talk about is the United States. And I think, you know, this is the one that we would be most familiar with and the platforms that we most often interact with. But obviously, we see, you know, the Chinese model adopting a lot of the approaches data sovereignty or digital sovereignty becoming more common as more and more countries seem to want to get control of their digital infrastructures and the digital technologies that their societies use. And in the past, that seemed more focused on what China was doing and the way that it was carving itself out from the United States. But now it increasingly seems like something that is saying we reject what China is doing and what the United States is doing, and we're trying to form something else. So I wonder
Starting point is 00:54:48 if you see any hope in movements around the world or approaches around the world that are trying to push back against this Dana colonialism model, what you see is most hopeful there. For sure. This concept of sovereignty is very interesting because obviously when it means autonomy, when it means independence, I think that's definitely what we want. That's why some of us are sort of interested in this non-aligned technologies movement to replicate the original non-aligned movement, which some of your listeners might be able to identify an agreement between nations in the Cold War, who basically said, we don't want capitalism and we don't want communism. We need this third way. We need this autonomy and this independence.
Starting point is 00:55:29 Today, obviously, it's a little bit more complicated because Europe, for instance, wants to be sovereign because it doesn't want to depend on the United States or China too much. Of course, they're finding out it's impossible not to do that. But this notion, this definition of sovereignty is always problematic when it comes to the state, because it always includes certain people, but it also excludes others. Sovereignty, even in the time of independence, when colonies got their independence, was used precisely in this way to say who is in and who is out. No surprises, those who were usually out were indigenous people, people of color, people who were considered not good enough to be part of the nation state.
Starting point is 00:56:15 Today, when it comes to Europe, for instance, as it talks about data sovereignty, it is still using data to exclude certain people, migrants, for instance, or when it comes to European or companies from the global north developing tools of war, advanced and new tools of war. Obviously, the people who are out in that case are civilians in these conflict zones who might become part of a list of targets without regards for their civilian status. So I think this concept of sovereignty, we need to be careful. It's promising, but we just need to be aware that traditionally, historically, it's been used to keep certain kinds of peoples out. Just one final point, building what the hope that Ulysses has just talked about, about more resistant uses of data. Why not think about community sovereignty? Why does this always have to be
Starting point is 00:57:10 at the level of the state? In a way, the core of what we're saying is that this is not going to be easy to resist this. This is a massive thing, as large as earlier colonialism, as much as one key aspect of capitalism. So individuals can't fight back. We can only fight together. That's the only way that resistance even begins to make sense. But on what terms? The things that Ulysses is an expert about with sovereignty at the level of states are particularly difficult area and few people are very optimistic about that today. But we know there are wonderful experiments going on in Latin America and Africa and the Middle East and different parts of the
Starting point is 00:57:49 world where communities are saying, you know what, we've had enough. We're just going to imagine a future without that way of doing things and on a different basis. And actually, we have just about enough tech to make that happen. And we, as we said at the beginning, we're not in any way against data. We're in favor of data collectives which track male violence against women. We're in favor of data collectives that track the violence of sovereign states against migrant populations, or track its use in warfare, and so on. We're absolutely in favor of that. What we need, and this is the reason we wrote the book, is to flip things over and say, hang on, why wouldn't we start from what communities want? Tim McGebrow had a beautiful
Starting point is 00:58:31 phrase like that. I can't remember the exact wording. Why wouldn't we start with what the people who are being harmed by data want, rather than what the people who are doing the harm want? Why can't we shift the terms of the debate? That's why we wrote the book. And we hope people more and more will start to listen to that and take it into their own hands. It's such a great point and such a great place to leave our conversation as well. You know, as I said earlier, I would highly recommend people pick up the book Data Grab. I really enjoyed it. Ulysses, Nick, thank you so much for taking the time to speak with me. I really appreciate it. Thank you, Paris. Great pleasure. Thanks a lot.
Starting point is 00:59:10 Ulysses Mejias is a professor at SUNY Oswego, and Nick Coltry is a professor at the London School of Economics, and they are the co-authors of Data Grab. Tech Won't Save Us is made in partnership with The Nation magazine and is hosted by me, Paris Marks. Production is by Eric Wickham, and transcripts are by Bridget Palou-Fry. Tech Won't Save Us relies on the support of listeners like you to keep providing critical perspectives on the tech industry. You can join hundreds of other supporters by going to patreon.com slash tech won't save us and making a pledge of your own. Thanks for listening. Make sure to come back next week. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.