TED Talks Daily - We’re doing AI all wrong. Here’s how to get it right | Sasha Luccioni

Episode Date: October 30, 2025

Artificial intelligence is changing everything — but at what cost? AI sustainability expert Sasha Luccioni exposes how tech companies' massive data centers are burning through energy and wrecking th...e planet. She introduces a powerful alternative: small but mighty AI models that could flip the script and make the technology smarter, fairer and sustainable.Interested in learning more about upcoming TED events? Follow these links:TEDNext: ted.com/futureyou Hosted on Acast. See acast.com/privacy for more information.

Transcript
Discussion (0)
Starting point is 00:00:00 You're listening to TED Talks Daily, where we bring you new ideas to spark your curiosity every day. I'm your host, Elise Hu. Sometimes it seems like it's all anyone can talk about. Will AI help to transform the future of humanity for the better? Or will it bring about the end of humanity as we know it? For AI sustainability expert Sasha Lucione, those questions missed the point. In her talk, she shares why she believes we're currently doing AI wrong at the expense of people and our fragile planet. Showing why we must make decisions with sustainability in mind, she paints a picture of a future in which AI is used for the sake of all humanity and the planet and not just a select few.
Starting point is 00:01:00 AI has been promised to transform the future of humanity. Or it's set to bring about the end of humanity as we know it. It really depends on who you ask. In my opinion, both of these statements are wrong, and what they do is they distract us from the real issue at hand. We're doing AI wrong at the expense of people and the planet. As it stands, a handful of large corporations are using huge capital to sell us large language models,
Starting point is 00:01:29 large language models, or LLMs, as the solution to all of our problems. Possibly because they think that they'll bring about superintelligence, emotional intelligence, basically whatever flavor of intelligence is trending in Silicon Valley these days, and in this race, they're building more and bigger data centers,
Starting point is 00:01:48 the people and the planet be damned. Meta is set to build a data center the size of Manhattan in the next few years, part of an investment of hundreds of billions of dollars towards a quest to develop super-intelligence. OpenAI recently announced the first phase of their Stargate Data Center in Texas. Once operational, it's set to emit 3.7 million tons of CO2 equivalents per year, as much as the whole country of Iceland.
Starting point is 00:02:20 XAI is currently being sued by the residents of South Memphis because of the air pollution caused by their 35 questionably legal gas turbines which are powering its data center colossus, exacerbating the health issues of the city's most vulnerable residents. And yet, for years, activists and scientists like myself have been sounding the alarm when it comes to AI's increasing unsustainability. Does this ring a bell? Remember Big Oil?
Starting point is 00:02:48 Well, now we have Big AI, following the exact same playbook, using more and more resources building bigger and bigger data structures and selling us the narrative that this is somehow, inevitable. But what if we could learn from the lessons of the past and use them to build a future in which AI is giving back to the planet instead of taking away from it, a future in which AI models are small but mighty
Starting point is 00:03:11 in which they are both better performing and more sustainable? To do this, we have to take back the power, pun intended, from the big AI companies and put it back into the hands of the developers, regulators, and users of AI. Today we use AI as if we were turning on all of the lights of the stadium just to find a pair of keys, using huge AI models, trained using the energy demands of a small city
Starting point is 00:03:39 just to tell us knock-knock jokes or help us figure out what to make for dinner. This is driven by a bigger-is-better mentality. This has become somewhat of a mantra in AI. Bigger models, more compute, bigger data sets, more energy equals better performance. And the pinnacle of this approach which are LLMs, models like ChatGPT,
Starting point is 00:03:59 which are trained specifically to be general purpose, able to answer any question, generate any haiku, and act as your therapist while they're at it. But this performance comes at a cost, because models that are trained to do all tasks use more energy each time than models that it can do a certain task at a time. In a recent study I led, we looked at using LLMs to answer simple questions, like, what's the capital of Canada?
Starting point is 00:04:24 And we found that compared to a smaller task, they use up to 30 times more energy. And as this energy use grows, so does their cost. Essentially, with the number of organizations that can afford to build and deploy what's considered state-of-the-art AI shrinking, becoming limited to a handful of big tech companies with millions of dollars to burn,
Starting point is 00:04:48 while startups, academics, and nonprofits are all left in the dust. So now this handful of big AI companies companies, largely gathered by the move-fast and break-things mentality, decides the future of a technology that can impact the lives of billions of people. But in the background of all this hubbub around the deep-seeks and the chao-chip-tees of the world, a revolution has been quietly building in recent months. This revolution is driven by small LMs, which are also language models,
Starting point is 00:05:18 but there are orders of magnitude smaller than traditional LLMs. The smallest of this family has around 135 more. million parameters, making it 5,000 times smaller than Deepseek's model. These models are flipping the script on the bigger is better mentality by using less data, less compute, less energy, and still having the same level of performance. The data used to train HuggingFaces small-al-land models was carefully curated to be 60% educational web pages, explicitly chosen based on the quality of their content. This also means that the models that are trained on this data are less likely to produce
Starting point is 00:05:55 misinformation, or toxicity, when we query them. And since the models are so small, they can run literally on your phone or in your web browser, giving you access to state-of-the-art AI in the palm of your hand without needing massive data centers. And above and beyond environmental impacts, they also have benefits when it comes to cybersecurity when it comes to data privacy and sovereignty,
Starting point is 00:06:18 giving users more power over the AI that they're using. And, since they're smaller and cheaper to, train, they give smaller AI companies the ability to connect with a community and to compete with big AI companies because they can actually afford to be training and deploying these models and adapting them to different uses and then sharing them back with the community. Proving that reduce, reuse, recycle also applies to AI. But the truth of the matter is that there's more to AI than just small LMs. And if we really want to make AI more sustainable, we have to be thinking beyond LLMs to using
Starting point is 00:06:54 all sorts of different approaches that can be really useful in our fight against climate change. Because, sure, Chad GPT can tell you which countries sign the Paris Agreement, but it can't predict extreme weather events, which requires an understanding of the physics of weather patterns and geography. And sure, Claude can explain the whys and the hows of climate change,
Starting point is 00:07:14 but it can't help a farmer decide when to plant their crops based on temperature, humidity, and historical weather patterns. There are so many other approaches in AI that use less energy and still are really useful in our fighting against climate change. For example, recently, a team of researchers funded by NASA trained the Galileo models, which can be used for all sorts of different tasks,
Starting point is 00:07:35 from crop mapping to flood detection without needing specialized hardware. This makes them accessible to governments and non-profits. And Rainforest Connection uses AI to do bioacoustic monitoring. That means that they listen to the sounds of rainforests across the world, identify species, and even detect the sounds illegal logging in real time. Their AI models are so small,
Starting point is 00:07:57 they run on old cell phones powered with solar panels. An open climate fix uses AI to analyze satellite imagery, weather forecasts, and topography data to predict the output of solar and wind installations, allowing us to move forward to decarbonizing energy grids around the world. This includes data centers because
Starting point is 00:08:16 currently they're powered by mostly coal and gas, but they could be renewable if we had the right tools. But another problem is, as users of AI, we don't know how much energy and AI model is using or how much carbon it's emitting when we use it. That means that we can't make decisions with sustainability in mind as we do for the food that we eat or for how we get around town.
Starting point is 00:08:40 This led me to create the AI Energy Score project in which we tested over 100 open-source AI models across a variety of different tasks, from text generation to images, and we assign them scores from one to five stars based on energy efficiency. So say that you forgot the capital of Canada again, it's Ottawa. You could use a model like Small Alam, which would use 0.007 watt hours to give you that answer.
Starting point is 00:09:08 Or you could use a model like Deepseek, which would use 150 times more energy for that answer. But, sadly, big AI companies didn't want to play ball want to play ball and evaluate their models with our methodology. And honestly, I can't blame them because the truth might only make them look bad. Because currently, we don't have the laws or incentives that we need to encourage AI companies to evaluate the environmental impacts of their models or to take accountability for them.
Starting point is 00:09:36 The EU AI Act started this process by introducing voluntary disclosures around the energy and resource use of AI models. But enforcing this act in Europe and eventually writing laws like this across, across the world will take time that we simply don't have given the speed and the scale of the climate crisis. But the good news is that we don't need to stay hooked on the AI sold to us by big AI companies today as we've stayed hooked on the coal and plastic and fossil fuels
Starting point is 00:10:08 that have been sold to us by big oil for all these decades. And in fact, instead of believing that the future of AI is already written, that it consists of huge LLMs powered by infinite amounts of energy that will somehow result in superhuman intelligence and magically solve all of our problems, we can take back the wheel and shape an alternative future for AI together. A future where AI models are small but mighty,
Starting point is 00:10:35 where they run on our cell phones and do the task they're meant to do without needing huge data centers. A future in which we have the information we need to choose one AI model over the other based on its carbon footprint. A future in which legislation exists that makes big AI companies take accountability
Starting point is 00:10:54 for the damage that they're causing to people and the environment. A future in which AI serves all of humanity and not just a handful of for-profit tech companies. With every prompt, every click, and every query, we can reinvent the future of AI to be more sustainable together. Thank you. That was Sasha Lucione, speaking at a TED Countdown event in New York in partnership with the Bezos Earth Fund in 2025.
Starting point is 00:11:30 If you're curious about Ted's curation, find out more at TED.com slash curation guidelines. And that's it for today. Ted Talks Daily is part of the TED Audio Collective. This talk was fact-checked by the TED Research team and produced and edited by our team, Martha Estefanos, Oliver Friedman, Brian Green, Lucy Little, and Tonica, Sung Marnivong. This episode was mixed by Christopher Faisie Bogan. Additional support from Emma Tobner and Daniela Balezzo. I'm Elise Hugh. I'll be back tomorrow with a fresh idea for your feed. Thanks for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.