TED Talks Daily - How a worm could save humanity from bad AI | Ramin Hasani

Episode Date: October 23, 2024

What if AI could think and adapt like a real brain? TED Fellow and AI scientist Ramin Hasani shares how liquid neural networks — a new, more flexible AI technology inspired by physics and l...iving brains — could transform how we solve complex problems.

Transcript
Discussion (0)
Starting point is 00:00:00 TED Audio Collective for TED Fellows Films, adapted for podcasts, just for our TED Talks daily listeners. TED's fellowship supports a network of global innovators, and we're so excited to share their work with you. Today, we'd like you to meet AI scientist Ramin Hassani. We talk about AI a lot here at TED because it's a world-changing technological development that's fast improving and risky, but it's hard for a layperson like me to really grasp how it functions. Ramin's new AI system, which he co-invented, addresses that issue head-on. His system gives us a lot more control and visibility into the mechanics behind the tech, making it safer and more trustworthy. After we hear from Ramin, stick around for his conversation
Starting point is 00:01:03 with TED Fellows Program Director Lily James Olds, all coming up after the break. from home as we settled down at our Airbnb during a recent vacation to Palm Springs, I pictured my own home sitting empty. Wouldn't it be smart and better put to use welcoming a family like mine by hosting it on Airbnb? It feels like the practical thing to do, and with the extra income, I could save up for renovations to make the space even more inviting for ourselves and for future guests. Your home might be worth more than you think. Find out how much at airbnb.ca slash host. And now our TED Talk of the day. My wildest dream is to design artificial intelligence that is our friend. You know, if you have an AI system that can help us understand mathematics, you can solve the economy of the world. If you have an AI system that helps us understand mathematics, you can solve the economy of the world.
Starting point is 00:02:05 If you have an AI system that can understand humanitarian sciences, we can actually solve all of our conflicts. I want this system to, given Einstein's and Maxwell's equations, take it and solve new physics, you know? If you understand physics, you can solve the energy problem. So you can actually design ways for humans to be the better versions of themselves. I'm Ramin Hassani. I'm the co-founder and CEO of Liquid AI. Liquid AI is an AI company built on top of a technology that I invented back at MIT. It's called Liquid Neural
Starting point is 00:02:41 Networks. These are a form of flexible intelligence as opposed to today's AI systems that are fixed, basically. So think about your brain. You can change your thoughts. When somebody talks to you, you can completely change the way you respond. You always have a mechanism that we call feedback in your system. So basically, when you receive information from someone as an input, you basically process that information and then you reply. For liquid neural networks, we simply got those feedback mechanisms and we added that to the system. So that means it has the ability of
Starting point is 00:03:16 thinking. That property is inspired by nature. We looked into brains of animals, and in particular, a very, very tiny worm called C. elegans. The fascinating fact about the brain of the worm is that it shares 75% of the genome that it has with humans. We have the entire genome mapped. So we understand a whole lot about the functionality of its nervous system as well. So if you understand the properties of cells in the worm, maybe we can build intelligent systems that are as good as the worm and then evolve them into systems that are better than even humans.
Starting point is 00:03:58 The reason why we are studying nature is the fact that we can actually, having a shortcut through exploring all the possible kind of algorithms that you can actually having a shortcut through exploring all the possible kind of algorithms that you can design. You can look into nature, that would give you like a shortcut to really faster get into efficient algorithms. Nature has done a lot of search, billions of years of evolution, right?
Starting point is 00:04:17 So we learned so much from those principles. I just brought a tiny principle from the worm into artificial neural networks and now they are flexible and they can solve problems in an explainable way that was not possible before. AI is becoming very capable, right? The reason why AI is hard to regulate is because we cannot understand the system.
Starting point is 00:04:36 Even the people who design the systems, we don't understand those systems. They are black boxes. With Liquid, because we are fundamentally using mathematics that are understandable, we have tools to really understand and pinpoint which part of the system is responsible for what. You're designing white box systems. So if you have systems that you can understand
Starting point is 00:04:57 their behavior, that means even if you scale them into something very, very intelligent, you can always have a lot of control over that system because you understand it. You can never let it go rogue. So all of the crises we're dealing with right now, you know, doomsday kind of scenarios, is all about scaling a technology that we don't understand. We liquidate our purposes to really calm people down
Starting point is 00:05:21 and show people that, hey, you can have very powerful systems that you have a lot of control and visibility into their working mechanisms. The gift of having something very super intelligence is massive and it can enable a lot of things for us. But at the same time, we need to have control over that technology because this is the first time that we're going to have a technology that's going to be better than all of humanity combined. That was Ramin Hassani, a 2024 TED Fellow. Stick around after the break to hear Ramin go deeper into his work.
Starting point is 00:06:02 And now a special conversation between TED Fellow Ramin Hassani and TED Fellow's Program Director Lily James-Olds. Hi, Ramin. It's so great to have you with us today. Thanks for having me. So does this mean we can all stop panicking about AI? Well, a little bit, yes. So we are moving in that direction. We're opening the black box. We are trying to improve the control that we have as designers of AI systems in a way that you have a lot more control on the outcomes, on the outputs of an AI system. And you can put boundaries around what you want them to do,
Starting point is 00:06:39 you know, and that controlled ability is something that we want to create for AI and build systems that are fundamentally and inherently understandable. Okay, I'm going to come back to that because I have a lot of questions on that. But just to start, so you say that looking into nature helped you and your team invent these liquid neural networks, and in particular, one specific worm that surprisingly shares a lot in common with humans. Now, this is totally wild to me. I had no idea that I was so closely related to a worm. Can you tell me a bit more about how this worm's brain inspired your discovery of liquid neural networks? The worm is called C. elegans. This is the first animal that we had in its entire nervous system mapped. You know,
Starting point is 00:07:22 neuroscientists anatomically connected all the connections that exist in the brain of the worm, 302 neurons. The scientists that designed this thing, they won Nobel Prizes. And the reason for that is just the fascinating fact that in the tree of evolution, 600 million years ago, we got split from this worm. So it shares 75% genetic similarity to humans. The fact that our nervous systems, our brains are actually inspired by the mapping of this kind of worm, I thought that this would be a very good place to get started. Also, you should know that the body of the worm is transparent. You can see inside how things happen. So you see that the neurons flash actually like under microscope when you look at the worm. So you can actually see how the neuron behaves
Starting point is 00:08:10 while you can record the brain activities of the worm. So you have a lot of data. So it becomes a very good model organism. So I started looking into this. I thought that, okay, so neurons and synapses are the same, almost the same in terms of functionality in this worm and in humans. So if we can understand on this worm how things work from the mathematical principles and how behavior emerges from a set of neural activities with mathematics that are involved, then we can take this and evolve this into better versions of itself, which became human brain. And maybe we can also evolve artificial intelligence that way. That's so crazy that that discovery came from nature so directly. So back to where you started this conversation. Right now, we don't have the transparency into how current AI systems work. As you said, it's a black box.
Starting point is 00:09:01 And you said that this is the problem and why we don't have control over these systems. I guess my first question is just how did we get to this point? You know, why weren't these AI systems built with transparency as a core tenant? The thing is, like, the AI systems were transparent and they are still traceable. You know, the problem that we have with these AI systems is the scale of these AI systems today. So you started taking this very simple mathematics, you know, simple if condition, you know, if something happens, neuron gets activated. If doesn't happen, the neuron turns off. Then we took this function and we scaled this technology. We scaled it into billions or now we are getting into trillions of parameters.
Starting point is 00:09:48 You know, so now a system, imagine you have trillion knobs that you have to turn. Now, if you want to go and reverse engineer, what are these trillions of knobs are actually doing? This becomes a non-tractable process. So you wouldn't be able to really say what each of these one out of trillions of knobs are actually doing
Starting point is 00:10:11 and what's the function of these things in the overall kind of behavioral generation of the generative AI system. That's why we call them black boxes. You know, when we scaled the models, we saw that much, much better and smarter behavior emerged from these AI systems. That's the excitement that we move towards, right? We always want to design systems that are
Starting point is 00:10:31 more fascinating, you know, getting closer, getting smarter than humans. And then that excitement sometimes prevents us from looking into the socio-technical challenges that these AI systems can bring, right? And that is something that we have to control. So how are the liquid neural networks different? So why are they more trustworthy? And why do we have more control over them at scale? That's a great question. So think about it like this.
Starting point is 00:11:01 When you're sitting on an airplane, you know, as a passenger, then the pilot turns on autopilot. You as a passenger completely trust that autopilot. Even if you don't understand that system, how is it that we trust that autopilot in action in such a safety critical task? The reason why you trust it is because the engineers who designed that whole system, they completely understand how that mathematics works. They go through multiples of testing so that they can get into this safety critical kind of system. That's the best type of explainability that you want to have. You know, you want the engineers who design the systems understand fully how the technology works. Now with liquid neural networks, the core mathematics is something that is tractable. That's why us engineers and scientists are being able to actually get into these systems.
Starting point is 00:11:51 And we have a lot of tools to really steer and put controls on top of this. Something that's been on my mind and many people's minds a lot is how can we make sure that AI systems are built on ethical frameworks and inclusive data? Data representation is one aspect. that AI systems are built on ethical frameworks and inclusive data. Data representation is one aspect. The ability of a human to understand also what happens inside a model is another aspect of it, right? Then these two together, data representation plus us being able to explain models, that's the road towards achieving safe artificial intelligence. So fascinating. I have to say this conversation does make me feel a little bit more at ease. So
Starting point is 00:12:29 thank you for taking the time to talk to us today. My last question is, if someone listening is interested in diving deeper into this topic, what resources would you recommend to them in terms of a book, a podcast or something else? I've given a lot of talks about liquid neural networks online, but more concentrated kind of material, you can find it on our website. We started the company around liquid neural networks and taking these technologies to the next level
Starting point is 00:12:53 and providing it to the society for developing safe AI. And this is liquid.ai. So this is where you can find all sorts of information. There are blog posts around like the research papers, talks, products, and everything around the topic. Amazing. Well, thank you so much, Ramin. Absolutely. Thank you. Support for this show comes from Airbnb. If you know me, you know I love staying in Airbnbs when I travel. They make my family feel most at home when we're away from home. As we settled down at our Airbnb during a recent vacation to Palm Springs, I
Starting point is 00:13:30 pictured my own home sitting empty. Wouldn't it be smart and better put to use welcoming a family like mine by hosting it on Airbnb? It feels like the practical thing to do, and with the extra income, I could save up for renovations to make the space even more inviting for ourselves and for future guests. Your home might be worth more than you think. Find out how much at Airbnb.ca slash host. To learn more about the TED Fellows program and watch all the TED Fellows films, go to fellows.ted.com. And that's it for today. TED Talks Daily is part of the TED Audio Collective. This episode was produced and edited by our team, Martha Estefanos, Oliver Friedman, Brian Green, Autumn Thompson, and Alejandra Salazar.
Starting point is 00:14:17 It was mixed by Christopher Fazi-Bogan. Additional support from Emma Taubner and Daniela Balarezo. I'm Elise Hu. I'll be back tomorrow with a fresh idea for your feed. Thanks for listening. Looking for a fun challenge to share with your friends and family? Today, TED now has games designed to keep your mind sharp while having fun. Visit TED.com slash games to explore the joy and wonder of TED games.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.