Theories of Everything with Curt Jaimungal - Physics Absorbed Artificial Intelligence & (Maybe) Consciousness

Episode Date: September 3, 2025

As a listener of TOE you can get a special 20% off discount to The Economist and all it has to offer! Visit https://www.economist.com/toe MIT physicist Max Tegmark argues AI now belongs inside physic...s—and that consciousness will be next. He separates intelligence (goal-achieving behavior) from consciousness (subjective experience), sketches falsifiable experiments using brain-reading tech and rigorous theories (e.g., IIT/φ), and shows how ideas like Hopfield energy landscapes make memory “feel” like physics. We get into mechanistic interpretability (sparse autoencoders), number representations that snap into clean geometry, why RLHF mostly aligns behavior (not goals), and the stakes as AI progress accelerates from “underhyped” to civilization-shaping. It’s a masterclass on where mind, math, and machines collide. Join My New Substack (Personal Writings): https://curtjaimungal.substack.com Listen on Spotify: https://open.spotify.com/show/4gL14b92xAErofYQA7bU4e Timestamps: - 00:00 - Why AI is the New Frontier of Physics - 09:38 - Is Consciousness Just a Byproduct of Intelligence? - 16:43 - A Falsifiable Theory of Consciousness? (The MEG Helmet Experiment) - 27:34 - Beyond Neural Correlates: A New Paradigm for Scientific Inquiry - 38:40 - Humanity: The Masters of Underestimation (Fermi's AI Analogy) - 51:27 - What Are an AI's True Goals? (The Serial Killer Problem) - 1:03:42 - Fermat's Principle, Entropy, and the Physics of Goals - 1:15:52 - Eureka Moment: When an AI Discovered Geometry on Its Own - 1:30:01 - Refuting the "AI Doomers": We Have More Agency Than We Think Links mentioned: - Max’s Papers: https://scholar.google.com/citations?user=eBXEZxgAAAAJ&hl=en - Language Models Use Trigonometry to Do Addition [Paper]: https://arxiv.org/abs/2502.00873 - Generalization from Starvation [Paper]: https://arxiv.org/abs/2410.08255 - Geoffrey Hinton [TOE]: https://youtu.be/b_DUft-BdIE - Michael Levin [TOE]: https://youtu.be/c8iFtaltX-s - Iceberg of Consciousness [TOE]: https://youtu.be/65yjqIDghEk - Improved Measures of Integrated Information [Paper]: https://arxiv.org/abs/1601.02626 - David Kaiser [TOE]: https://youtu.be/_yebLXsIdwo - Iain McGilchrist [TOE]: https://youtu.be/Q9sBKCd2HD0 - Elan Barenholtz & William Hahn [TOE]: https://youtu.be/A36OumnSrWY - Daniel Schmachtenberger [TOE]: https://youtu.be/g7WtcTATa2U - Ted Jacobson [TOE]: https://youtu.be/3mhctWlXyV8 - The “All Possible Paths” Myth [TOE]: https://youtu.be/XcY3ZtgYis0 SUPPORT: - Become a YouTube Member (Early Access Videos): https://www.youtube.com/channel/UCdWIQh9DGG6uhJk8eyIFl1w/join - Support me on Patreon: https://patreon.com/curtjaimungal - Support me on Crypto: https://commerce.coinbase.com/checkout/de803625-87d3-4300-ab6d-85d4258834a9 - Support me on PayPal: https://www.paypal.com/donate?hosted_button_id=XUBHNMFXUX5S4 SOCIALS: - Twitter: https://twitter.com/TOEwithCurt - Discord Invite: https://discord.com/invite/kBcnfNVwqs Guests do not pay to appear. Theories of Everything receives revenue solely from viewer donations, platform ads, and clearly labelled sponsors; no guest or associated entity has ever given compensation, directly or through intermediaries. #science Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 When Michael Faraday first proposed the idea of the electromagnetic field, people were like, what are you talking about? You're saying there is some stuff that exists, but you can't see it, you can't touch it. That sounds like total non-scientific ghosts. Most of my science colleagues still feel that talking about consciousness as science is just bullshit. But what I've noticed is when I push them a little harder about why they think it's bullshit, they split into two camps that are in complete disagreement with each other. You can have intelligence without consciousness.
Starting point is 00:00:28 And you can have consciousness without intelligence. Your brain is doing something remarkable right now. It's turning these words into meaning. However, you have no idea how. Professor Max Tegmark of MIT studies this puzzle. You recognize faces instantly, yet you can't explain the unconscious processes. You dream full of consciousness while you're outwardly not doing anything. Thus, there's intelligence without consciousness.
Starting point is 00:00:58 and consciousness without intelligence. In other words, they're different phenomena entirely. Tegmark proposes something radical. Consciousness is testable in a new extension to science where you become the judge of your own subjective experience. Physics absorbed electromagnetism and then atoms and then space time, and now Tegmark says it's swallowing AI. In fact, I spoke to Nobel Prize winner Jeffrey Hinton about this specifically.
Starting point is 00:01:27 Now, to Max Tegmark, the same principle that explains why light bends and water may actually explain how thoughts emerge from neurons. I was honored to have been invited to the Augmentation Lab Summit, which was a weekend of events at MIT last week. This was hosted by MIT researcher Dunya, Baradari. The summit featured talks on the future of biological and artificial intelligence, brain computer interfaces, and included speakers such as Stephen Wolfram and Andreas Gomez-Emilson. conversations with them will be released on this channel in a couple weeks, so subscribe to get notified. Or you can check the substack, kurtjymongle.com, as I release episodes early over there. A special thank you to our advertising sponsor, The Economist.
Starting point is 00:02:11 Among weekly global affairs magazines, The Economist is praised for its non-partisan reporting and being fact-driven. This is something that's extremely important to me. It's something that I appreciate. I personally love their coverage of other topics that aren't just. politics-based as well. For instance, the Economist has a new tab for artificial intelligence on their website, and they have a fantastic article on the recent DESE Dark Energy Survey. It surpasses, in my opinion, Scientific Americans' coverage. Something else I love, since I have ADHD,
Starting point is 00:02:44 is that they allow you to listen to articles at 2x speed, and it's from an actual person, not a dubbed voice. The British accents are a bonus. So if you're passionate about expanding your knowledge and gaining a deeper understanding of the forces that shape our world, I highly recommend subscribing to The Economist. It's an investment into your intellectual growth, one that you won't regret. I don't regret it. As a listener of Toe, you get a special discount. Now you can enjoy The Economist and all it has to offer for less.
Starting point is 00:03:15 Head over to their website, Economist.com slash T-O-E to get started. Make sure you use that link. That's www.economist.com slash tow to get that discount. Thanks for tuning in, and now back to the explorations of the mysteries of the universe with Max Tegmark. Max, is AI physics? There was a Nobel Prize awarded for that, but what are your views? I believe that artificial intelligence has gone from being not physics to being physics, actually. you know one of the best ways to insult physicists is to tell them that their work isn't physics
Starting point is 00:03:55 as if somehow there's a generally agreed on boundary between what's physics and what's not or between what's science and what's not but i find like the most obvious lesson we get if we just look at the history of science is that the boundary has evolved some things that used to be considered scientific by some, like astrology has left, the boundary has contracted, so that's not considered science now. And then a lot of other things that were purported as being non-scientific are now considered obviously science. Like, well, I sometimes teach the electromagnetism course. And I remind my students that when Michael Faraday first proposed the idea of the electromagnetic field
Starting point is 00:04:42 people were like what are you talking about you're saying there is some stuff that exists but you can't see it you can't touch it that sounds like ghosts like total non-scientific bullshit you know and they really
Starting point is 00:04:58 gave them a hard time for that and the irony is not only is that considered part of physics now but you can see the electromagnetic field it's in fact the only thing we can see because light is an electromagnetic wave. And after that, things like black holes, things like atoms, which Max Planck famously said is not physics, you know,
Starting point is 00:05:18 even talk about what our universe was doing 13.8 billion years ago have become considered part of physics. And I think AI is now going the same way. I think that's part of the reason that Jeff Hinton got the Nobel Prize in physics. Because what is physics? To me, physics is all about looking at some complex, interesting system, doing something and trying to figure out how it works. We started on things like the solar system and atoms,
Starting point is 00:05:50 but if you look at an artificial neural network that can translate French into Japanese, that's pretty impressive too. And there's this whole field that started blossoming now, that I also had a lot of fun working in called Mechanistic Interpretability where you study an intelligent artificial system to try to ask these basic questions, like, how is it work? Are there some equations that describe it?
Starting point is 00:06:21 Are there some basic mechanisms and so on? And in a way, I think of traditional physics, like astrophysics, for example, is just mechanistic interpretability applied to the universe. and Hopfield also got the Nobel Prize last year he was the first person to show that hey you can actually write down
Starting point is 00:06:43 an energy landscape so potential energy on the vertical axis how the potential energy depends on where you are and think of each little valley as a memory you might wonder like
Starting point is 00:06:59 how the heck can I like store information in an egg cart and say if it has 25 valleys in it. Well, very easy, you know, you can put the marble in one of them. Now, that's log 25 bits right there. And how do you retrieve what the memory is? You can look, where is the marble? Hopfield had this amazing physics insight.
Starting point is 00:07:27 If you think of there as being any system whose potential energy function has many, many, many different minima that are pretty stable. You can just use it to store information. But he realized that that's different from the way computer scientists used to store information. It used to be like the whole von Neumann paradigm, you know, with a computer. You're like, tell me what's in this variable. Tell me what number is sitting in this particular address. You go look here.
Starting point is 00:07:54 Right. That's how traditional computers store things. But if I say to you, you know, twinkle, twinkle. Uh-huh. Little Star. Yeah, that's a different kind of memory retrieval, right? I didn't tell you, hey, give me the information that's stored in those neurons over there. I gave you something which was sort of partial part of the stuff and you filled it in.
Starting point is 00:08:20 This is called associative memory. And this is also how Google will give you something. You can type something in that you don't quite remember and it'll give you the right thing. And Hopfield showed, coming back to the egg carton, all right, that if you put, if you don't remember exactly what, suppose you want to memorize the digits of pie and you have an energy function as a function,
Starting point is 00:08:43 where the actual minimum is it exactly 3.15, et cetera. Yeah. But you don't remember exactly what pie is. It's like three something. Yes. So you put a marble at three, you let it roll. as long as it's in the basin of attraction, whose minimum is at Pi, it's going to go there.
Starting point is 00:09:01 So to me, this is an example of how something that's felt like it had nothing to do with physics, like memory, can be beautifully understood with tools from physics. You have an energy landscape, you have different minima, you have dynamics, the Hopfield network. So I think, yeah, it's totally fair that Hinton and then Hopfield got an enterprise in physics. And it's because we're beginning to understand that we can expand again the domain of what is physics to include very deep questions about intelligence, memory, and computation.
Starting point is 00:09:38 What about consciousness? So you mentioned that Faraday with an electromagnetic field that was considered to be unsubstantiated, unfalsifiable nonsense, or just ill-defined. Consciousness seems to be at a similar stage where many scientists or many physicists tend to look at the way that consciousness studies or consciousness is studied
Starting point is 00:09:58 or consciousness is talked about is, well, firstly, what's the definition of consciousness? You all can't agree, there's phenomenal, there's access, et cetera. And then even there, what is it? And then the critique would be, well, you're asking me for a third-person definition of something that's the first-person phenomenon,
Starting point is 00:10:13 okay, so how do you view consciousness in this? Yeah, I love these questions. I feel the consciousness is actually... probably the final frontier the final thing which is going to end up ultimately in the domain of physics right now on the controversial borderline so let's go back to Galileo say
Starting point is 00:10:36 right if if he dropped a grape and a hazelnut he could predict exactly when they were going to hit the ground and that how far they fall grows like a parabola as a function of time.
Starting point is 00:10:54 But he had no clue why the grape was green and the hazelah was brown. Then came Maxwell's equations and we started to understand that light and colors is also physics. And we got the equations for it. And then we got, he couldn't figure out Galileo either why the
Starting point is 00:11:10 grape was soft and why the hazelah was hard. Then we got quantum mechanics and we realized that all these properties of stuff could be calculated from the Schrodinger equation and also brought into physics. And then And then we started, and then intelligence seemed like such a holdout. But we've already talked about now how, if you start breaking it into components like
Starting point is 00:11:32 intelligent, like memory, and we can talk more about computation and learning, how that can also very much be understood as a physical process. So what about consciousness? Yeah, so I'd say most of my science colleagues still feel that. talking about consciousness as science is just bullshit. But what I've noticed is when I push them a little harder about why they think it's bullshit, they split into two camps that are in complete disagreement with each other.
Starting point is 00:12:04 Half of them roughly will say, oh, consciousness is bullshit because it's just the same thing as intelligence. Okay. And the other half will say consciousness is bullshit because obviously machines can't be conscious, which is obviously totally inconsistent. with saying it's the same thing as intelligence. You know, what's really powered the AI revolution in recent decades
Starting point is 00:12:30 is just moving away from philosophical quibbles about what is intelligence really mean in a deep philosophical sense. And instead, making a list of tasks and saying, can my machine, can it do this task? Can it do this task? And that's quantitative. You can train your systems to get better at the task. And I think you'd have a very hard time if you went to the neural
Starting point is 00:12:52 Groups conference and argue that machines can never be intelligent, right? So if you then say intelligence is the same as consciousness, you're predicting that machines are conscious then if they're smart. But we know that consciousness is not the same as intelligence, just like some very simple introspection. We can do it like right now. So, for example, what did it say here? I guess we shouldn't do product.
Starting point is 00:13:24 No, I don't mind. Let's do this way. What is it saying? It's towards more interpretable AI with sparse auto encoders by Joshua Engels. Great. This is a PhD thesis of my student, Josh Engels, or the master's thesis. So, like, how did you do that computation? Like 30 years ago, if I gave you just a bunch of numbers that are the actual red, green,
Starting point is 00:13:47 and blue strengths of the pixels and this, and ask you what does it say? people didn't, this is a hard problem. Even harder is if you just open your eyes and ask who is this and you say it's Max, right? But you can do it like this. But think about how it felt. Do you actually feel that you, if you open your eyes and you see this as Max
Starting point is 00:14:09 that you know the algorithm that you used to recognize my face? No. Same here. For me, it's pretty obvious. It feels like my consciousness, there's some part of my information process thing that is conscious and it's kind of got an email from the face recognition module saying
Starting point is 00:14:27 you know face recognition complete the answer is you know it's it's so and so so in other words you do something when you recognize people that's quite intelligent but not conscious right and i would say actually a large fraction of what your brain does you're just not conscious about you you find out about the results of it often after the fact So you can have intelligence without consciousness. It's the first point I'm making. And second, you can have consciousness without intelligence, without accomplishing any tasks. Like, did you have any dreams last night?
Starting point is 00:15:06 None that I remember. But have you ever had a dream that you remember? Yeah. So there was consciousness there. But if someone was just looking at you lying there in the bed, you probably weren't accomplishing anything, right? So I think it's obvious that consciousness is not the same. you can have consciousness without intelligence and vice versa. So those who say that consciousness equals intelligence are being sloppy.
Starting point is 00:15:32 Now, what is it then? My guess is that consciousness is a particular type of information processing and that intelligence is also a typical type of information processing, but that there's a Venn diagram like this. There are some things that are intelligence and intelligent and conscious. Some are conscious, and some of them are conscious, but not very intelligent. And so the question then becomes to try to understand, can we write down some equations or formulate some principles
Starting point is 00:16:06 for what kind of information processing is intelligent and what kind of information processing is conscious? And I think my guess is that, something to be conscious, there's at least some sufficient conditions that it probably has to have. There has to be information, a lot of information there, something to be the content of consciousness, right? The, there's an Italian scientist, Julio Tononi, who has put a lot of creative thought to this and triggered enormous controversy also who argues that there's one
Starting point is 00:16:55 one necessary condition for consciousness is what he calls integration basically that if it's going to subjectively feel like a unified consciousness like your consciousness it cannot consist of two information processing systems that don't communicate
Starting point is 00:17:15 with each other because if conscious is the way information feels when it's being processed, right, then if this is the information that's conscious and it's just completely disconnected from this information, there's no way that this information can be part of what it's conscious of. Just a quick question. Ultimately, what's the difference between information processing, computing, and communication?
Starting point is 00:17:42 So communication, I would say, is just a very simple, special case of information processing. You have some information here. and you make a copy of it the goes ends up over there it's a volleyball you send over yeah volleyball you send over copy this to that
Starting point is 00:17:54 but computation can be much more complex than that and then the you see that was information processing communication and what was the third word computation and yeah
Starting point is 00:18:10 so computation and information processing I would say is more or less the same thing then you can try to classify different kinds of information processing depending on how complex it is and mathematicians have been doing an amazing job there even though they still don't know whether p equals n p and so on but just coming back to consciousness again i think a mistake many people make when they think about their
Starting point is 00:18:34 own consciousness is like can you see the beautiful sunlight coming in here from the window and some colors and so on it's the to have the this model this model that somehow you're actually conscious of that stuff, that the content of your consciousness somehow is the outside world? I think that's clearly wrong because you can experience those things when your eyes are closed, when you're dreaming, right? So I think the conscious experience
Starting point is 00:19:05 is intrinsic to the information processing itself. What you are actually conscious about when you look at me isn't me. It's the world model that you have and the model you have in your head right now of me, and you can be conscious of that whether you're awake or whether you're sleep. And then, of course, you're using your senses
Starting point is 00:19:24 and all sorts of analysis tools to constantly update your inner world model to match relevant parts of what's outside and that's what you're conscious of. That's what you're conscious of. So what Tononi is saying is that the information processing has to be such that there's no way of this
Starting point is 00:19:49 that it isn't actually just secretly two separate parts that don't communicate at all and cannot communicate with each other because then they would basically be like two parallel universes that were just unaware of each other
Starting point is 00:20:04 and you wouldn't be able to have this feeling that it's all unified. I actually think that's a very reasonable criteria and he has a particular formula he calls phi for measuring how integrated things it is are and the things that have a high fire are more conscious.
Starting point is 00:20:22 I wasn't completely sure whether that was the only formula that had that property. So I wrote a paper once to classify all possible formulas that have that property and it turned out there was less than a hundred of them. So I think it's actually quite interesting to test if any of the other ones
Starting point is 00:20:38 fit the experiments better than his. but just to close and finish up on the why people say consciousness is bullshit though I think ultimately the main reason is either they feel it sounds too philosophical or they say oh you can never test consciousness theories because how can you test if I'm conscious or not when you all you can observe as my behavior right but the here is a misunderstanding I'm more actually I'm much more optimistic can I tell you about an experiment I where you can test the consciousness theory? No, of course.
Starting point is 00:21:16 So suppose you have someone like Julia Tannone or anyone else who has really stuck their name, neckout, okay, and written down a formula for what kind of information processing is conscious, okay? And suppose we put you in one of our MEG machines here at MIT or some future scanner that can read out in a massive amount of your neural data in real time, okay, and you connect that to this computer that uses that theory to make predictions about what you're conscious of, okay?
Starting point is 00:21:53 And then now it says, I predict that you're consciously aware of a water bottle. And you're like, yeah, that's true. Yes, theory. And then it says, okay, now I predict that you're, I see information processing there, about regulating your pulse and I predicted you're consciously aware of your heartbeat
Starting point is 00:22:14 you're like no I'm not you've never ruled out that theory actually it made a prediction about your subjective experience and you yourself can falsify that right so first of all it is possible
Starting point is 00:22:30 for you to rule out the theory to your satisfaction that might not convince me because you told me that you told me that you weren't aware of your carpeed, maybe I think you're lying or maybe something, whatever. But then you can go, okay, hey, Max, why don't you try this experiment? And I put on my, I put on my MEG helmet and I work with this. And then it starts making some incorrect assumptions about what
Starting point is 00:22:55 I'm experiencing and I'm now also convinced that it's ruled out. It's a little bit different from how we usually rule out theories, but at the end of the day, anyone who cares about this can be convinced that this theory sucks and belongs on the garbage dump of history, right? And conversely, suppose that this theory just again and again and again and again keeps predicting exactly what you're conscious of and never anything that you're not conscious about, you would gradually start getting kind of impressed, I think. And if you moreover read about what goes into this theory and you said, wow, this is a beautiful formula and it just kind of philosophically makes sense that these are the criteria
Starting point is 00:23:34 that consciousness should have and so on, you might be tempted now to try to extrapolate the and I wonder if it works also on on other biological animals, maybe even on computers and so on. You know, this is not altogether different from how we've dealt with, for example, general relativity, right?
Starting point is 00:23:53 So you might say you can never, it's bullshit to talk about what happens inside black holes because you can't go there and check and then come back and tell your friends or publish your findings and physical review letters, right? But what we're actually testing is not some philosophical ideas about black holes.
Starting point is 00:24:17 We're testing a mathematical theory, general relativity. I have it there in a frame by my window, right? And so what's happened is we tested it on the perihelian shift of mercury, how it's not really going in an ellipse, but the ellipse is processing a bit. We tested it, and it worked. We tested it on how gravity bends light. And then we extrapolated it to all such a stuff way beyond what Einstein had thought about. Like, what would happen when our universe was a billion times smaller in volume?
Starting point is 00:24:51 And what would happen when black holes get really close to each other and give off gravitational waves? And it just passed all these tests also. So that gave us a lot of confidence in the theory and therefore also in the predictions that that we haven't been able to test yet, even the predictions that we can never test, like what happens inside black holes. So now, so this is typical for science, really.
Starting point is 00:25:17 If someone says, you know, I like Einstein, you know, I like what it did for predicted for black hole, for gravity in our solar system, but I'm going to opt out of the black hole prediction. You can't do that. It's not like, oh, I want coffee, but decaf, you know. If you're going to buy the things,
Starting point is 00:25:37 theory, you need to buy all its predictions, not just the ones you like. And if you don't like the predictions, well, come up with an alternative to general relativity. Write down the math and then make sure that it correctly predicts all the things we can test. And good luck, because some of the smartest humans on the planet have spent 100 years trying and failed, right? So if we have a theory of consciousness in the same vein, right, which correctly predicts the subjective experience on whoever puts on this device. and test the predictions for what they are conscious about, and it keeps working.
Starting point is 00:26:11 I think people will start taking pretty seriously also what it predicts about coma patients who seem to be unresponsive, whether they're having locked in syndrome or in a coma, and even what it predicts about machine consciousness, whether machines are suffering or not. And people who don't like that, they will then be incentivized to work harder
Starting point is 00:26:33 to come up with an alternative theory that at least predicts subjective experience. So this was my, I'll get off my soapbox now, but this is why I strongly disagree with people who say that consciousness is all bullshit. I think there's actually more saying that because there's an excuse to be lazy and not work on it. Hi, everyone.
Starting point is 00:26:53 Hope you're enjoying today's episode. If you're hungry for deeper dives into physics, AI, consciousness, philosophy, along with my personal reflections, you'll find it all on my substack. Subscribers get first access to new episodes, new posts as well, behind the scenes insights, and the chance to be a part of a thriving community of like-minded pilgrimers. By joining, you'll directly be supporting my work and helping keep these conversations at the cutting edge.
Starting point is 00:27:19 So click the link on screen here. Hit subscribe, and let's keep pushing the boundaries of knowledge together. Thank you and enjoy the show. Just so you know, if you're listening, it's c-U-R-T-J-A-M-M-G-A-L.org, Kirt-J-I-M-M-M-G-A-L.org. So in the experiment where you put some probes on your brain in order to discern which neurons are firing or what have you. So that would be a neural correlate. I'm sure you've already thought of this. Okay, so that you're correlating some neural pattern with the bottle.
Starting point is 00:27:50 And you're saying, hey, okay, I think you're experiencing a bottle. But then technically, are we actually testing consciousness or testing the further correlation of that it tends to be that when I ask you the question, are you experiencing a bottle? and we see this neural pattern that's correlated with you saying yes so it's still another correlation is it not well but you're not trying to convince when the experiment is being run on you
Starting point is 00:28:14 you're not trying to convince me it's just you talking to the computer you are just doing experiments basically on the theory there's no one else involved no other human and you're just trying to convince yourself so you sit there and you have all sorts of thoughts you might just decide to click
Starting point is 00:28:31 to close your eyes and think about your favorite place in Toronto to see if it can predict that you're conscious of that, right? And then you might also do something else which you know you can do unconsciously and see if you can trick it into predicting that you're conscious of that information
Starting point is 00:28:49 that you know is being processed in your brain. So ultimately you're just trying to convince yourself that the theory is making incorrect predictions. I guess what I'm asking is in this case, I can see being convinced that it can read my mind in the sense that it can roughly determined what I was seeing. But I don't see how that would tell this other system that I'm conscious of that. In the same way that we can see what files are on a computer. Doesn't mean that
Starting point is 00:29:12 those files are, or when we do some cut and paste, we can see some process happening. Well, you're not trying to convince the computer. The computer is coded up to just make the predictions from this putative theory of consciousness, this mathematical theory, right? And then your job is just to see, oh, those are the wrong equations or the right equations, and the way you'd ascertain that is to see whether it correctly or incorrectly predicts what you're actually subjectively aware of. We should be clear that we're defining consciousness here just simply as subjective experience, right? Which is very different from talking about what information is in your brain. You have all sorts of memories in your brain right
Starting point is 00:29:52 now that you probably haven't thought about for months. And that's not your subjective experience right now. And even again when I open my eyes and I see a person there's a computation happening to figure out exactly who they are. There's all sort of detailed information in there probably about
Starting point is 00:30:12 some angles, about their ears and stuff which I'm not conscious about it at all. And if the machine incorrectly says and I'm conscious about that, again, the theory has failed. So it's quite hard. It's like if you look at my messy desk or I show you a huge amount of information in your brain or in this book.
Starting point is 00:30:30 And suppose there's some small subset of this which is highlighted in yellow, you know. And you have to have a computer that can predict exactly what's highlighted in yellow. It's pretty impressive if it gets it right. And in the same way, if it can accurately predict exactly which information in your brain is actually stuff that you're subjective aware on.
Starting point is 00:30:49 Okay, so let me see if I understand this. So in the global workspace theory, you have like a small desktop and pages are being sent to the desktop but only a small amount at any given time. I know that there's another metaphor of a spotlight, but whatever, let's just think of that. So this desktop is quite small relative to the globe.
Starting point is 00:31:05 Yeah. Okay, relative to the city, relative to the globe for sure. So our brain is akin to this globe because there's so many different connections, there's so many different words, there could possibly be. If there's some theory that can say, hey, this little thumbtack is what you're experiencing.
Starting point is 00:31:21 Yeah. And you're like, actually, that is correct. Yeah. Okay. Exactly. So the global workspace theory, great stuff. you know, but it is not sufficiently predictive to do this
Starting point is 00:31:30 experiment. It doesn't have a lot of equations in it, mostly words, right? So we don't have, no one has actually done this experiment yet. I would love to do it, for someone to do it, where you have a theory that's efficiently physical, mathematical, that it can
Starting point is 00:31:46 actually stick its neck out and risk being proven faults all the time. I guess what I was saying, just to wrap this up, I'm gosh, this is that. Yes, that is extremely impressive. I don't even know if that can technologically be done. Maybe it can be approximately done, but regardless
Starting point is 00:32:02 we can for sure falsify theories. But it still wouldn't suggest to an outside observer that this little square patch here or whoever is experiencing this square patch is indeed experiencing the square patch. But you already know
Starting point is 00:32:18 that you're experiencing the square patch. I know. Yes, that's the key thing. You know it. I don't know it. I don't know that. I don't know that you know, but you can convince yourself that this theory is false or that this theory is increasingly promising, right? That's the catch. And I just want to stress, you know, people sometimes say to me that, oh, you can never prove for sure that something is conscious. We can never prove anything with physics. A little dark secret, but we can never prove anything.
Starting point is 00:32:50 We can't prove that general relativity is correct. You know, probably it's wrong. Probably it's just, really good approximation. All we ever do in physics is we disprove theories. But if, as in the case of general relativity, some of the smartest people on Earth have spent over a century trying to disprove something and they still have failed, we start to take it pretty seriously and start to say, well, you know, it might be wrong, but we're going to take it pretty seriously as a really good approximation, at least, for what's actually going on, you know. That's how it works in physics, and that's the best we can ever get with consciousness also, something which people have,
Starting point is 00:33:30 which is making strong predictions and which we've, despite trying really hard to fail to falsify, so we start to start joining our respect, we start taking it more seriously. You said something interesting. Look, we can tell, or you can tell you. You can tell this theory of consciousness is correct for you
Starting point is 00:33:48 or you can convince yourself. This is super interesting because earlier in the conversation we're talking about physics, what was considered physics and what is no longer considered physics, so what is this amorphous boundary, or maybe it's not amorphous, but it changes. Yeah, absolutely changes. Do you think that's also the case for science?
Starting point is 00:34:05 Do you think science, to incorporate a scientific view of consciousness, quote-unquote, is going to have to change what it considers to be science? I think, I'm a big fan of Carl Popper. I think I personally consider things scientific if we can falsify them. if there's at least if there's no if no one can even think of a way in which we could even
Starting point is 00:34:28 conceptually in the future with arbitrary funding you know and technology tested I would say it's not science I think Popper didn't say if it can be falsified then it's science
Starting point is 00:34:38 it's more that if it can't be falsified it's not science I'll agree with that I'll agree with that also for sure but what I'm saying is consciousness is
Starting point is 00:34:46 can not a theory of consciousness that's willing to actually make concrete predictions about what you personally is subject to the experience cannot be dismissed like that because you can falsify it. If it predicts just one thing, that's wrong. If it's, right, then you falsify it.
Starting point is 00:35:05 And I would encourage people to stop wasting time on philosophical excuses for being lazy and try to build these experiments. That's why I think we should. And, you know, we saw this happen with intelligence. People had so many quibbles, but, I don't know, I have to define intelligence, and whatever. And in the meantime, you got a bunch of people who started rolling up their sleeves
Starting point is 00:35:24 and saying, well, can you build a machine that beats the best human in chess? Can you build a machine that translates Chinese into French? Can you build a machine that figures out of fold proteins, you know? And amazingly, all of those things have now been done, right? And what that's effectively done is just made people redefine intelligence. as the ability to accomplish tasks, ability to accomplish goals. That's what people in machine learning will say, if you ask them, what do they mean by intelligence?
Starting point is 00:36:00 And the ability to accomplish goals is different from having a subjective experience. The first I call intelligence, the second I call consciousness. And it's just getting a little philosophically. I mean, it's quite striking throughout the history of physics, how oftentimes we've vastly delayed physics breakthroughs just by some curmudgeoningly arguing
Starting point is 00:36:26 that it's impossible to make this scientific. For example, extra solar planets. People were so stuck with this idea that all the other solar systems had to be like our solar system with a star and then some small rocky planets near it and some gas giants farther out. So they're like, yeah, no point in even looking around other stars because we can't see Earth-like planets.
Starting point is 00:36:51 Eventually, some folks decided to just look anyway with the Doppler method to see if stars were going in little circles because something was orbiting around. And they found these hot Jupiter's, like the gigantic Jupiter-sized thing going closer to the star than Mercury is going to our sun. Wow. But they could have done that 10 years earlier, you know,
Starting point is 00:37:12 if they hadn't been intimidated by these convergens who said, don't look. So my attitude is don't listen to the curmudgeon's. If you have an idea for an experiment you can build that's just going to cut into some new part of parameter space and experimentally test the kind of questions that have never been asked, just do it.
Starting point is 00:37:36 More than half the time when people have done that, there was a revolution, you know. When Carl Jansky wanted to build the first X-ray telescope and look at x-rays from the sky. For example, people said, what a loser. There are no x-rays coming from the sky. Do you think you're like dentists out there?
Starting point is 00:37:55 I don't know what. And then you found that there is a massive amount of x-rays, even coming from the sun, you know. Or people decided to look at them. Basically, whenever we were opened up another wavelength with telescopes, we've seen new phenomena we didn't even know existed. Or when Leuvenhook built the first microscope, up, do you think he expected to find these animals that were so tiny you couldn't see them
Starting point is 00:38:22 with a naked eye? Of course not, right? But he basically went, orders of magnitude in a new direction in an experimental parameter space. And there was a whole new world there, right? So this is what I think we should do with consciousness. And with intelligence, this is exactly what has. This is exactly what has happened. If we segue a little bit into that topic, I think there's too much pessimism in science. If you should go back, I don't know, 30,000 years ago, you know, if you and I were living in a cave, sitting and having this conversation, you know, we would probably have figured, well, you know, look at those little white dots in the sky here. They're pretty nifty. So we wouldn't have any Netflix to distract us with. And we would know that some of our friends
Starting point is 00:39:22 had come up with some cool myths for what these dots in the sky were. And, oh, look, that one, maybe it looks like an archer or whatever. But, you know, since you're a guy who likes to think hard, you'd probably have a little bit of a melancholy tinge that, you know, we're never really going to know what they are. You can't jump up and reach them. I can climb the highest tree and they're just as far away. And we're kind of stuck here on our planet, you know, and maybe we'll starve to death, you know, and 50,000 years from now if there's still people,
Starting point is 00:39:54 you know, life for them is going to be more or less like it is for hours. And boy, oh boy, would we have been too pessimistic, right? We hadn't realized. We had totally, that we were the masters of underestimation. Like, we massively underestimate and not only the size of what, existed, that everything we knew of was just a small part of this giant spinning ball, Earth, which was in turn just a small part of a grander structure of the solar system, part of a galaxy, part of a galaxy cluster, part of a supercluster, part of a universe,
Starting point is 00:40:30 maybe part of a hierarchy or parallel universes. But more importantly, we also underestimated the power of our own minds to figure stuff out. And we didn't even have to fly to the stars. to figure out what they were. We just really kind of have to let our minds fly. And, you know, Aristarchus of Samos over 2,000 years ago was looking at a lunar eclipse.
Starting point is 00:40:56 And some of his friends were probably like, oh, this moon turned red. It probably means we're all going to die or the gods and omen from the gods. And he's like, hmm, the moon is there. The sun just sat over there. So this is obviously Earth's shadow being cast on the moon. And actually, the edge of the shadow of Earth is not straight.
Starting point is 00:41:22 It's curved. Wait, so we're living on a curved thing. We're maybe living on a ball. Huh. And wait a minute, you know, the curvature of Earth's shadow there is clearly showing that Earth is much bigger than the moon is. And he went down and calculated how much bigger Earth is than the moon. And then he's like, okay, well, I know that.
Starting point is 00:41:43 I know Earth is about 40,000 kilometers because I read that Eratosthenes had figured that out and I know the moon I can cover it with my pinky so it's like half a degree in size so I can figure out how what the actual physical sizes of the moon. It was ideas like this that started like breaking this curse of just
Starting point is 00:42:08 of of overdone pessimism we started to believe in ourselves a little bit more and here we are now with the internet with artificial intelligence with all these
Starting point is 00:42:26 little things you can eat that prevent you from dying of pneumonia my grandfather Sven died of a stupid kidney infection could have been treated with penicill and it's amazing you know how how much excessive pessimism there's been and i think we still have a lot of it unfortunately uh that's why i want to come back to this thing that if someone has yeah there's no better way to fail at something to convince yourself that it's impossible you know
Starting point is 00:43:03 And look at AI, I would say whenever with science we have started to understand how something in nature works that we previously thought of as sort of magic, like what causes the winds or the seasons, et cetera, what causes things to move, we were able to historically transform that into technology that often did this better and could serve us more. So we figured out how we could build machines that were stronger than us and faster than us. We got the Industrial Revolution. Now we're figuring out that thinking is also a physical process, information processing, computation. And Alan Turing was, of course, one of the real pioneers in this field. And he...
Starting point is 00:43:58 He clearly realized that the brain is a biological computer. He didn't know how the brain worked. We still don't exactly. But it was very clear to him that we could probably build something in those much more intelligent and maybe more conscious too. Once we figured out more details, I would say from the 50s, when the term AI was coined, not far from here.
Starting point is 00:44:29 Dartmouth, the field has been chronically overhyped. Most progress has gone way slower than people predicted, even than McCarthy and Minsky predicted for that Dartmouth workshop and so on. But then something changed about four years ago when it went from being overhyped to being underhyped. Because I remember very vividly like seven years ago, six years ago, most of my colleagues here at MIT and most of my AI colleagues in general,
Starting point is 00:44:54 we're pretty convinced that we were decades away from passing the Turing test, decades away from building machines that could master language and knowledge at human level. And they were all wrong. They were way too pessimistic because it already happened. You can quibble about whether it happened with chat GPT4 or when it was exactly, but it's pretty clear it's in the past now. so if people could be so wrong about that maybe they were wrong about more and sure enough since then AI has gone from being kind of high school level to kind of college level
Starting point is 00:45:36 to in many areas being PhD level to professor level to even far beyond that in many areas in just four short years so so prediction after prediction has been crushed now where things have happened faster. So I think we have gone from the overhyped regime to the underhyped regime and this is of course the reason why so many people now are talking about maybe we'll reach broadly human level
Starting point is 00:46:01 and cup two years or five years depending on which tech CEO you talk to or which professor you talk to. But it's very hard now for me to find anyone serious who thinks we're 100 years away from it. And then
Starting point is 00:46:16 of course you have to think about, go back and reread your touring right so he said in 1951 that once we get machines that are vastly smarter than us in every way they can basically perform better than us on all cognitive tasks the default outcome is that they're going to take control and from there on earth will be run by them not by us just like we took over from from other apes and irving j good pointed out in the 60s that last sprint from being kind of roughly at our little bit
Starting point is 00:46:55 better than us to being way better than us can go very fast because as soon as we can replace human AI researchers by machines who don't have to sleep and eat and can think a hundred times faster and can copy all their knowledge to the others. You know,
Starting point is 00:47:09 every doubling in quality from then on might not take months or years like it is now, but the sort of human R&D time scale. it might happen every day or on the time scale of hours or something. And we would get this sigmoid ultimately where we shift away from the sort of slow exponential progress that technology has had ever since the dawn of civilization where you use today's technology to build tomorrow's technology,
Starting point is 00:47:39 which is so many percent better and to an exponential which goes much faster. First, because it's now humans are out of the loop, don't slow things down. out. And then eventually it plateaus into a sigmoid when it bumps up against the laws of physics. No matter how smart you are, you're probably not going to send information faster than light, and general relativity in quantum mechanics put limits and so on. But my colleague Seth Lloyd here at MIT has estimated they were still about the million, million, million, million, million times away from the limits from the laws of physics. So it can get pretty crazy pretty quickly and
Starting point is 00:48:18 it's also Alan I keep discovering more stuff as Stuart Russell dug out this fun quote from him in 1951 that I wasn't aware of before where he also talks about you know how what happens when we reach this threshold
Starting point is 00:48:34 and he he's like well don't worry about this control loss thing now because it's far away but I'll give you a test so you know at the pay attention, you know, the canary in the coal mine,
Starting point is 00:48:49 the touring test, as we call it now. And we already talked about how that was just passed. And this reminds me so much of what happened around 1942 when Enrique Fermi built the first self-staining nuclear chain reaction under the football stadium in Chicago, right?
Starting point is 00:49:10 That was like a touring test for nuclear weapons when the physicists who found out about this then they totally freaked out not because the reactor was at all dangerous it was pretty small you know it wasn't any more dangerous than chat GPT is but today but but because they realize oh that was the canary in the coal mine you know that was the the last big milestone we had no idea how to meet and the rest is just engineering about AI now, I think that we obviously don't have
Starting point is 00:49:49 AI that are better than us at, or as good at us at AI development, but it's mostly engineering, I think, from here on out. We can talk more about the nerdy details of how it might happen. It's not going to be large language models scaled. It's going to be other things.
Starting point is 00:50:09 Like, in 1942, Two, I'm curious, actually, if you were there visiting Fermi, you know, how many years would you predict it would have taken from then until the first nuclear explosion happened? How many years? Difficult to say maybe a decade. Uh-huh. So then it happened in three.
Starting point is 00:50:29 Could have been a decade. Probably got sped up a bit because of the geopolitical competition that was happening during World War II. And similarly, it's very hard to say now. Is it going to be three years? It's going to be a decade. But there's no shortage of competition, fueling it again. And as opposed to the nuclear situation, there's also a lot of money in it.
Starting point is 00:50:51 So I think this is the most interesting time, an interesting fork in the road in human history. And if Earth is the only place in our observable universe with telescopes, then it's, whether it's actually conscious us about the universe at large, this is probably the most interesting fork in the road in the last 13.8 billion years for our universe too
Starting point is 00:51:18 because there's so many different places this could go, right? And we have so much agency and steering in a good direction rather than a bad one. Here's a question I have when people talk about the AIs taking over.
Starting point is 00:51:32 I wonder, so which AIs? So is Claude considered a competitor to open AI in this AI space from the AI's perspective? Does it look at other models as you're an enemy because I want to self-preserve? Does Claude look at other instances so you have your own Claude chats? Are they all competitors? Is every time it generates a new token? Is that a new identity? So it looks at what's going to come next and before as, hey, I would like you to not exist
Starting point is 00:52:01 anymore because I want to exist. Like what is the continuing identity that would make us say that the AIs will take over? Like what is the AI there? Yeah, those are really great questions. I mean, the very short answer is people generally don't know. I'll say a few things. First of all, we don't know whether Claude or GPD5 or any of these other systems are having a subjective experience or not, whether they're conscious or not. Because as we talked about for a long time,
Starting point is 00:52:32 we do not have a consensus theory of what kind of information processing has a subjective experience, what consciousness is, that we don't need necessarily for machines to be conscious for them to be a threat to us. If you're chased by heat-taking missile, you know, you probably don't care whether it's conscious in some deep philosophical sense. You just care about what it's actually doing, what its goals are. And so let's shift, let's just switch to talking about just behavior of systems, you know, in physics, we typically think about the behavior as determined by the past through causality, right? Why did this phone fall down? Because gravity pulled on it, because there's an Earth planet down here. When you look at what people do, we usually instead
Starting point is 00:53:22 interpret, explain why they do in terms of not the past, but the future, that's some goal they're trying to accomplish. If you see someone scoring a beautiful goal in a soccer match, You could be like, yeah, it's because their foot struck the ball in this angle and therefore action equals reaction, blah, blah, blah, but more likely you're like, they wanted to win. And when we build technology, we usually build it with a purpose in it. So people build heat seeking missiles to shoot down aircraft. They have a goal. We build mousetraps to kill mice. And we train our AI systems today, our large language models, for example,
Starting point is 00:54:02 to make money and accomplish certain things. But to actually answer your question about what the system, if we would have a goal to see, to collaborate with other systems or destroy them or see them as competitors, you actually have to ask, what does the system actually have it? Is it meaningful to say that this AI system as a whole has a coherent goal? And that's very unclear, honestly. you could say at a very trivial level that chat GPT has the goal
Starting point is 00:54:40 to correctly predict the next token or word in a lot of texts because that's exactly, that's how we trained it, so-called pre-training. You know, you just let it read all the internet and look at and predict which words are going to come next. You let it look at pictures and predict what's more in them and so on. But clearly, they're able to. to have much more sophisticated goals than that.
Starting point is 00:55:06 Because it just turns out that in order to predict, like if you're just trying to predict my next word, it helps if you make a more detailed model about me as a person and what my actual goals are and what I'm trying to accomplish, right? So these AI systems have gotten very good at simulating people. Say this sounds like a Republican. And so if this Republican is writing about immigration, he's probably going to write this.
Starting point is 00:55:35 Based on what they wrote previously, they're probably a Democrat. So when they write about immigration, they're more likely to say these words. This one is, the Democrat is more likely to maybe use the word undocumented immigrant, whereas the Republican might predict they're going to say illegal alien, you know. So they're very good at predicting modeling people's goals. But does that mean they have the goals?
Starting point is 00:56:00 themselves? Now, if you're a really good actor, you're very good at modeling people with all sorts of different goals, but does that mean you have the goals, really? This is not a well-understood situation. And when companies spend a lot of money on what they call aligning in AI, which they bill as giving it good goals, what they are actually in practice doing is just affecting its behavior. So they basically punish it when it says things that they don't want it to say and encourage it.
Starting point is 00:56:40 And that's just like if you train a serial killer, you know, to not say anything that reveals his murder's desires. So I'm curious, if you do that and then the serial killer stops ever dropping any hints about wanting to knock someone off, you know, would you feel that you've actually changed this person's goals to not want to kill anyone?
Starting point is 00:57:00 Well, the difference in this case would be that the AI's goals seem to be extremely tied to its matching of whatever fitness function you give it, whereas in the serial killer case, their true goals are something else and their verbiage is something else. It seems like in the LLM, in LLM's cases. Yeah, but when you train an LLM, I'm talking about the pre-training now where they read the whole internet basically you're not telling it to be kind or anything like that you're just really training it to be have the goal of predicting and then in the so-called fine-tuning reinforcement learning for human feedback is the nerd phrase for it yes there you look at different answers that it could give and you say
Starting point is 00:57:43 I want this one not that one but you're again not explaining to it you know like I have a two-and-a-half year old I have a two-year-old son right this guy and you know my idea for how to make him a good person is to help him understand or the value of kindness. My approach to parenting is not to be mean to him if he ever kicks somebody without any explanation. I want him rather to internalize the goal of being a kind person and that he should value the well-being of others, right? And that's very different from how we do reinforcement learning with human feedback and it's frankly not at all we I would stick my neck out and say we have no clue really what if any goals chat GPT has it acts as if it has goals yeah you know
Starting point is 00:58:39 but if you kick a dog every time it tries to bite someone it's going to also act like it doesn't want to bite people but like who knows or the serial killer case it's quite possible that it doesn't have any particular set of unified goals at all so this is a very important thing to study and understand because if we're ever going to end up living with machines that are way smarter than us, right? Then our well-being depends on them having actual goals now to be the treat us well, not just having said the right buzzwords before they got the power. So we both have lived with entities that were more than us, our parents, when we were. We we were little, and it worked out fine, because they really had goals to be nice to us, right?
Starting point is 00:59:31 So we need some deeper, very fundamental understanding of the science of goals in AI systems. Right now, most people who say that they've aligned goals to AIs are just bullshitting, in my opinion. They haven't. They've aligned behavior, not goals. and I think I would encourage any physicists and mathematicians watching this who think about getting into AI to think
Starting point is 01:00:00 I would encourage them to think to consider this because physicists have one of the things that's great about physicists is physicists like you have a much higher bar on what they mean by understanding something than engineers typically do engineers will be like, well, yeah, it works.
Starting point is 01:00:23 Let's ship it. Whereas as a physicist, you might be like, but why exactly does it work? Can I actually go a little deeper? Is there some, can I write down an effective field theory for how the training dynamics works? Can I model this somehow? Can I, this is what the kind of thing that Hopfield did with memory.
Starting point is 01:00:46 This is the sort of thing that Jeff Hinton has done. And we need much more of this to have an actual satisfactory theory of intelligence, what it is, and of goals. If we actually have a system, an AI system that actually has goals, and there's some way for us to actually really know what they are, then we would be in a much better situation than we are today. We haven't solved the problems because this AI, if it's very loyal, might be owned by someone
Starting point is 01:01:17 who orders to do horrible things and so on, and program in horrible goals into it. But at least then we'll have the luxury of really talking about what goals AI systems should have. A great word was used, understand.
Starting point is 01:01:37 That's something I want to talk about. What does it mean to understand? Before we get to that, I want to linger on your grandson for a moment. So when you're training your, yes, your son. I have a grandson too, though, actually, he's also super cute. So when you're training your son, why is that not you're a human, you're giving feedback,
Starting point is 01:01:55 it's reinforcement, why is that not RLHF for the child? And then, well, you'd wonder, well, what is the pre-training stage? What if the pre-training stage was all of evolution, which would have just given rise to his nervous system by default? And now you're coming in with your RLHF and tuning not only his behavior, but his goals, simultaneously. So let's start with that second part. Yeah. So first of all,
Starting point is 01:02:22 the way RLHF actually works now is that American companies will pay one or two dollars an hour to a bunch of people in Kenya and Nigeria to sit and watch the most awful graphical images and horrible things and then they keep clicking on which of the different what you keep classifying them
Starting point is 01:02:43 and is this something that should be, is okay? or not okay and so on. It's nothing the way like the way anyone watching this podcast treats their child where they really try to help the child understand in a deep way
Starting point is 01:02:57 and want. Second, the actual architecture of the transformers and more scaffolding systems being built right now are very different from our limited understanding of how a child's brain works
Starting point is 01:03:14 so no we're certainly not um we can't just say the clip we're going to declare victory move on from this we just like i said before that people i think have used um some philosophical excuses to avoid working hard on the consciousness problem i think some people have made the philosophical excuses to avoid just asking this very sensible question of of goals before we talk about understanding, can I talk a little bit more about goals? Please. Yeah, because so if we talk about goal-oriented behavior first,
Starting point is 01:03:53 there's less emotional baggage associated with that, right? Let's define goal-oriented behavior as behavior that's more easily explained by the future than by the past, more easily explained by the effects that it is going to have than by what caused it, okay? Interesting. So again, to come back to this, If I just take this thesis here and I bang it, you know,
Starting point is 01:04:19 and you ask why did it move? Reading, playing, learning. Stellist lenses do more than just correct your child's vision. They slow down the progression of myopia. So your child can continue to discover all the world has to offer through their own eyes. Light the path to a brighter future with stellar lenses for myopia control. Learn more at SLO. and ask your family eye care professional for SLR Stellis lenses at your child's next visit.
Starting point is 01:04:51 You could say the cause of it moving was because I, another object, my hand bumped into it, action equals reaction, other words, this impulse given to it, et cetera, et cetera. Or you could say, but the goal oriented, you could view it as goal oriented behavior, thinking, well, Max wanted to illustrate a point, he wanted it to move. so he did something that made it move right and that feels like the more economic description in this case and it's interesting
Starting point is 01:05:20 even in basic physics we actually see stuff which can sometimes be more so first thing I want to say is there is no right and wrong description both of those descriptions are correct right
Starting point is 01:05:33 so look at the water in this bottle here again if you put a straw into it you know it's going to look bent because light rays bend when they cross the surface into the water, you can give two different kinds of explanations for this. The causal explanation will be like, well, the light ray came there,
Starting point is 01:05:52 there were some atoms in the water, and they interacted with the electromagnetic field and blah, blah, blah, blah. And after a very complicated calculation, you can calculate the angle that goes that way. But you can give a different explanation from us principle and say that actually the light ray took the path that was going to get it there
Starting point is 01:06:09 the fastest. if if this were instead a beach and this is an ocean and you're working a summer job as a lifeguard and you want to risk and you see a swimmer who's in trouble here how are you going to go to the swimmer you're going to again go in the path that gets you there the fastest so you'll run longer distance through the air on the beach and then a shorter distance through the water you know then clearly that's goal-oriented behavior right For us, for the photon, though? Well, both the descriptions are valid. It turns out in this case that it's actually simpler to calculate the right answer if you do from us principle and look at the goal-oriented way. And now, then we see in biology, so Jeremy England, who used to be a professor here, realized that in many cases non-equilibrium thermodynamics can also be understood
Starting point is 01:07:03 sometimes more simply through goal-oriented behavior. like if you, if you, um, no, suppose I put a bunch of sugar on the floor. And no life form ever enters the room, come back in the, and including the, you know, Phyllis, who keeps this nice and tidy here. Then it's still going to be there in a year, right? But if there are some ants, sugar's going to be gone pretty soon. And entropy will have increased faster that way, because the sugar was, eaten and there was dissipation.
Starting point is 01:07:41 And Jeremy England showed actually that there's a general principle in non-equilibrium thermodynamics where systems tend to adjust themselves to always be able to dissipate faster. To be able to, if you have a thing, if you have some, there's some kinds of liquids where you can put some stuff where if you shine light at one wavelength, it will rearrange itself so that it can absorb that wavelength better, dissipate the heat faster. And you can even think of life itself a little bit of that. Life basically can't reduce entropy, right, in the universe as a whole, and it can't beat the second law of thermodynamics.
Starting point is 01:08:19 But it has this trick where it can keep its own entropy low and do interesting things, retain its complexity and reproduce by increasing the entropy of its environment even faster. So if I understand the increasing of the entropy in the environment is the side effect, but the goal is to lower your own entropy. So, again, you can have, there are two ways of looking at it. You can look at it all as just a bunch of atoms bouncing around and causally explain everything.
Starting point is 01:08:51 But a more economic way of thinking about it is that, yeah, life is doing the same thing that liquid that rearranges itself to absorb sunlight. It's a process that just increases the overall entropy production in the universe, It makes the universe messier, faster, so that it can accomplish things itself. And since life can make copies of itself, of course, those systems that are most fit at doing that are going to take over everything. And you get an overall trend towards more life, more complexity, and here we are having a conversation, you know.
Starting point is 01:09:27 So where I'm going with this is to say that gold-oriented behavior, at a very basic level, was actually built into the laws of physics. That's why Rob Fremont's principle. There are two ways you can think of physics, either as the past causing the future, or as deliberate choices made now to cause a certain future. And gradually, our universe has become more and more goal-oriented, as we started getting more and more sophisticated life forms, now us. And we already are at the point,
Starting point is 01:10:02 a very interesting transition point now, where the amount of atoms that are in technology that we built with gold in mind is becoming comparable to the biomass already. And it might be if we end up
Starting point is 01:10:16 in some sort of AI-powered future where life starts spreading into the cosmos, near the speed of light, etc., that the vast majority of all the atoms are going to be engaged in goal-oriented behavior so that a universe is
Starting point is 01:10:34 becoming more and more goal-oriented. So I wanted to just anchor it a little bit in physics again, since you love physics, right? And say that I think it's very interesting for physicists to think more about the physics of goal-oriented behavior. And when you look at an AI system, oftentimes what plays the role of a goal is actually just a loss function or a reward function.
Starting point is 01:11:02 you have a lot of options and there's some sort of optimization trying to make the loss as small as possible. And anytime you have optimization, you'd say you have a goal? Yeah, but just like it's a very lame and banal goal for a light rate, refract a little bit
Starting point is 01:11:20 in water to get there as fast as possible, and that's a very sophisticated goal if someone tries to raise their daughter well or write a beautiful movie or symphony, there's a whole spectrum of goals but yeah building a system
Starting point is 01:11:39 that's trying to optimize something I would say absolutely is a goal-oriented system yeah I'm just going to inquire about are they equivalent so I see that whenever you're optimizing something you have to have a goal that you're optimizing towards sure but is it the case that any time you have a goal there's also optimization so anytime someone
Starting point is 01:11:54 uses the word goal you can think there's going to be optimization involved and vice versa that's a wonderful question actually Richard Feynman famously asked us question. He said that all lots of physics he knows about can actually be derived from an optimization principle, except one. And he wondered if there was one. So I think this is an interesting open question. You just threw it out there. I mean, look at you. I would expect that your actions cannot be really accurately modeled by writing down a single goal that you're just trying to maximize. I don't think that's how human beings in general operate.
Starting point is 01:12:32 What I think actually is happening with us and goals is a little different. You know, our genes, according to Darwin, the goal-oriented behavior they exhibited, right, even though they weren't conscious, obviously, our genes, was just evolutionary fitness, you know, make a lot of successful copies of themselves, right? That's all they cared about. so then it turned out that they would reproduce better if they also develop bodies around them with brains that could do a bunch of information processing
Starting point is 01:13:04 and get more food and mate and stuff like that. But evolution also became quite clear that if you have an organism like a rabbit if it would have to every time it was going to decide whether to eat this carrot or flirt with this girl rabbit, you know, go back and recalculate,
Starting point is 01:13:28 what is this going to do to my evolute? How many times the expected number of fertile offspring I have? That rabbit would just die of starvation and those genes would go out of the gene pool. They didn't have the cognitive capacity to always anchor every decision it made in one single goal. It was computationally unfeasible. It would always be running this actual optimization
Starting point is 01:13:51 that the genes cared about, right? So what happened instead in rabbits and humans and what we in computer science call agents of bounded rationality, where we have limits to how much we can compute, was we developed all these heuristic hacks. Like, well, if you feel hungry, eat. If you feel thirsty, you know, drink. If there's something that tastes sweet or savory, you know, eat more of it. fall in love, make babies, you know, these are clearly proxies ultimately for what the genes cared about,
Starting point is 01:14:28 you know, making copies of themselves because you're not going to have a lot of babies if you die of starvation, right? But now that you have your great brain, what it is actually doing is making its decisions based on all these juristics that themselves don't lead to, correspond to any unique goal. anymore like any person watching this podcast right who's ever used birth control would have so pissed off
Starting point is 01:14:54 their genes if the genes were conscious because this is not at all what the genes wanted right the genes just gave them the incentive to make love because the genes that would make copies of the genes the person who used birth control was well aware what the genes wanted and was like screw this i don't want to have a baby at this point in my life so there's been a rebellion in the goal-oriented behavior of people against the original goal that we were made with and replaced by these heuristics that we have our emotional drives and desires and hunger and thirst and et cetera
Starting point is 01:15:35 et cetera that are not anymore optimizing for anything specific and they can sometimes go work out pretty badly like the obesity epidemic and so on and I think the machines today, the smartest AI systems, are even more extreme like that than humans. I don't think they have anything. I think they're much more,
Starting point is 01:15:57 I think humans still tend to end up, especially the more for those who are, who like introspection and self-reflection, are much more prone and likely to have at least somewhat consistent strategy for their life or gold than... Chat GPT has. It's a completely random mishmash of all sorts of things.
Starting point is 01:16:24 Understanding. Understanding, yes. Oh, that's a big one. I've been toying with writing a paper called artificial understanding for quite a long time, as opposed to artificial consciousness and artificial intelligence. And the reason I haven't written it is because it's, It's a really tough question. I feel there is a way of defining understanding
Starting point is 01:16:53 so that it's quite different from both consciousness and intelligence, although also a kind of information processing, or at least a kind of information representation. I thought you were going to relate it to goals, because if I understand goals are related to intelligence, sure, but then also the understanding of someone else's goals seems to be related to intelligence
Starting point is 01:17:16 so for instance in chess you're constantly trying to figure out the goals of the opponent and if I can figure out your goals prior to you figuring out mine or ahead of yours or what happened then I'm more intelligent than you now you would think that
Starting point is 01:17:30 the ability to reliably achieve your goals is what is intelligence but it's not just that because you can have an extremely simple goal that you always achieve like the photon here it's just following some principle but we have goals even the person on the beach
Starting point is 01:17:43 with the swimming hypothetically even that we'll fail at but we're more intelligent than the photon so but we're able to model the photon's goal the photon's not able to model our goal so I thought you were going to say well that modeling is related to understand
Starting point is 01:18:00 yeah that I agree with it for sure modeling is absolutely related to understanding goals I view is different I personally think of intelligence as being rather independent of goals. So I would define intelligence as ability to accomplish goals. You know, you talked about chess, right? There are tournaments where computers play chess against computers to win.
Starting point is 01:18:30 Have you ever played losing chess? It's a game where you're trying to force the other person to win the game. No. They have computer tournaments for that too. Interesting. So you can actually give a computer the goal which is the exact opposite of a normal chess computer and then you can say that the one the losing chess tournament
Starting point is 01:18:48 is the most intelligent again. So this right there shows that being intelligent doesn't, it's not the same as having a particular goal. It's how good you are at accomplishing them, right? I think a lot of people also make the mistake of saying, oh, we shouldn't worry about what happens with powerful AI because it's going to be so smart, it'll be kind, dramatically to us, you know, if Hitler had been smarter, do you really think the world would
Starting point is 01:19:15 have been better? I would guess that it would have been worse, in fact, if he had been smarter and when World War II and so on. So there's, Nick Bostrom calls us the orthogonality thesis that intelligence is just an ability to accomplish whatever goals you give yourself or whatever goals you have. And I think understanding is a component of intelligence, which is very linked to modeling, as you said. Maybe you could argue that it even is the same, the ability to have a really good model of something,
Starting point is 01:19:57 another person, as you said, or of the universe, our universe, if you're a physicist, right? And I'm not going to give you some very glib definition of understanding or artificial understanding because I view it as an open problem but I can tell you one anecdote of something which felt like artificial understanding to me so me and some of my students here at MIT
Starting point is 01:20:23 we were very interested in so we've done a lot of work including this thesis here that randomly happens to be lying here is about how you take AI systems and you do something smart then you figure out how they do it. So one particular task we trained an AI system to do is just to implement,
Starting point is 01:20:48 to learn group operations abstractly. So a concrete example, suppose you have 59, the number is zero through 58, okay? And you're adding them modulo 59. So you say like one plus two is three, but 57 plus 3 is 60 well that's bigger than 59 so you subtract off 59 you say it's one same principle as clock the same exactly as a clock and I'm so glad you said clock because that's
Starting point is 01:21:21 your model in your brain about modular arithmetic you think of all the numbers sitting in a circle it goes after 10 and 11 comes 12 but then comes one so what happened was we there are 59 times 59s are about 3,600 pairs of numbers, right? We trained on the system on some fraction of those to see if we could learn to get the right answer. And the way the AI worked was it learned to embed and represent each number, which was given to it just as a symbol. It didn't know five, whether it had anything to do with the number five, as a point
Starting point is 01:21:58 in a high-dimensional space. So we have these 59 points in a high-dimensional space, okay? and then we trained another neural work to look at these representations. So you give it this point and this point, and it has to figure out, okay, what is this plus this mod 59? And then something shocking happened. You train it, train it, it sucks, it sucks, and then it starts getting better on the training data.
Starting point is 01:22:24 And then as a sudden point, it suddenly also starts to get better on the test data. So it starts to be able to correctly answer questions for pairs of numbers it hasn't seen yet. So it somehow had a eureka moment where it understood something about the problem. It had some understanding. So I suggested to my students,
Starting point is 01:22:43 why don't you look at what's happening to the geometry of all these points that are moving around, these 59 points that are moving around in this high dimensional space during the training? I told them to just do a very simple thing, principal component analysis where you try to see if they mostly lie in a plane
Starting point is 01:23:00 and then you can just plot that 59 points. And it was so cool what happened. You look at this. You see 59 points that's looking very random. They're moving around. And then at exactly the point when the Eureka moment happens, when the AI becomes able to answer questions it hasn't seen before, the points line up on a circle.
Starting point is 01:23:21 A beautiful circle. Interesting. Except not with 12 things, but with 59 things now because that was the problem it had. So to me, this felt like the AI had reached understanding, about what the problem was. It had come up with a model, or as we often call it, a representation of the problem.
Starting point is 01:23:42 In this case, in terms of some beautiful geometry, and this understanding now enabled it to see patterns in the problem so that it could generalize all sorts of things that it hadn't even come across before. So I'm not able to give a beautiful, succinct, fully complete answer to your question
Starting point is 01:24:07 of how to define artificial understanding but I do feel that this is an example a small example of understanding we've since then seen many others we wrote another paper where we found that when large language models do arithmetic they represent the numbers on a helix like a spiral shape
Starting point is 01:24:26 and I'm like what is that well the long direction of it can be thought of like representing the numbers in analog, like you're farther this way if the number is bigger. But by having them wrap around on a helix like this, you can use the digits if it's ten to go around. There were actually several helixes.
Starting point is 01:24:45 There's a hundred helix and a ten helix. So I suspect that one day people will come to realize that more broadly when machines understand stuff, and maybe when we understand things also, it has to do with coming up with the same patterns and then coming up with a clever way of representing the patterns such that the representation itself goes a long way towards already giving you the answers you need. This is how I'm a very visual thinker when I do physics
Starting point is 01:25:21 or when I think in general. I never feel I can understand anything unless I have some geometric image in my mind. Actually, Feynman talked about this. Feynman said that there's the story of him and a friend who can both count to 60, something like this, precisely. And then he's saying to his friend, I can't do it if you're waving your arms in front of me or distracting me like that.
Starting point is 01:25:44 I remember. But I can, if I'm listening to music, I can still do this trick. And he's like, I can't do it if I'm listening to music, but you can wave your arms as much as you like. And Feynman realized he, Feynman, was seeing the numbers, one, two, three. That was his trick, was to have a mental image as projected.
Starting point is 01:26:00 Yes. And then the other person was having a metronome. But the goal or the outcome was the same, but the way that they came about it was different. There's actually something in philosophy called the rule following paradox. So you probably know this. There are two rule following paradoxes. One is Kripki and one is the one that I'm about to say. So how do you know when you're teaching a child that they've actually followed the rules of arithmetic? So you can test them 50 plus 80, et cetera, 50 times 200. And they can get it correct every single time. They can even show you their reasoning, but then you don't know if that actually fails at 6,000 times 51 and the numbers above that. You don't know if they did some special convoluted method to get there. Exactly. All you can do is say you've worked it out in this case, in this case, in this case. That's actually, we have the advantage with computers
Starting point is 01:26:46 that we can inspect how they understand. In principle, but when you look under the hood of something like Chias-GPD, all you see is billions and billions of numbers, and you oftentimes have no idea what, when all these matrix multiplications and things like this, you have no idea really what it's doing. But mechanistic interpretability, of course, is exactly the quest to move beyond that
Starting point is 01:27:09 and see how does it actually work. And coming back to understanding and representations, there is this idea known as the platonic representation hypothesis that if you have two different machines, or I would generalize it to people also who both reach a deep understanding of something, there's a chance that they've come up with this similar representation. In Feynman's case, there were two different ones.
Starting point is 01:27:34 Right. But there probably aren't, they're probably, at most, there's probably a few ones that are one or a few that are really good. That seems like a hard case to make. But there's a lot of evidence coming out for it now, actually. You can, already many years ago, there was this team where they just took, you know, in chat, GPT and other AI systems, all the words. and word parts, they call tokens, get represented as points in a high-dimensional space.
Starting point is 01:28:06 And so this team, they just took something which had been trained only on English books and another one, English language stuff, and another one trained only on Italian stuff. And they just looked at these two point clouds and found that there was a way they could actually rotate them to match up as well as possible. And it gave them a somewhat decent English-Italian dictionary,
Starting point is 01:28:26 so they had the same representation. And there's a lot of recent papers, quite recent ones even, that are showing that, yeah, it seems like the representations of one that large language model like Chachyb-T, for example, is in many ways similar to the representations that other ones have. We did a paper, my student, my grad student, Don Juan Beck and I, where we looked at family trees. So we took the Kennedy family tree, a bunch of royalty family trees, et cetera. And we, this train the AI to correctly predict, like, who is the son of whom, who is the uncle of whom, is so-and-so a sister of whom. We just asked all these, we asked all these questions. And we also incentivize the large language model to learn it with as little, in as simple way as possible, by not giving it the arbitrary, by limiting the resources it had. And then when we looked inside, we discovered something amazing.
Starting point is 01:29:31 We discovered that, first of all, a whole bunch of independent systems had learned the same representation. So you could actually take the representation of one and literally just like rotate it around and stretch it a little bit and put it into the other and it would work there. And then when we looked at what it was, they were trees. We never told, we never told anything about family trees, but it would draw like, here is this king so-and-so, and then here are the sons and this and this. And then it could use that to know that, well, if someone is farther down, they're a descendant, et cetera, et cetera, et cetera. So that's yet another example, I think, in support of this platonic representation hypothesis, the idea that understanding probably has something, or often has something to do with capturing patterns in some, often in a beautiful geometric way, actually. So I wanted to end on the advice that you.
Starting point is 01:30:27 received from your parents, which was about don't concern yourself too much what other people think, something akin to that. It was differently worded, but I also wanted to talk about what are the misconceptions of your work that other colleagues even have, that you have to constantly dispel. And another topic I wanted to talk about was the mathematical universe. Oh, the easy stuff. So there are three, but we don't have time for all three. If you could think of a way to tie them all together, then feel free like a gymnast or juggler, but otherwise then I would like to end on the advice from your parents. Okay. Well, the whole reason I spent so many years thinking about whether we are all part of a mathematical structure and whether a universe actually is mathematical rather
Starting point is 01:31:12 than just described by it is, of course, because I listened to my parents because I got so much shit for that. And I just felt, yeah, I think I'm going to do this anyway, because to me it makes logical sense. I'm going to put the ideas out there. And then in terms of misconceptions about me, I think, one misconception I think is that somehow I don't believe that being falsifiable is important for science. I usually talked about earlier. I'm totally on board with this. And I actually argue that
Starting point is 01:31:55 if you have a predictive theory about anything, gravity, consciousness, et cetera, but means that you can falsify it. So that's one. And another one, probably the one I get most now because I've stuck my neck out a bit about AI and the idea that actually
Starting point is 01:32:15 the brain is a biological computer and actually we're likely to be able to build machines that we could totally lose control over, is that some people like to call me a doomer, which is, of course, just something they say when they ran out of arguments. It's like if you call someone a heretic or whatever. And so I think what I would like to correct about that
Starting point is 01:32:38 is I feel actually quite optimistic. I'm not a pessimistic person. I think that there's way too much pessimism floating around about humanity's potential. One is people, oh, we can never figure out and make any more progress on consciousness. We totally can. If we stop telling us that it's impossible and actually work hard,
Starting point is 01:33:05 some people say, oh, you know, we can never figure out more about the nature of time and so on unless we can detect gravitons or whatever. We totally can. There's so much progress that we can make if we're willing to work hard. And in particular, I think the most pernicious kind of pessimism we suffer from now is this meme
Starting point is 01:33:33 that it's inevitable that we are going to build superintelligence and become irrelevant. it is absolutely not inevitable but you know if you tell yourself that something is inevitable it's a self-fulfilling prophecy right this is convincing a country that's just been invaded that it's inevitable that they're going to lose the war if they fight it's the old it's siop game in town right so of course if there's someone who has a company and they want to build stuff and they don't want you to have any laws that make them accountable. They have an incentive to tell everybody,
Starting point is 01:34:15 oh, it's inevitable that this is going to get built. So don't fight it, you know. Oh, it's inevitable that humanity is going to lose control over the planet. So just don't fight it and, hey, buy my new product. It's absolutely not inevitable. You know, you could have had people, people say it's inevitable, for example, because they say people will,
Starting point is 01:34:38 always build technology that can give you money and power. That's just factually incorrect, you know. You're a really smart guy. If I could do cloning of you and, like, start selling a million copies of you on the black market, I could make a ton of money. We decided not to do that, right? They say, oh, if we don't do it, China's going to do it. Well, there was actually one guy who did human cloning in China, and you know what
Starting point is 01:35:06 happened to him? no he was sent to jail by the Chinese government oh okay people just didn't want that they thought we could lose control over the human germline and our species let's not do it so there is no human cloning happening now we could have gotten a lot of military power with bio weapons you know then professor Matthew messelson at Harvard said to Richard Nixon you know we don't want there to be a weapon of mass destruction that's so cheap that all our Saras can afford it.
Starting point is 01:35:38 And Nixon was like, huh, it makes sense, actually. And then Nixon used that argument on Brezhnev, and it worked, and we got a bio-weapons ban, and now people associate biology mostly with curing diseases, not with building bio-weapons. So it's absolutely not, it's absolute BS, this idea that
Starting point is 01:36:03 we're always going to build any technology that can give power or money to some people. If there's a... We have much more control over our lives and our futures. We have much more... We have much more control over our futures than some people like to tell us that we have. We are much more empowered than we thought.
Starting point is 01:36:24 I mentioned that if we were living in a cave 30,000 years ago, we might have made the same mistake and thought we were doomed to just always be at risk of getting eaten by tigers and starve to death. That was too pessimistic. We had the power to invent, through our thought,
Starting point is 01:36:42 develop a wonderful society and technology where we could flourish. And it's exactly the same way now. We have an enormous power. What most people actually want to make money on AI is not some kind of sand god that we don't know how to control. It's tools. AI tools.
Starting point is 01:36:58 People want to cure cancer. People want to make their business more efficient. Some people want to make their armies stronger and so on. you can do all of those things with tool AI that we can control. And this is something we work on in my group, actually. And that's what people really want. And there's a lot of people who do not want to just be like, okay, yeah, it's been a good run, you know,
Starting point is 01:37:21 hundreds of thousands of years. We had science and all that. But now let's just throw away the keys to Earth to some alien minds that we don't even understand what goals they have. Most Americans in polls think that's just a terrible idea. Republicans and Democrats, there was an open letter by evangelicals in the U.S. to Donald Trump saying, you know, we want AI tools.
Starting point is 01:37:41 We don't want some sort of uncontrollable superintelligence. Pope Leo has recently said he wants AI to be a tool, not some kind of master. You have people from Bernie Sanders to Marjorie Taylor Green. They've come out on Twitter, you know, saying we don't want SkyNet. We don't want to just make humans. economically obsolete. So it's not inevitable at all, I think. And if we can just, remember, we have so much agency in what we do, what kind of future
Starting point is 01:38:14 we're going to build. If we can be optimistic and just think through what is a really inspiring, globally shared vision for not just curing cancer, but all the other great stuff we can do, then we can totally collaborate and build that future. the audience member now is listening they're a researcher they're a young researcher they're an old researcher
Starting point is 01:38:38 they have something they would like to achieve that's extremely unlikely that's criticized by their colleagues for even them proposing it and it's nothing nefarious it's something that they find interesting and maybe beneficial to humanity whatever
Starting point is 01:38:49 what is your advice two pieces of advice first of all of order half of all the greatest breakthroughs in science were actually trash-talked at the time so just because someone
Starting point is 01:39:08 says that your idea is stupid doesn't mean it is stupid a lot of people ideas you should be willing to abandon your own ideas if you can see the flaw and you should listen to structured criticism against it but if you feel you
Starting point is 01:39:23 you really understand the logic of your ideas better than anyone else and it makes sense to you then keep pushing it forward you know and the second piece of advice I have is you might worry then like I did when I was in grad school that if I only worked on stuff that my colleagues thought was bullshit like thinking about the many world's interpretation
Starting point is 01:39:45 of quantum mechanics or multiverses from then my next job was going to be in McDonald's you know then my advice is um hadn't you just had your bets you know spend enough time working on things that gets appreciated by your peers now so that you can pay your bills so that your career continues ahead. But carve out a significant chunk of your time to do what you're really passionate about in parallel. If people don't get it, well, don't tell them about it at the time. You know, I, I, um,
Starting point is 01:40:27 from startup to growing and managing your business. That's why they have a dedicated small business advice hub on their website to provide tips and insights on business banking to entrepreneurs, no matter the stage of business you're in. Visit td.com slash small business advice to find out more or to match with a TD small business banking account manager. And that way you're doing science, for the only good reason,
Starting point is 01:40:59 which is that you're passionate about it. And it's a fair deal to society to then do a little bit of chores for society to pay your bills also. That's a great way of viewing it. And it's been quite shocking for me to see actually how many of the things that I got most criticized were
Starting point is 01:41:14 or was most afraid of talking openly about when I was a grad student, even papers that I didn't show my advisor until after he signed my PhD thesis and stuff, have later actually I'm pretty picked up and I actually feel
Starting point is 01:41:30 the things that I feel have been most impactful were generally in that category you're never going to be the first to do something important if you're just following everybody else Max thank you
Starting point is 01:41:44 thank you hi there Kurt here if you'd like more content from theories of everything and the very best listening experience then be sure to check out my substack at kirtjymongle.org.
Starting point is 01:42:00 Some of the top perks are that every week you get brand new episodes ahead of time. You also get bonus written content exclusively for our members. That's c-U-R-T-J-A-I-M-U-N-G-A-L.org. You can also just search my name and the word substack on Google. Since I started that substack, it somehow already became number two. in the science category. Now, substack for those who are unfamiliar is like a newsletter, one that's beautifully formatted, there's zero spam, this is the best place to follow the content of this channel that isn't anywhere else. It's not on YouTube, it's not on Patreon, it's exclusive to the
Starting point is 01:42:44 substack, it's free, there are ways for you to support me on substack if you want, and you'll get special bonuses if you do. Several people ask me like, hey, Kurt, you've spoken to so many people in the fields of theoretical physics, of philosophy, of consciousness. What are your thoughts, man? Well, while I remain impartial in interviews, this substack is a way to peer into my present deliberations on these topics. And it's the perfect way to support me directly. Kurtjymungle.org or search Kurtzimungle substack on Google. Oh, and I've received several messages, emails, and comments from professors and researchers saying that they recommend theories of everything to their students.
Starting point is 01:43:32 That's fantastic. If you're a professor or a lecturer or what have you, and there's a particular standout episode that students can benefit from or your friends, please do share. And of course, a huge thank you to our advertising. sponsor The Economist. Visit Economist.com slash Toe toe to get a massive discount on their annual subscription. I subscribe to The Economist and you'll love it as well. Toe is actually the only podcast that they currently partner with. So it's a huge honor for me and for you, you're getting
Starting point is 01:44:06 an exclusive discount. That's economist.com slash tow, T-O-E. And finally, you should know this podcast is on iTunes, it's on Spotify, it's on all the audio platforms. All you have to do is type in theories of everything and you'll find it. I know my last name is complicated, so maybe you don't want to type in Jai Mungle, but you can type in theories of everything and you'll find it. Personally, I gain from re-watching lectures and podcasts. I also read in the comment that Toll listeners also gain from replaying. So how about instead you relisten on one of those platforms like iTunes, Spotify, Google Podcasts, whatever podcast catcher you use, I'm there with you. Thank you for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.