Theories of Everything with Curt Jaimungal - Founder of Cellular Automata Unifies Biology, Computation, & Physics

Episode Date: September 10, 2025

Get 50% off Claude Pro, including access to Claude Code, at http://claude.ai/theoriesofeverything As a listener of TOE you can get a special 20% off discount to The Economist and all it has to offer!... Visit https://www.economist.com/toe In this episode, I speak with Stephen Wolfram—creator of Mathematica and Wolfram Language—about a “new kind of science” that treats the universe as computation. We explore computational irreducibility, discrete space, multi-way systems, and how the observer shapes the laws we perceive—from the second law of thermodynamics to quantum mechanics. Wolfram reframes Feynman diagrams as causal structures, connects evolution and modern AI through coarse fitness and assembled “lumps” of computation, and sketches a nascent theory of biology as bulk orchestration. We also discuss what makes science good: new tools, ruthless visualization, respect for history, and a field he calls “ruliology”—the study of simple rules, where anyone can still make real contributions. This is basically a documentary akin to The Life and Times of Stephen Wolfram. I hope you enjoy it. Join My New Substack (Personal Writings): https://curtjaimungal.substack.com Listen on Spotify: https://open.spotify.com/show/4gL14b92xAErofYQA7bU4e SUPPORT: - Become a YouTube Member (Early Access Videos): https://www.youtube.com/channel/UCdWIQh9DGG6uhJk8eyIFl1w/join - Support me on Patreon: https://patreon.com/curtjaimungal - Support me on Crypto: https://commerce.coinbase.com/checkout/de803625-87d3-4300-ab6d-85d4258834a9 - Support me on PayPal: https://www.paypal.com/donate?hosted_button_id=XUBHNMFXUX5S4 SOCIALS: - Twitter: https://twitter.com/TOEwithCurt - Discord Invite: https://discord.com/invite/kBcnfNVwqs Guests do not pay to appear. Theories of Everything receives revenue solely from viewer donations, platform ads, and clearly labelled sponsors; no guest or associated entity has ever given compensation, directly or through intermediaries. Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 Reading, playing, learning. Stellist lenses do more than just correct your child's vision. They slow down the progression of myopia. So your child can continue to discover all the world has to offer through their own eyes. Light the path to a brighter future with stellus lenses for myopia control. Learn more at SLOR.com. And ask your family eye care professional for SLR Stellist lenses at your child's next visit. I'm shocked.
Starting point is 00:00:29 Oh my gosh. something that completely violates the intuition that I've always had. It's something that I said couldn't happen. From discrete space to Darwinian evolution, to entropy and the second law, Stephen Wolfram's computational view of the universe makes claims about all of these in a unified fashion. Today's episode is a treat. If you're a fan of this channel theories of everything, then you're likely someone who enjoys surveying large swaths of lessons from disparate fields, attempting to see how they all relate and integrate. Same with me, Kurt Jymungle.
Starting point is 00:01:02 Now, today, Stephen Wolfram outlines how this polymathic disposition has helped him solve to his satisfaction, some of the major outstanding problems in fields as diverse as computer science, fundamental physics, and biology. This is a journey through his life in science, where I tease out the lessons he's learned throughout his career and how you can apply them yourself if you also want to make contributions. I was honored to have been invited to the Augmentation Lab Summit,
Starting point is 00:01:29 a weekend of events at MIT last month, hosted by MIT researcher Dunya Baradari, the summit featured talks on the future of biological and artificial intelligence, brain computer interfaces, and included speakers such as the aforementioned Stephen Wolfrum and Andreas Gomez-Emelson, subscribe to the channel to see the upcoming talks.
Starting point is 00:01:49 Stephen, welcome. Thank you. It's a pleasure. How does one do good sense? It's an interesting question. I mean, I've been lucky enough to have done some science, I think it's fairly interesting over the course of years. And I wonder, how does this happen? And I look at kind of other people doing science and I say, how could they do better science? You know, I think the first thing to understand is when does good science get done? And the typical pattern is some new tools, some new methodology gets developed, maybe some new paradigm, some new way of thinking about things. And then there's a a period when there's low-hanging fruit to be pricked. Lasts, maybe five years, maybe 10 years, maybe a few decades. And then some field of science gets established, it gets a name, and then there's a long grind for the next hundred years or something that people are doing,
Starting point is 00:02:40 sort of making incremental progress in that area. And then maybe some new methodology gets invented, things live and up again, and one has the opportunity to do things in that period. I have been lucky in my life, because I've kind of alternated between developing technology, and doing science, maybe about five times in my life. And that cycle has been very healthy. It wasn't intentional, but it's been worked out really well because I've spent a bunch of time developing tools that I've then been able to use to do science.
Starting point is 00:03:10 The science shows me things about how to develop more tools, and the cycle goes on, so to speak. So I've kind of had the opportunity to be sort of have first dibs on a whole bunch of new tools because I made them, so to speak. And that's let me do a bunch of things in science. that have been exciting and fun to do. I mean, I think a bunch of science I've done, I was realizing recently that it's also a consequence
Starting point is 00:03:32 of sort of a paradigmatic change, this idea taking the idea of computation seriously. And by computation, the fundamental thing I mean is you're specifying rules for something, and then you're letting those rules run, rather than saying, I'm going to understand the whole thing at the beginning. It's kind of a more starting from the foundation's point of view.
Starting point is 00:03:53 of view. Well, what I realized, actually very recently, and it's always surprising how long it takes one to realize these somewhat obvious features of history of science, even one's own history, is that, you know, I've been working on a bunch of things in fundamental physics and foundations of mathematics, foundations of biology, a bunch of other areas where I'm looking at the foundations of things, using a bunch of the same kinds of ideas, the same kinds of paradigms. And I'm realizing that a bunch of what I'm doing is kind of following on from what people did about 100 years ago, maybe sometimes a little bit more than 100 years ago. And I was wondering, why is it that, you know, a bunch of things I'm interested in, I'm going
Starting point is 00:04:33 back and looking at what people did 100 years ago, I'm saying they got stuck, I think we can now make progress. What happened? I think what happened is that in the 1800s, there was this kind of push towards abstraction. There was this idea that you could make formal versions of things. It happened most notably in mathematics, where kind of the idea emerged that sort of mathematics, was just a formal game where you're defining axioms and then seeing their consequences. It wasn't a thing about actual triangles
Starting point is 00:05:01 or whatever else. It was an abstract exercise. And once people had ground things down to that level of abstraction, same kind of thing happened with atoms and so on in the structure of physics. Once people had ground things down to that sort of deconstructed level,
Starting point is 00:05:17 they had something where they didn't know what to do next. Because what turns out to be the case is that that's sort of the setup for computation is you've ground things down to these kind of elementary primitives and then computation takes over and that's the thing that uses those primitives to do whatever is going to happen.
Starting point is 00:05:36 And so I think a lot of what got stuck then was closely related to this phenomenon of computational reducibility that I studied for the last 40 years or so that has to do with even though you know the rules it will not necessarily be the case that you can kind of jump ahead and say what will happen. You may just have to follow those rules step by step and see what happens.
Starting point is 00:05:58 You may not be able to make the big theory that sort of encompasses that describes everything that happens. And I think what happened in a bunch of these fields is people kind of ground things down to the primitives. Then they effectively discovered computational irreducibility, implicitly discovered it, by the fact that they couldn't make progress. And I think that, and things like Girdle's theorem, which are reflections of computational irreducibility, were kind of the other signs, more direct signs of that phenomenon. But the end result was people got to these primitives and then they couldn't get any further. And now that we actually understand something about the kind of computational paradigm,
Starting point is 00:06:39 we can see to what extent you can get further and the kinds of things that you can now say. So it's kind of interesting to me to see that. I mean, one particular area where I've only learned the history recently, and I'm kind of shocked that I didn't know the history earlier, is about the discreetness of space. So, you know, back in antiquity, people were arguing back and forth, you know, is the universe continuous or discrete? You know, are there atoms?
Starting point is 00:07:02 Does everything flow? And nobody knew. End of the 19th century, finally one had evidence that, yes, matter is made of molecules and atoms and so on, matter as discreet. Then, first decade of the 20th century became clear you could think of light as discreet. That had been another debate for a long time. And at that time, first few decades of the 20th century, most physicists were sure this whole discreetness thing was going to go all the way, that everything was going to turn out to be
Starting point is 00:07:31 discrete, including space. I didn't know that because they published very little about this. And the reason was they couldn't make it work. After relativity came in, it was like, how to do you? we make something that is like space, but is capable of working like relativity says space should work. And it was also somewhat confused by the idea of space time and the similarities between time and space, which were more mathematical than physical, really. That kind of confused the story. But I think then the thing that became clear was, I think, 1930 was, I think, the time
Starting point is 00:08:09 when particularly Werner Heisenberg was, I think, one of the ones who was really, you know, space is discreet. He had some, I think, I need to go look at his archives and things, but I think he had some kind of discrete cellular model of space, and he couldn't really make it work. And then
Starting point is 00:08:25 eventually he said, forget about all of this. I'm just going to think about processes in physics as being these particles come in, something happens in the middle, but we're not going to talk about that. We're just going to say it's an S matrix, and then things go out. And so that's when he started just saying, I'm going to look at the S matrix, this thing that just says how what goes in is related to what comes out.
Starting point is 00:08:46 I'm not going to talk about what's in the middle because I got really stuck thinking about sort of the ontology of what's in the middle. So then after that, you know, the sort of quantum field theory and quantum mechanics and so on started working pretty well. People forgot about this idea of discrete space. And in fact, the methods that they had would not have allowed them to have much interesting to say at that time. And finally, through sort of series of events that are kind of mildly interesting in my own life, I kind of came to realize how you could think about that in computational terms and how you could actually make all of that work. One of the cautionary tales for me is this question about, is matter even discreet or continuous?
Starting point is 00:09:30 The people had argued about people like Ludwig Boltzmann had gotten, you know, had sort of said towards the end of the 19th century, he kind of said, He believed very much in the atomic theory of matter. He was like, nobody else believes this. He said, you know, I'm going to write down what I have to say about this. He says, one man can't turn back the tide of history. I'm going to write down what I know about this so that when eventually this is rediscovered, when eventually people realize this is the right idea, they won't have to rediscover everything.
Starting point is 00:10:00 Well, in fact, it's kind of a shame because in 1827, I think, a chap called Robert Brown, who was a botanist, had observed this little pollen grains being kicked discreetly when they were on water or something, and that it was realized eventually that brown in motion is direct evidence for the existence of molecules. So if Boltzmann had known the botany literature, he would have known that, in fact, there was evidence for molecules that existed. Just those connections were not made. And so for me, that's a kind of cautionary tale because in modern times, you know, I think about, what does it take to kind of find experimental implications of the kinds of theories that I've worked on? And, you know, some of those things are difficult.
Starting point is 00:10:49 And it's like, well, you know, it might be 200 years until we can do this kind of investigation, or it might cost, you know, $10 billion to set up this giant space-based thing. But it might also be the case that, in fact, you know, somebody in 1972 observed, exactly a phenomenon that is the thing that would be what I'm looking for. And, you know, in fact, one of the things that's kind of ironic, and I've seen this a bunch of times in my career in science, is that when people don't kind of have a theory that says how an experiment should come out, and they do the experiment, and the experiment comes out differently from the way they expected, they say, I'm not going to publish this. This must be
Starting point is 00:11:35 wrong. And so a lot of things which later on one might realize were, you know, when you have a different theory, I might realize, gosh, you know, that experiment should have come out a different way. It got hidden in the literature, so to speak, which is, you know, this is a feature of the kind of institutional structure of science, but it's something where, you know, if one's lucky, somebody would have said, sort of done the honest experiment and just said, this is what we found. And, you know, even though this doesn't agree with the theory that so far we understand, so to speak. And so I'm actually been using LLMs quite a bit to try and do this kind of thematic searching of the scientific literature to try and figure out whether you could save the $10 billion, not do the experiment now, but just use the thing that already got figured out. But, you know, I think in terms of, I don't know, my efforts to do science, as I was saying,
Starting point is 00:12:29 I think one of the things that is sort of a critical feature is methodology and when new methodologies open up. And sort of I've been kind of lucky to be alive at a time when kind of computation and computers first make it possible to do kind of experiments with computers, so to speak. and, you know, that said, I built lots of tooling to be able to do that. But I think the thing that's always interesting about doing computer experiments, and particularly this is a consequence in the end of this whole computational irreducibility idea, is almost every computer experiment I ever do comes out in a way that I didn't expect. In other words, I'll do something, been doing some things even the last week or so, on particular domain, where, you know, I've got some theory about how,
Starting point is 00:13:17 it will come out. And if I didn't have any theory about how it would come out, I wouldn't do the experiment. First point is, the experiment has to be easy enough for me to do that on a whim I can do the experiment. It can't be the case that I've got to spend a month figuring out to do the experiment, because then I've got to be really sure about why it's worth doing. You know, it's got to be something where the tooling is such that I can do the experiment easily. I happen to have spent the last 40 years building that tooling. And, you know, it's available to everybody else in the world, too. But, you know, I'm going to be able to, I'm, as far as I'm concerned, I'm the number one user of, you know, Wolfram Language and so on, as far as I'm concerned.
Starting point is 00:13:54 I doubt I'm the person who uses it the most of everybody out there, but I'm the user I care about the most, so to speak. And, you know, it's a very nice thing to be able to take an idea that I have and be able to quickly translate that into something where I can, you know, make it real and do an experiment on it. So I have to have some idea about what's going to happen. others wouldn't do the experiment. But then I do it, and the thing I typically say to people who work with me is we have to understand the computational animals are always smarter than we are. So, you know, do the experiment, and you find things that you never expected to find. Why, you give an example, a specific one? Yeah, I mean, so, let's see, I was looking at almost any kind of simple program.
Starting point is 00:14:42 the question is, what kind of thing can it do? And so, for instance, let's say, looking at simple programs that rewrite networks. You run the network rewrite thing, what's it going to do? Well, sometimes it builds this elaborate geometrical structure. I had no idea it was going to do that. I thought it was going to make this messy kind of sort of random looking thing. But no, actually it builds this very sort of organized, elegant structure in some particular cases. Or, for example, I've been looking most recently, as it happens, at Lambda calculus,
Starting point is 00:15:23 very old model of computation that I happen to have never studied before, but have a particular reason to be interested in right now. And it's a question of, okay, I'll give you an example that happened to me just a few days ago. So there's a very simple specification, simple program. the program will run for a while. They're different versions of this program. You can enumerate lots of different possibilities. It will typically run for a while and, well, usually it will stop.
Starting point is 00:15:50 Sometimes it won't stop. It will just keep going. Sometimes it will go on repetitively and so on. And so I was looking at a bunch of these things and I was wondering, you know, what's the maximum lifetime of one that eventually stops? So I studied a bunch of different cases and there was, you know, I thought, okay, I found it. it's a lifetime, I don't know what, a few hundred or something. And I could find that with some simple experiments.
Starting point is 00:16:13 But then there was one where I started looking at it and I started looking at pictures of it and I'm kind of a seasoned hunter at this point for these kinds of things. And there was something about it that didn't seem quite right. There was something about one that I thought was going to just go on forever, but it seemed like it was doing things
Starting point is 00:16:30 that might not allow it to go on forever. So I pushed it a bit harder, push it harder. Sorry, what do you mean by you pushed it harder? Push the program harder. I let it run for longer. I let it run overnight on a network of computers. I mean, that's in practice what I did. And then come back in the morning and, oops,
Starting point is 00:16:50 it ran for, I don't know, some tens of thousands of steps and then stopped. Never expected that. And then there was another one where I, having seen that particular phenomenon, there was another one where I kind of suspected this one is going to stop. And that one stops after some, well, let's see, that one I kind of found a method for figuring out how it will develop. And that one, I think, stops after a few billion steps. So these are things you just, you don't expect. And, you know, this is very typical of what happens in this computational universe.
Starting point is 00:17:28 It's a bit similar to the physical universe. There are things that happen in the physical universe that you don't expect. But it's particularly sort of in your face when you can see, you know, in the physical universe, you don't necessarily know what the underlying rules for something are. So you could always be wondering, you know, do I just not know enough about how snowflakes form or something? But in this case, you know the rules. You know exactly what went in. And yet, you'll sort of have this forced humility of realizing you're not going to be able to figure out what happens. And sometimes you'll be quite wrong in your guess about what's going to happen.
Starting point is 00:17:58 So, you know, one of the principles in doing that kind of science is, you know, So you'll do the experiments, they'll often come out in ways you don't expect. You kind of just have to let the chips fall where they fall, which is something in doing science that can be psychologically very difficult. You know, you had to have had some kind of theory that caused you to start doing the experiment. Otherwise, you wouldn't have done the experiment. And so there's something of a psychological pressure to say, look, I had this theory. This theory has to be correct.
Starting point is 00:18:30 You know, something went wrong with my experiment. You know, let me tweak my experiment. Let me ignore that part of the experiment or something, because I'm sure this theory must be correct. One of the things that, you know, is an important thing that I kind of learned long, long, long ago now is just let the damn chips fall where they'll fall. And turns out one of the things that's happened to me is sometimes some of the, sometimes these chips fall in places that very much violate various kinds of prejudices that I have. and it's just like I'm more interested in where the chips actually fall than in supporting some prejudice that I have and as it's turned out in the end what's happened is
Starting point is 00:19:11 sometimes several years later I realize actually the way those chips fell was more consistent with my prejudice than I could ever have imagined so I'll give you an example so you know a lot of things I've done have been sort of deeply deconstructive of the way the universe works that is, they're very non-human interpretations of what goes on in the universe and so on. They're very, you know, the universe is some giant hypergraph of things. It's a very kind of humanly meaningless object. It doesn't have sort of a resonance with kind of our human sensibilities and so on.
Starting point is 00:19:48 It's deeply abstract, sort of deeply deconstructed in a sense. And yet, as a person, I'm quite a people enthusiast. You know, I like people. I find people interesting. I work with people. I've been a company full of people and so on. And so for me, it was always something of a conflict that on the one hand, I'm interested in people. On the other hand, the things I'm doing in science are deeply deconstructive of anything sort of human about what's going on in science. And that was the situation in my world for a couple of decades. And then I realized more recently that the nature of the observer is actually critical. In the end, in the end, sort of the ideas about the Ruliad and so on, the kind of entangled limit of all possible computations, that's sort of the ultimately deconstructed, dehumanized thing. But what you then realized, what I realized eventually, is how our perception of the laws of physics depends critically on our nature as observers within the Ruliad.
Starting point is 00:20:48 In other words, from going from a completely dehumanized view of science, that is this totally abstracted Ruliad, turns out. the humans are actually really important in giving us the science that we have. So even though I was sort of unhappy in some sense of some sort of psychological prejudice, I was unhappy with the idea that everything is deeply dehumanized because I kind of like humans, in the end, a couple of decades later, I realized actually the humans are kind of much more at the center of things than I had ever expected. So that was kind of an interesting realization.
Starting point is 00:21:25 Another one like that that I resisted for a long time is these things I call multi-way systems, which I had invented back in the early 90s and which I had thought by the late 90s that was a possible view of how quantum mechanics might work that there are these kind of many paths of history and that are being followed by the universe and so on. I really resisted that idea
Starting point is 00:21:47 because I felt sort of egotistically that I didn't want it to be the case that there were all these different possible paths of history, and the one I was experiencing was just one of those paths, or something like that. That was the assumption that I had made about what the implications of this idea of multi-way systems would be. What I realized when we did the physics project in 2019 was that, in fact, that isn't the right picture. That the idea of multi-way systems and the idea that there are these many paths of history, that's the right story. But the thing to realize is we are embedded as observers in this universe that is branching all the time.
Starting point is 00:22:25 And the critical point, then, is that we are branching as well. So from this idea, from this first sort of naive idea, that when you have something where you have many branches of history, that our sort of our experience must just go down one branch, that's really not the right picture. Actually, there are a couple of issues. One is that the branches can merge, and the other is that our experience can span many branches. We are extended objects. Our minds are extended objects in this branchial space and the space of these possible branches. I didn't realize that until 2019.
Starting point is 00:23:03 And that makes it, that means that my sort of ignoring multi-way systems for close to 30 years was a piece of sort of incorrect prejudice. And I was kind of lucky enough that eventually kind of started thinking about, look, might as well try and take this seriously. and see what its consequences actually are. It was something, actually, Jonathan Gorard was one of the people who's like, you should take these more seriously. I'm not sure he saw what the outcome would be,
Starting point is 00:23:33 but that was a, you know, why are you resisting this so much? It's perhaps a, so that was, but so it's a, it's always an interesting thing. When you have kind of this sort of, you have to have a belief about how things are going to work, Otherwise, you don't even look there, but you have to just believe the experiments you're doing or the things you figure out. I mean, for me, if I was doing, I mean, back in the day when I was long ago, when I was sort of first doing physics, I worked a bunch with Dick Feynman, who was a physicist who one of his great strengths was, he was a really good human calculator. And I can't do that. I'm a good computer calculator, but not a good human calculator. I built these computer tools
Starting point is 00:24:22 because I wasn't a very good human calculator. But Dick Feynman was really good at doing these calculations and getting to the right answer. And then he would go back and say, he didn't think anybody would be impressed by the fact that he got to the right answer by doing this complicated calculation.
Starting point is 00:24:36 He thought people would only be impressed if he could come up with this really simple kind of intuitive explanation of what was going on, which he often managed to come up with. Then he would throw away the calculation. Never tell anybody about the calculation. And everybody would be like,
Starting point is 00:24:48 how did you possibly figure out this kind of intuitive thing? And they'd all think, oh, it must be simple to come out with this intuitive thing. It wasn't simple. It was the result of some long calculation, which he didn't think anybody would be impressed with because he found it easy to do those things. You know, I don't, you know, for me, the only kind of thing where I know, I know what I'm talking about
Starting point is 00:25:08 is I do a computer experiment. It comes out in a certain way. The computer does what the computer does. There's no kind of sort of, I might have made a mistake somewhere type situation. I mean, I think if I look at the kinds of things that I've tended to do in science, they sort of mix what one might think of as kind of philosophy and what one might think of as this kind of very detailed, kind of solid computational experiments and so on.
Starting point is 00:25:34 I mean, that turns out to be, for me, has been sort of a powerful methodology for dealing with things. To go from, on the one side, sort of a general, almost philosophical understanding of how things might work. and then sort of the challenges to be able to sort of think computationally, fluently enough that you can go from that sort of philosophical understanding to say, okay, here's the program I should run that is a manifestation of that philosophical understanding
Starting point is 00:26:02 and then let's see what it actually does. And then I don't have to worry, am I getting it wrong because the program just does what the program does. And it's kind of, you know, I find it kind of charming when my new kind of science book came out in 2002. people saying, but it's wrong. It's like, what does that mean? What about it was wrong?
Starting point is 00:26:24 I don't know, but I mean, it's kind of people assume that you could sort of have got the wrong answer by doing the wrong calculation or something. But this is the nature of computer experiments. You just, you know, you specify the rule, you run the program, the program does what the program does. There's no, you know, no humans are involved. No possibility of error exists, so to speak. you can be wrong in the interpretation of what's happening. You can be wrong in the belief that what's happening in the computer experiment is relevant to something else.
Starting point is 00:26:54 But the actual experiment itself, it just is what it is. Now, you can be confused, I will say. And the number one source of confusion is when people don't look at everything that happens in the experiment. So people say, as a certain tendency in science, people have had this idea that, you know, being scientific is about generating numbers. And so one quite common type of mistake is to say, well, you know, there's a lot of detailed stuff going on underneath,
Starting point is 00:27:25 but I'm just going to plot this one curve as the result. And that means that you don't really get to see sort of the detail of what's happening. You're just seeing this one sort of summary, and sometimes that one summary can be utterly confusing. It can just lead you into kind of the thinking the wrong thing. And so for me, you know, being able to have sort of the highest bandwidth thing that I think we have to kind of understand what's going on is our visual system and being able to sort of visualize what's happening in as much detail as possible, I've always found very important. And often when I do projects, in fact, I just got bitten with this in the very latest project that I was doing. I always try to make sure that I front the effort to make the best possible visualizations.
Starting point is 00:28:10 Interesting. Because if, you know, the thing that one does that's a mistake, is to do the project with kind of crummy visualizations and then say, and now I'm going to present it and I'm going to make a really good visualization. Then you do the really good visualization and then you're like, oh gosh, there's something I can now see
Starting point is 00:28:26 that I didn't see while I was doing the project. So I was just a little bit bitten with that because there was a particularly complicated kind of visualization that I didn't go to the effort to make quite early enough in the project that I'm currently doing. And so I'm just having to redo a bunch of things because I realize that I can understand them
Starting point is 00:28:41 much more clearly using this sort of more sophisticated visualization technique. But that's a, you know, that's kind of, you know, you have to, that's just one of these things when you kind of start sort of thinking computationally about things, this idea that you can see as deep into the computation as possible is important, rather than saying all I care about here is this thing of plotting this one curve because that's what scientists have done for the last, you know, a couple of hundred years. You know, I think one of the things I realized only very recently about my own personal sort of scientific journey is back in the early 80s, I started doing a bunch of computer experiments, visualizing sort of the computations that were going on, figuring things out from that.
Starting point is 00:29:26 And for me, doing that in 1981 or something like that was completely obvious. It was like, how could you ever not think about doing something like that? But I was, the question was, why was that obvious to me? And, you know, it turns out what I had been doing for several years previously was building my first big computer system, which was a system for doing algebraic computation, symbolic algebraic computation. And I had gotten into doing that because I was doing particle physics. In particle physics, one of the things you get to spend a lot of time doing is computing Feynman diagrams. Feynman diagrams are this way of working out, well, actually this S-Matrix thing that I'd mentioned earlier, a particular way of doing that. sort of the best way we know to do it. I have to say, as a sort of footnote to this, Dick Feinman
Starting point is 00:30:13 always used to say about Feynman diagrams, they're kind of the stupidest way to do this kind of calculation. There's got to be a better way, he said. I remember one time telling him, if you work out there's sort of a series that you generate a Feynman diagrams, and I think at the Kth order in these Feynman diagrams, I'd worked out that the computational complexity of doing these Feynman diagrams went like k factorial to the fifth power. So as you go to higher orders, it takes unbelievably much more difficult to work things out. And so
Starting point is 00:30:43 I was telling the Static of Feynman, he's like, yeah, this is a stupid way to work things out. There's got to be a better way to do it. We haven't known what that better way is. I'm sort of excited right now because I finally think I understand kind of in a bigger, more foundational picture of what Feynman diagrams really are
Starting point is 00:30:59 and how one can think about them in a way that does allow one to sort of go underneath that formalism and potentially work things out in a much more effective way. It's a, I mean, to sort of, it's a spoiler for some things that I'm still working on, but essentially in Feynman diagrams, you're drawing these diagrams that sort of say an electron goes here and then it interacts with a photon, and then the photon interacts with the other electron and so on. It's a diagram of sort of interactions. And what I realized is that really those diagrams are diagrams about causality. They're diagrams that show the lines that
Starting point is 00:31:33 represent, here's an electron. Really, the electron is basically a carrier of a causality. There's an event that happens when electron interacts with a photon, and that event has a causal effect on some other event, and that causal connection is represented by this electron line, so to speak, in this Feynman diagram. That way of understanding things allows one to connect sort of Feynman diagrams to a bunch of things that have come up in our physics project, to do with things we call multi-weight causal graphs and there's a whole rather lovely theory that's starting to emerge about these things
Starting point is 00:32:07 but I haven't figured it all out yet. But in any case, that's a relevant side show. But back in the late 70s, I was trying to get computers to do these very ugly, nasty calculations of Feynman diagrams because that was the only way we knew to work out the consequences of quantum field theory
Starting point is 00:32:26 in those days, particularly QCD, which was a young field in those days. days. And I, as the teenage me, had the fun of being able to work out a bunch of calculations about QCDs for the first time. And, you know, now they're well-known, sort of classic kinds of things. But then they were fresh and new because it was a new field. But it was, one of the things that had happened was I had built a bunch of capability to do symbolic computation, algebraic computation. And one of the features of doing algebraic computation is that you don't just get a number as the answer. If the answer to your calculation, if the computer spits out
Starting point is 00:33:04 17.4, there's not a lot you can do with 17.4. There's not a lot of intuition you can get from 17.4 on its own. But when the computer spits out this big, long algebraic expression, it has a lot of structure. And one of the things that I had kind of learned to do was to get intuition from that structure. And so, for example, if you're doing, I don't know, you're doing integrals, let's say. I've never been good at doing integrals by hand, but I became what I learned from doing thousands of integrals by computer was kind of things about the intuition about the structure of what happens in integrals.
Starting point is 00:33:36 And that allows one to kind of make, to sort of jump ahead and see this complicated integral. It's going to be, I think, intuitively, roughly of this structure, and that's a big clue in actually being able to solve the thing on a computer. So the thing that I realized only very recently is that my kind of experience in doing algebraic computation had gotten me used to the idea that what a computer produces will be a thing that has structure from which you can get intuition.
Starting point is 00:34:02 And so when I started thinking about sort of actual simple programs and what they do, it was sort of obvious to me, really was obvious, that I should just make some sort of visualization of what all the steps that were going on were, because I expected to get intuition from kind of the innards of the computation, so to speak. However, one of the things that I will say is that I started studying, in that case, cellular automata, back in 1981, and I found out a bunch of things about cellular automata. I thought they were pretty interesting.
Starting point is 00:34:36 And I had generated back in 1981, I generated a picture of this thing called Rule 30, which is a particular cellular automaton that's been my all-time favorite. And it has the feature that has a very, very simple rule, but you started off from just this one black cell. It makes this complicated pattern. Many aspects of that pattern look for all practical purposes random. It's something that my intuition back in 1981 said couldn't happen.
Starting point is 00:35:04 It said, you know, if the rule was simple enough, there will be a trace of that simplicity in the behavior that's generated. And so when I, you know, I generated a picture of Rule 30, I even put it into a paper I published, but I didn't really pay attention to it because my intuition was so strongly, nothing like that can happen. Then, actually, sort of methodologically amusing, I think June of, 1984, I happened to get a high-resolution laser printer. They were a new thing. There were big, clunky objects at that time. And I thought I was going to go on some plane flight, and I thought I'll make some cool pictures from my new laser printer. So I printed out Will 30 at high-resolution
Starting point is 00:35:45 and took it with me, and I'm starting to look at it. And it's like, hmm, what an earth is going on here? I finally sort of really started, well, actually some other things that happened. I had also studied other aspects of cellular automata and how they related to theory of computation and so on. And that kind of primed me for really looking more seriously at this picture and realizing, oh my gosh, this is something that is completely violates the intuition that I've always had, that, you know, to get something complicated, you need complicated rules in some sense, or you need a complicated initial condition. This is something new and different. And it is kind of amusing that I realized that it really, at that point, I was primed enough from the other
Starting point is 00:36:26 things I'd studied, particularly about computation theory, that it only took a couple of days before I was like sort of telling people about, you know, I was going to some conference. And actually, I found recently there was a transcript of a Q&A session at that conference where I'm kind of talking as if I'd known it forever about, you know, how Rule 30 works and so on. But actually, I'd only known it for two days. But an important point about that was, First of all, it's something I had kind of quotes discovered, but I had not understood. I had not internalized it. But to be able to internalize it required me to build up a bunch of other context from studying,
Starting point is 00:37:07 well, a bunch of things about cellular automata, a bunch of things about computation theory. Given that priming, I was able to actually understand this point about what Rule 30 is and what it means. Now, for example, this phenomenon of computational irreducibility, I'm now at sort of 40 years and counting since I came with that idea, and I'm still understanding what its implications are, whether it's for, you know, AI ethics or for, you know, proof of work for blockchains or a whole bunch of different areas. I'm only now understanding, in fact, the thing that I've understood most recently, I would say, is that I think a lot of the story of physics is a story of the interplay between computational irreducibility and our kind of limitations as observers of the world.
Starting point is 00:37:51 that it's something like, for example, the simplest case is the second law of thermodynamics, where kind of the idea is you've got a bunch of molecules bouncing around and the second law says most of the time it will seem like those molecules get sort of more random in the configurations that they take on. And that's something people have wondered about since the mid-1800s,
Starting point is 00:38:13 and I had wondered about it was one of the first things I got really interested in in physics back when I was 12 years old or so was this phenomenon of how does randomization happen in the second law? And what I finally have understood is that it happens because the sort of computational irreducibility of the underlying dynamics of these molecules bouncing around, yet that's sort of, there's an interplay between that
Starting point is 00:38:39 and the sort of computational limitations of us as observers. Because we as observers aren't capable of decrypting the sort of computation that happened in this underlying structure, in these underlying collisions, we just have to say, oh, it looks random to me, so to speak. And so this phenomenon of computational reducibility, sort of at 40 years and counting, I'm still understanding sort of the implications of this particular idea. And I mean, another thing to say about sort of the progress of science, which I can see in my own life, and I can also see from history of science, is it can take a long time.
Starting point is 00:39:16 Once you've had some sort of paradigmatic idea, it can take a long time for one to understand the implications of that idea. And I know for things I've done, I'm fully recognized the fact that it's taken me sometimes 30 years or more to understand what it was that I actually really discovered. It's kind of like if other people don't figure it out for 50 or 100 years, that's kind of par for the course. because it took me 30 years to figure out what the significance of this or that thing was. I mean, I see that also in technology development. In building world from language, we build certain paradigmatic ideas. We have certain ideas about the structure of what can be in the language. And then it can sometimes take a decade before we really realize, given that structure, here's what you can build.
Starting point is 00:40:05 You kind of have to get used to the ideas. You have to grind around the ideas for a long time before you kind of get to the next step. in kind of seeing what's possible from. I kind of see it as sort of this tower that one's building of ideas and technology in that case. And the higher you are on the tower, the further you've built up on the tower, the further you can see into the distance, so to speak,
Starting point is 00:40:26 about what other things might be possible. Now, I think sometimes when people sort of hear about history of science, discoveries that are being made, and so on, there's a certain tendency for people to think that science works by, and somebody wakes up one day and just figures out some big thing. Spell that out.
Starting point is 00:40:48 So you know how in physics we like to reduce something that's complex into something that's more elegant, more efficient, more simple? Turns out you can do that with dinner. Hello Fresh sends you exactly the ingredients you need. They're pre-measured, they're pre-portioned, so you don't have to do this superposition
Starting point is 00:41:09 of too much cilantro versus why is there no cilantro or what have you collapsing in your kitchen every night. They have literally 40 different recipes every week, vegetarian, calorie smart recipes, protein heavy recipes, which is my favorite. Whatever your diet function looks like. And they all take approximately 15 minutes to a half hour. That's less time than it takes me to figure out whether time is fundamental. I've been using it myself. It's the one experiment that in my life has always had reproducible results. It's delicious, it's easy, it saves me time from living on black coffee
Starting point is 00:41:48 while editing episodes at 2 a.m. Personally, my favorite part, though, is that it's an activity I can do with my wife. Thus, it not only serves as dinner, but as a bonding enterprise. So if you want to make dinner the most predictable part of your day, go to hellofresh.com
Starting point is 00:42:07 slash theories of everything. 10 FM to get 10 free meals plus a free item for life one per box with an active subscription free meals applied as discounts on first box new subscribers only varies by plan again that's hellofresh.com slash theories of everything 10 fm all one word hello fresh the best way to cook just got better how do you go from maxwell's equations right my own efforts in studying history of science and my own experience in my life doing science is that simply never happens it's there's basically many years usually a decade of build-up to whatever it is one is going to potentially discover i mentioned what happened with me in rule 30 once
Starting point is 00:43:01 i was adequately primed it was kind of like it all happened very quickly but that priming took many years. And that's the thing that, you know, usually what's reported in the storybook, so to speak, is only that final moment after that priming of when you realize, well, actually, this fits together in that way and so on. And you can then sort of describe what happened. One of the things I learned about Einstein recently was that in 1904, he'd written several papers about a very different subject. He'd written a bunch of papers about thermodynamics, particularly about the Second Law of thermodynamics. I think it was much influenced by Boltzmann, who was a very philosophically oriented physicist. It was a
Starting point is 00:43:46 person who sort of believed you could figure out things about physics just by thinking about them in the style of natural philosophy, so to speak, rather than sort of being driven by, you know, experiment to experiment type thing. And Boltzman figured out a bunch of things about atoms, Boltzmann had basically had the core ideas of quantum mechanics, although of discreteness and so on. That's what Plank picked up when he studied the black body radiation problem. But in any case, you know, at the time 1904 was still very much, can you prove the second law of thermodynamics from some underlying principles that did not just sort of introduce the second law as a law of physics. Could you prove it from some underlying mechanical principles?
Starting point is 00:44:33 And many people have been involved in that. Plank was trying to do that, actually. Plank was trying to do that when he discovered the quantum mechanics, the way of sort of understanding black body radiation in terms of discrete photons. I mean, that's a weird story, because Plank, people had wondered, why do things get more random? And they kept on saying, to get randomness, you have to have some magic source of randomness. So Planck's idea was that infrared radiation, radiative heat, would be sort of the magic source of randomness that would sort of produce heat in everything
Starting point is 00:45:10 and lead to that randomness. And so he was actually studying that question when experiments came out about blackbody radiation, and he then noticed that these calculations that Boltzman had done a lot earlier, where Boltzman, just as a matter of sort of mathematical convenience, had said, let's assume energy is discrete and had then worked out what the consequences of that were. Plank said, well, actually, if you say that energy really is discrete, you fit this data a lot better
Starting point is 00:45:36 than otherwise. It took Plank another decade to actually believe that this was more than just a mathematical trick. Einstein was the one who really sort of said photons might be real a few years later. But in any case, the thing was interesting there was that Einstein was using kind of this natural philosophy, philosophical approach to science as a way to think about how things might work. And he tried applying that to thermodynamics in 1904, and he didn't get it right. You know, he didn't figure out the second law. He, I mean, as we now know, sort of the paradigmatic ideas that you need to figure out the second law come from ideas about computation and so on, which were another close to 100 years in the future, so to speak. But it's sort of interesting
Starting point is 00:46:20 that, you know, he was applying those kinds of philosophical thinking ideas, and it was a misfire in thermodynamics. It was a hit in relativity, in the photoelectric effect, and the existence of photons, and also in Brownian motion. But it's kind of, it's sort of, it's an interesting sort of footnote to the history of science. I think, you know, another point that one realizes there is there are things that there is an ambient level of understanding that will allow one to make progress in that area and there are things where there isn't. And in fact, Einstein himself, I think 1916,
Starting point is 00:46:59 he wrote to somebody, you know, in the end, space will turn out to be discreet, but right now we don't have the tools necessary to see how this works, which was very smart of him. I mean, that was correct. You know, it took another hundred years to have those tools. But it's a thing where when you think about science, another issue is, are you at the right time in history, so to speak? Is the ambient sort of understanding of what's going on sufficient to let you make the progress you want to make?
Starting point is 00:47:28 So one area I've been interested in recently is biology, where biology has been a field which really hasn't had a theory. You know, the best, the closest one gets to a theory in biology is natural selection from 1859. And, but that's a very, it's, you know, if we say, well, why is biology the way it is, you know, is there a, when we look at a biology textbook, what's the theoretical foundation of a biology textbook? We haven't known that. And the question is, can there be a theory of biology? Most biologists don't really imagine there as a theory of biology. They just say, we're collecting this data, we do these experiments, we work out these probabilities of things if you're doing medicine. That's the typical approach. And, you know, we don't imagine. that there is an underlying theory. It's not like physics, where we imagine there might be some underlying sort of primitive theory, so to speak. Sorry, why wouldn't computational irreducibility
Starting point is 00:48:21 come into play in the biological case for you to say, or for a biologist, to say that that's the reason why there is no toe for biology? Yep. Well, that's some truth to that. Are my jumping ahead? No, no, that's good, good inference. I mean, that's some... I think the reason that biology looks as complicated
Starting point is 00:48:41 as it does is precisely because of computational reducibility. And the thing that surprised me, this is another sort of story of my life in science. I had worked on cellular automata back in the early 80s. It's kind of funny fact is that this was days before the internet, so you couldn't look things up as easily, or at least before the web. And, you know, people had seen that I was working on cellular automata. So I got invited to all the theoretical biology conferences, because they figured this must be about, you know, biological kinds of things. And it's kind of funny because looking back now, I was kind of, the period in the 1980s was a time when there was sort of a burst of somewhat interest in theoretical biology. There'd been another one in the 1940s, and it kept on sort of dying
Starting point is 00:49:28 off. But in 1980s, there was something of a burst of interest. And I realized that, in fact, somebody who's working with me now whose biologists sort of went back and looked at some of these conferences and things. And I keep on saying, I don't really know much about biology. And he kept on saying, but you were there at all these key conferences. Wait, just a moment. It's one thing for you to be invited because of cellular automata. It's another thing for you to accept it. Why did you go if you thought, hey, this is irrelevant?
Starting point is 00:49:52 Oh, no, I didn't think it was irrelevant. I thought it was interesting. I just, you know, had a certain... I see. And, you know, when I talked about cellular automata there and talked about the way that cellular automata are relevant to things like the growth of organisms and so on, I thought it was interesting and other people thought it was interesting. And it turned into a whole sort of subfield.
Starting point is 00:50:11 of people studying things. But I didn't think, when it comes to sort of the foundations of biology and things like, you know, why does natural selection work? I didn't think it had anything much. Well, I wasn't sure if it had nothing to say about that. Back in the mid-80s, I tried to see if it had something to say about that. I tried to see if you could take cellular automata, which have these definite underlying rules.
Starting point is 00:50:35 Those little underlying rules are a little bit like genomic sequences, so to speak, which, where the genome is specifying the rules for building an organism. So in a cellular automata, you can think about it the same way. The underlying rules specify the rules for building these patterns that you make in cellular automata. So the obvious question was,
Starting point is 00:50:53 could you evolve those rules like natural selection does? Could you make mutations and selection and so on on those underlying rules? And could you see that produce some sort of interesting patterns of growth? So I tried that in 1975. I didn't find anything interesting. Do you mind explaining how you tried that? Because look, the way that it works with cellular automata like Rule 30 is you have some grid.
Starting point is 00:51:17 How long is the grid? Maybe 30 or 40? No, no. It's an infrared grid. Okay. So you have some grid and then you have some rules. But at no point do you change the rules? Correct. Okay.
Starting point is 00:51:26 So even if you were to change the rules, are you changing the rules for only finite parts of the grid? No, no. This is saying, instead of looking, and it's not going to work so well with something as simple as Rule 30, because you can go from Rule 30 to Rule 31. Rule 30 and Rule 31 behave very, very differently. By the time you're dealing with sort of Rule a billion, rule a billion and one potentially doesn't behave really that differently from Rule a billion.
Starting point is 00:51:53 Because many of the cases in the rule, the rule is saying, if you see right near where you are, if you see cells red, green, white, then make this. And by the time you've got enough colors and so on, you can imagine that some of those particular rules won't even be used when a typical pattern is produced. It'll be something, so you can make some small changes to those rules and expect that maybe it won't make a huge change in what comes out. At least in the short run. Yes, yes. I mean, well, that's a complicated issue because it depends on what subpart of the rules
Starting point is 00:52:28 you end up selecting as you make this pattern. I mean, it's usually the case that when you make, you produce some pattern from a cellular automaton, for example. Every little sub-patch is typically using only some small subset of the rules, and it's using them in some particular way, and it maybe makes some periodic little sub-patch, and then something comes along that's using some different part of the rule, sort of crashes into that and destroys the simplicity that existed in that region. So in the case of Rule 30, there are only eight cases in its rule,
Starting point is 00:53:02 So you don't get to making any change to that is a big change, so to speak. By the time you have something with, like, say, 27 cases in its rule, you can kind of imagine, you know, making a little change in that is less significant to the behavior that will occur. But, okay, back in mid-1980s, I tried this, I tried looking at slightly bigger rules and making small changes. I didn't find anything interesting. Regarding biology?
Starting point is 00:53:27 Yes. Okay. Yes. No, I was interested in kind of a model for natural selection. It was a time when artificial life was first being talked about and worked on. And, you know, Chris Langton had done a bunch of nice stuff with my cellular automata and thinking about artificial life and so on. And it was, it was seen sort of the obvious thing to do.
Starting point is 00:53:47 But it didn't work. And so for years, I was like, it's just not going to work. And then last year, actually, I was interested in the foundations of machine learning. And the big lesson of machine learning that we particularly learned, in 2011 was in a neural net, if you bash it hard enough, it will learn stuff. And, you know, that wasn't obvious. Nobody knew that you could get sort of a deep learning, you know, a deep neural net to recognize images of cats and dogs and things. And sort of by accident, it was discovered that if you just leave the thing learning long enough, if you leave it training long enough,
Starting point is 00:54:23 it comes out and it's actually succeeded in learning something. So that's a new piece of intuition that you can have a system like a neural net. Neuronet is much more complicated than one of these cellular automata. You can have a system like that. You just keep bashing it, you know, a quadrillion times or something, and eventually it will successfully achieve some fitness function, will learn something, we'll do the thing you want it to do. So that was a piece of new intuition. So I thought, let me just run these, you know, this, I was writing something about actually foundations and machine learning. And I thought, let me just try this experiment that I, you know, I thought I'd tried it in the 1980s and it hadn't worked, but now we know something different from our
Starting point is 00:55:02 studies of neural nets, let me try running it a bit longer, and by golly, it worked. And I felt a bit silly for not having discovered this any time in the intervening, you know, 40 years or something, but nevertheless, it was really cool that it worked. And that meant that one could actually see, for the first time, kind of sort of more, in a more detailed way, how natural selection operates. And what's really going on in the end is computational irreducibility lets one go from these quite simple rules, these very elaborate kinds of behavior. The fitness functions, the things that sort of are determining whether you survive or not, those are fairly coarse in biology, but let's imagine there's a, you have a coarse fitness function, like what's the
Starting point is 00:55:49 overall lifetime of the pattern before it dies out, let's say, or how wide does the pattern get? You're not saying how it gets wide. You're not saying particularly, but just saying how wide does it get? Turns out with those kinds of course fitness functions, you can successfully achieve high fitness, but you achieve it in this very complicated way. You achieve it by sort of putting together these pieces of irreducible computation. And that means that so in the end, the answer, I think, to why does biological evolution work is that it is the same story as what happens in, physics and mathematics, actually, it is an interplay between underlying computational irreducibility and the computational boundedness of, quotes, observers of that computation.
Starting point is 00:56:35 So in the case of physics, the observers are us doing experiments in the physical world. In the case of mathematics, it's mathematicians looking at the structure of mathematics. In the case of biology, the observer is kind of the environment. It's the fitness function. The fitness function is kind of the analog of the observer, and the fitness function is saying you're a success if you achieve this kind of course subjective. And the reason that biological evolution works is that kind of there's so much power in the underlying irreducible computation that you're able to achieve many of these kind of
Starting point is 00:57:10 course fitness functions. So, you know, if you imagine that the only way, I don't know, an organism could survive is if it breaks out of its egg and immediately it computes a thousand primes and does all kinds of other weird things, right? It's not going to survive. I did that.
Starting point is 00:57:31 Right. It's so I did when I was born. As one does. But, you know, that's, one isn't going to be able to hit that particular very complicated target. What actually happens is the fitness functions are much coarser than that.
Starting point is 00:57:49 and that's why biological evolution has been able to work. But if you say, well, what's actually going on inside, what's going on inside is pieces of computational irreducibility being stuck together in a way that happens to achieve this course fitness function. By the way, this is the same thing as what's going on in machine learning, I think. In machine learning, it's the same story that the fitness function in that case
Starting point is 00:58:11 is you're trying to achieve some training objective and you do that by sort of fitting together of these lumps of irreducible computation. The analogy I've been using is it's kind of like building a stone wall. If you're doing kind of precise engineering, you might build a wall by making precise bricks and putting them together in a very precise way. But the alternative is you can make a stone wall where you're just picking up random rocks off the ground and noticing, well, this one more or less fits in here, let me stick
Starting point is 00:58:39 that in that way, and so on. That's what's going on in machine learning. You're sort of building this stone wall. if you then say, well, why does this particular feature of this machine learning system work the way it does? It's, well, it's because we happen to find that particular rock lying around on the ground or because we happen to go down this particular branch in the random numbers that we chose for the training. And so it's kind of an assembly of these random lumps of irreducible computation. That's what we are too.
Starting point is 00:59:07 And that's, I mean, in biology has, you know, the biology, the history of biology on Earth, you know, the dice has been rolled in particular ways. We have stuck together these sort of lumps of a reducible computation, and we make the organism that we are today. There are many other possible paths we could have taken, which would have also achieved a bunch of fitness objectives, but it was just a particular historical path that was taken. One of the things that's kind of a sort of slightly shocking thing to do
Starting point is 00:59:34 is to take one of these evolved cellular automata that looks very elaborate, has all these mechanisms in it, has all these patches that do particular things and that fit together in these interesting ways and so on. And you ask in L.LM, say, write a description of this pattern in the style of a biology textbook. And it's kind of shocking because it sounds just like biology because it's successfully describing, you know,
Starting point is 00:59:57 there's a, I don't know, it makes up, sometimes it can make up names for things. You know, there's a distal, you know, a distal triangle of this and so on. And, you know, interacting with this and this and this. And you go and you open a biology. textbook and it reads kind of just the same. It's a description, you know, biology and the detail in biology is a description of this particular sort of sequence of pieces of computational
Starting point is 01:00:24 irreducibility that got put together by the history of life on earth and that make us as we are today. Now, you know, since we care about us, it's very worthwhile to study that detailed history, that detailed, you know, lump of computational irreducibility that is us. But if you want to make a more general theory of biology, it better generalize beyond the details of us and the details of our particular history. And the thing that I've been doing actually most recently is I think we are the beginnings of finding a way to talk about sort of any system that was adaptively evolved. So in the case of the, if we look at all possible rules that could be going on in biology, many of those rules weren't be ones that would have been found by
Starting point is 01:01:15 adaptive evolution with course fitness functions. So a way to think about this. So the thing, we look at sort of what is biology doing? Biology is, if nothing else, a sort of story of bulk orchestration and molecular processes. There are all these, you know, one might have thought at some time in the past that biology is sort of just chemistry. And by that I mean, in chemistry, in chemistry, we sort of imagine we've got liquids and so on, and they just have molecules randomly bumping into each other. But that's not what biology mostly seems to be doing. Biology is mostly a story of detailed assemblies of molecules
Starting point is 01:01:51 that orchestrate, you know, this molecule hooks into this one and then does this and et cetera, et cetera, et cetera. And sort of discoveries in molecular biology keep on being about how orchestrated things are, not how this molecule randomly bumps into this other molecules. Right, it's sophisticated. Yes, but it also has this sort of mechanism. to it. It has a, it's not, it's not just random collisions. It's, this molecule is guided into
Starting point is 01:02:16 doing this with this molecule and so on. It's a, it's a big tower of things, a bit like in these evolved cellular automata, which also do what they do through this big tower of detailed sort of applications of rules and so on. But so what I'm interested in is to have a theory of bulk orchestration. That's something that can tell one about what happens in any system that is sort of bulk orchestrated, which can include things like a microprocessor, let's say, which has its own sort of complicated set of things that it does. A microprocessor is not well described by the random motion of electrons. It's something different from that.
Starting point is 01:02:53 But what is it? What can you, and does that theory that you make of the microprocessor depend on the details of the engineers who designed it, or are there necessary features of any system that has been built to achieve certain coarse-grained purposes. So I'm sort of, I'm going down this path. We're not there yet. I mean, in physics, the idea of statistical mechanics is an idea a bit like this, because the idea of statistical mechanics is, once you have enough molecules in there, you can start making conclusions just by looking at sort of averages based on what all possible configurations of molecules are like without having to have any details about the
Starting point is 01:03:34 particulars of, you know, these collisions and those collisions and so on. So it's a, the sort of the statistics wins out. There's lots of stuff there is more important than the details of what each of the pieces of stuff does. In the case of biology, there's one additional thing you seem to know, which is that the stuff you have has rules that were subject to some kind of adaptive evolution. And even though we don't know what the purpose was, you know, when, when people use sort of natural selection as a theory in biology, they look at some weird creature and they say it is this way because. And sometimes that because is less convincing than other times. But that's been the model of how one makes the theory in biology. And so I think what I'm interested in is,
Starting point is 01:04:23 is there a theory that's independent of what the because is. Just that there is a because, is that there is some coarse-grained fitness that's achieved, is that enough of a criterion to tell you that something about sort of this bulk orchestration, this limit of a large number of things subject to that kind of constraint? Actually, I figured out something this morning about this. We'll see whether it pans out. But I have one of the things that's really, really striking about these kinds of systems where one has done this kind of adaptive evolution on a large scale is you just make pictures of them and they just look very organic. It's a strange thing. And I think I have an idea about how to sort of characterize that more precisely that may lead
Starting point is 01:05:07 one to something that is sort of a, I mean, what one's looking for is something that's kind of a theory of bulk orchestration has, for example, information theory has been a general theory of just sort of the all possibilities type situation with data or with statistical mechanics and so on. So, you know, this is, but sort of it's interesting methodologically perhaps when I'm working on something like this, it's, it's an interesting, for me, it's an interesting mixture of thinking quite sort of philosophically about things and then just doing a bunch of experiments, which often will show me things that I absolutely didn't expect. And I suppose there's a third branch to that, which is doing my homework, which is, okay, so what have other
Starting point is 01:05:52 people thought about this. And that what I found is that, you know, when I try and learn some field, I often spend years kind of accumulating knowledge about that field. And I'm, you know, I'm lucky enough I bump into the world experts on this field from time to time and I'll ask them a bunch of questions. And usually I'll be kind of probing the foundations of the field. And one of the things I learned about sort of doing science is you might think, you know, the foundations of a field are always much more difficult to make progress in. than some detail, you know, high up in the tree of that field. This is often not the case,
Starting point is 01:06:29 particularly not if the foundations were laid down 50 or 100 years ago or something. Because what's happened is, you know, what you've discovered is when you talk to the people who are sort of in the first generation of doing that field of science and you say, well, what about these foundations? They'll say, good question. We wondered about that. We're not sure those foundations are right, you know, et cetera, et cetera, et cetera. Now you go five academic generations later,
Starting point is 01:06:51 and people will say, of course those foundations are right. How could you possibly doubt that? You know, it just becomes their building on top of this, you know, this thing that's far away from the foundations. Well, often the foundations are in a sense very unprotected. Nobody's looked at them for decades, maybe longer. And often the ambient methodologies that exist have changed completely. In modern times and things I've done a lot, sort of the computational paradigm is the biggest such change. And then you go, look at these foundations, and you realize, gosh, you can actually say things about these foundations, which nobody has even looked at for ages because they were just building many layers above those foundations. And so I think that's been one of the things I've noticed, and it's one of the, you know, for people who are sort of doing science and they want to make some progress in some particular area, it's like, well, what is the foundational question of this field? And, you know, sometimes there are people will sometimes have to think about that quite a bit. You know, what really is the question that we're really trying to answer in this field,
Starting point is 01:07:56 foundationally, not the thing that is the latest thing that the latest papers were talking about and so on, but what really is the foundational question? And then you say, well, can you make progress on that foundational question? And quite often the answer is yes. And even the effort to figure out what the foundational question is is often very useful. But by the way, when you do make progress on a foundational question, the kind of the trickle down into everything else is dramatic, although often the stream trickles in a different area
Starting point is 01:08:25 than the existing stuff had been built. So in other words, you make progress on the foundations, and now there are new kinds of questions about that field that you get to be able to answer, even if the existing questions you don't make very much progress on, they've been well worked out by the existing foundations. You make new foundations, you can kind of answer a new collection of questions. So I think that's a sort of a typical pattern.
Starting point is 01:08:50 And for me, kind of this effort, I tend to try and sort of ambiently understand some field for a while. And then I'll typically form some hypothesis about it. Hopefully I'll be able to turn it into something kind of computational. And then I can do the experiments. I can know that I'm actually getting the right answer and so on. And then I try and go back. And often at that point, I'll try and understand how does this relate to what people have thought about before. And sometimes I'll say, wow, that's a great connection.
Starting point is 01:09:18 You know, this person had figured out this thing, you know, 50 years ago that, you know, was pretty close to what I was talking about now. I mean, like the idea of computational irreducibility, for example, once you have that idea, you can go back and say, well, when did people almost have that idea before? Like Newton, for example, had this statement that, you know, he'd worked out celestial mechanics and calculus for the motion of planets. and he made this statement that, you know, what did he say? He said something like to work out the motions of all these different planets is beyond, he said, if I'm not mistaken, the force of any human mind. So he had figured out that there would be, you know, even though in principle he could calculate the motions of these planets, he very diplomatically said it was, in his time, it was
Starting point is 01:10:08 beyond the force of any human mind, the force of, you know, the mind of God, would be capable of working it out because, after all, the planets were moving the way they were moving. But that was sort of an interesting precursor that you can go back and see that already, you know, he was thinking about those kinds of things.
Starting point is 01:10:24 But sometimes, and sometimes you find that people just utterly missed something that later seems quite obvious. Like, for example, one of the things in physics is the belief in space time. The belief that time is something very similar to space, which is something that's been quite pervasive
Starting point is 01:10:40 in the last hundred years of physics, and I think it's just a mistake. I think Einstein didn't really believe that. The person who brought in that idea was Herman Minkowski in 1919 who kind of noticed that this thing that Einstein had defined, this kind of distance metric,
Starting point is 01:10:56 the proper time, was, you know, T squared minus X squared, minus Y squared or whatever, minus Z squared. And Minkowski was a number theorist and he'd been studying quadratic forms, sums of things with, you know,
Starting point is 01:11:09 with squares in them and so on. He's like, this is a quadratic form. It's so great. And look, you know, time enters into this quadratic form just like space. Let's bundle these things together and talk about space time. And I think that was, as I say, I think that was a mistake. I think that misled a lot of people. I think my own view, which is pretty, I think, it's clear that this is the way things work is that sort of the nature of space as kind of the extent of some data structure effectively, some hypergraph, for example, that the nature of that is very different
Starting point is 01:11:41 from the kind of computational process that is the unfolding of the universe through time. The fact that those things seem very different at first, it's then a matter of sort of mathematical derivation to find out the relativity works and makes them enter into equations and things together, but that's not their underlying nature. But in a case, that's a thing where in that particular case,
Starting point is 01:12:06 it was, you know, that was a thing where when you go back and look at the history, you say, why do people believe in space time? Why do people believe that space and time the same kind of thing? You eventually discover that piece of history and you say they weren't off in the wrong direction. Sometimes you're like, wow, they figured it out.
Starting point is 01:12:20 They really were in the right direction. Or they were in the right direction, but they didn't have the right tools or whatever else. And for me, that's a very important grounding to know that I know what I'm talking about, so to speak. Like recently, I was studying the Second Ororthodynamics. I finally think after sort of 50 years of thinking about it that I finally nailed down how it really works and how it sort of arises from this
Starting point is 01:12:43 kind of interplay between computational irreducibility and our nature as observers. And I was like, let me really check that I'm actually getting it right. And, you know, I'd known about the second law for a very long time. And the second law is one of these things, which in a textbook, you'll often see the textbook say, oh, you can derive that entropy increases. You say, well, actually, you can flip around this argument and say, that the same argument will say that entropy should decrease. And the, you know, my favorite is the books where the chapter ends. This point is often puzzling to the student. It's been puzzling to everybody else, too. But, you know, the question was, why did people come up with the things
Starting point is 01:13:23 they came up with in this area? And so I went and kind of untangled the whole history of the second law, which I was surprised nobody had written before. Although after I figured it out, I wasn't as surprised because it's really complicated. And it requires kind of understanding various points of view that people had that are a little bit tricky to understand. But I think, you know, in the end, I feel very confident that how what I figured out fits
Starting point is 01:13:47 in to what people have known before, what people have sort of be able to do experiments on and so on. And that's a, you know, that for me is an important sort of step in the kind of the philosophy, the computational experiments, the homework, so to speak. I mean, I find that, you know, if I study some
Starting point is 01:14:05 field and I'm like trying to read all these papers and so on and it's a complicated field with lots of complicated formalism. I just, I find it difficult to absorb all that stuff. For me, it's actually easier just to work it out for myself and then see where those chips have fallen and then go back and figure out what the history is and see how what I've done relates to that history. Maybe sometimes it's not so common, I have to say, but sometimes I'll discover something in the history where I say, that's an interesting idea. And I can use that idea. and something that I'm trying to do. It tends to be the case that I've kind of learned enough of the ambient kind of history
Starting point is 01:14:42 beforehand that I'm not usually surprised at that level. But it's always a humbling experience to learn some new field because you always, I mean, I feel a field I've been trying to learn for a while as economics, which is deeply related to kind of structure of human society and so on. And I'm sort of, I'm still at the stage. It happens whenever I learn a field. that for a while, every new person I talk to will tell me something I didn't know that makes me very confused about what is actually going on in the field.
Starting point is 01:15:12 I feel like I'm just at the, you know, just at the crest of the hill now for economics. I'm not, I'm just, it's starting to be the case that I've heard that idea before, and I'm beginning to understand how it fits into the sort of global set of things that I'm thinking about. And I do think, by the way, that it's going to turn out that sort of, There's a, you know, economics like biology, well, economics thinks it has more theories than biology thinks it has. And it's, but, you know, there's sort of a question of what kind of a thing is economics, what are, what are its foundational questions, what can one actually understand? I'm not there yet, but I think it's, it is really clear to me that the methodologies that I've developed are going to be very relevant to that. I don't know how it will come out.
Starting point is 01:16:01 That's one of these things you have to let the chips fall as they will. will, you know, I don't know how it's going to come out. I don't know whether, I mean, I have kind of prejudices about what I'm going to learn about, you know, cryptocurrencies and things like this, which is an interesting case because it's kind of a case where sort of all there is is the kind of economic network. There isn't the kind of obvious underlying human utility of things. I mean, just to give a preview of some of the thinking there, one of the questions in economics is, what is value? What makes something valuable? And my proto theory of that, which is some subject to change because it's not fully worked out at all, is in the end, the main thing that's valuable is essentially computational reducibility.
Starting point is 01:16:47 In other words, in the world at large, there's lots of computational irreducibility, lots of things that are unpredictable, lots of things you can't do quickly and so on. But we humans have one thing in fairly short supply, and that's time, because at least for the time being, we're mortal, we have only a limit. amount of time. So for us, anything that kind of speeds up what we can achieve is something that is valuable to us. And computational reducibility is the possibility of finding these little pockets where you can kind of jump ahead, where you're not stuck just going through, letting things work as they work, so as well. So I think the, you know, my proto theory is that the ultimate concept of the ultimate source of value is pockets of computational reduce stability. The fact that you can sort of put together a smartphone and it's a whole smartphone
Starting point is 01:17:39 rather than having to, you know, get all the ingredients together and just go step by step, so to speak. If we had infinite time, we could build every smartphone from scratch ourselves. But because we only have finite lifetimes, it's worthwhile for us to, you know, to have the finished goods, so to speak. You know, it's the very beginning. I'm just trying to understand these things. So roughly speaking, anything that saves us time, will be valuable, or at least what is valuable will save us time. And then what saves us time is something that's computationally reducible. Yes, yes, that's the idea.
Starting point is 01:18:13 And I mean, I think there's sort of questions about kind of what, you know, when you invent something, how do you build on that invention? How do you take that lump of reducibility and make use of it and so on? What is the value of that invention? That's not something usually taken into account in economics. You know, there's the scarcity of stuff, but not the value of this idea and so on. And in biology, we can see the same thing happening. There is some sort of piece of reducibility, some mechanism that you see being found,
Starting point is 01:18:52 and then that mechanism is reused. I don't understand how this stuff works yet. But this is, you know, in the end, it's kind of a theory that allows one to understand something about kind of about the function of things as well as about the mechanism by which the things occur, so to speak. But, you know, this is... So for me, I mean, in terms of doing a project like that, it's sort of this mixture of sort of the philosophy of what's going on,
Starting point is 01:19:19 the kind of conceptual framework, a bunch of computer experiments to see what actually happens, and then sort of a doing one's homework understanding of how this fits into what other people have figured out. And it's, you know, a thing that I've done for the last 30 years or so now is, you know, in academia, it's often like you write a paper and then you're like, get some citations, you know, let's boom, boom, boom, this is whatever. I have to say, I find it amusing that people have pointed out to me that citations, to my own stuff, sort of got corrupted in various databases. And so people are now copying the utterly incorrect citations that are just complete nonsense. and you can kind of see that they didn't look at anything.
Starting point is 01:20:03 They're just like, click, click, click. You know, now I've papered my paper, so to speak, by putting in the right citations. The only thing is amusing to me right now is because papers aren't on paper anymore. It's starting to be the case that you can have the citations where they'll cite all thousand authors explicitly. It's very nice and egalitarian.
Starting point is 01:20:24 And it means that the length of the, you know, I'm always, when I'm reading papers, I'm always like something exciting is about to happen. and something. He's going to exactly. Oh, shit, we reached the end of the paper. Nothing happened. Right. And I thought I hadn't reached the end because the thumb on my, you know, as I'm moving down on the window. You've seen the percentage. Yeah, yeah, right. It's like there's still a lot of way to go. But no, actually, it's just pages and pages and pages of citations. But, you know, what I've always thought, at least for the last 30 years, the much more interesting thing is the actual narrative history of the ideas. That's the thing that really matters. It's not. You're not, you know, it's like it's nice to be able to sort of cite your friends or whatever else you're doing. But what's much more significant for sort of the history of ideas is can you actually thread together how this relates to other ideas that have come up?
Starting point is 01:21:15 And so when I wrote new kind of science, I put a huge amount of effort into the historical notes at the back. And people, you know, it's like, oh, you didn't cite my thing. Well, read the frigging note. It's like, and I said, did I get it right? Yes, actually, it's a very, very good portrayal of what actually happened, and which I think is, and I think that's, for me, that's a much more useful thing than like, boom, I copied this citation from some database that had it wrong anyway. And I think, and that's, you know, it's part of the story. I mean, to me, you know, when you do a piece of science, there's the doing of the science and there's the kind of explaining of the science and there's the contextualizing. of the science, for me, kind of the effort of exposition is critical to my process in doing science. I mean, the fact that I'm going to write something, talk about something is very
Starting point is 01:22:12 important to me actually understanding what I'm talking about, so to speak. And I, you know, when I write expositions of things, I try and write expositions that I intend anybody to be able to understand. And that is a, you know, that's a big sort of constraint on what one does, because if one doesn't know what one's talking about, it's really hard to explain it in a way that anybody has a chance to understand. And so that, for me, has been an important constraint in my efforts to do science, is, can I explain it? And people sometimes think, it's more impressive if you explain the science in this very elaborate technical way. That's not my point of view. It's more impressive if you can grind it down to be able to explain it in a way that anybody
Starting point is 01:22:52 can have, who puts the effort in, can actually understand it. And it forces you to understand what you're talking about much more clearly and it prevents the possibility of you just sort of floating over the formalism and you know completely missing the point but I think I mean you know one of the things that you know it's sort of the doing of science there's a one another question is who should be doing science you know it's I mean I do science as a hobby and I do it because I discover interesting things and I think that's fun if I wasn't discovering interesting things, I just wouldn't do it. It's not what I do for a living. You know, I run a tech company for a living. It's something that has been kind of a, you know,
Starting point is 01:23:39 I started off earlier in my life. I did science for a living, so to speak. But for, it's been, what, 40 years or something since I did science for a living, so to speak. So before we get to who should do science, we started this conversation about what is good science. And I wanted to understand your position, not only your views on science, but your position in the history of science and you took us through the past 50 years up until even this morning. So that's great. What's missing there is even a definition of what is science. I opened with what is good science and then I should have said, well, what is science to begin with? And then we can also contrast that with what is bad science. Fair enough. Understand what is good science. To me, I think science has been
Starting point is 01:24:22 traditionally, about taking the world as it is, the natural world, for example, and somehow finding a way to produce a human narrative about what's going on there. So in other words, science is this attempt to bridge from what actually happens in the world to something that we can understand about what happens in the world. That's what the act of doing science is that effort to find this thing, which kind of is the explanation that fits in a human mind, so to speak, for what's going on in the world. That's been the traditional view of science. Now, there are things that call themselves sciences,
Starting point is 01:24:58 that's a computer science, that really isn't about that. That's really a different kind of thing. Computer science, if it was a science like that, will be what I now call Ruliology, the study of simple rules and what they do, the kind of the computation in the wild, so to speak. So it's sort of a misnomer of a science, but that's the tradition of how it's called.
Starting point is 01:25:20 But so, you know, for me, the science is this what is it that we humans can understand that relates to what's what actually happens out there in the world so to speak now you know i called my big book a new kind of science for a reason because i saw it as being a different twist on that idea of what science is because when you have these simple rules and you can only know what they do by just running them you have a slightly different case you have something where you can understand the essence of what's happening, the primitives, but you cannot understand,
Starting point is 01:25:54 you can only have kind of a meta-understanding of the whole arc of what happens. You can't expect what people have been hoping for from, for example, the physical sciences, where you say, you know, now I'm going to wrap my arms around the whole thing, I'm going to be able to say everything about everything that's going to go on there.
Starting point is 01:26:12 So it's a slightly, it's a different kind of science, hence the title of the book, so to speak. But so, you know, that's my view of sort of what science is, I would say that good science tends to be science that, at least science that I think is good, is science that has some sort of foundational, it has sort of some foundational connections. It's, you know, there is science, maybe I shouldn't say good science. I should say, say, high leverage science, science that's that we can be fairly certain is going to have importance in the future, so to speak.
Starting point is 01:26:48 you know, when you're at some tentacle, some, you know, some detail of some detail, the, you know, that is maybe the detail of the detail will open up some crack that will let you see something much more foundational, but much of the time that won't happen. And I think the thing that to me makes science that sort of high leverage science is a science where sort of the thing you're explaining, the thing you're talking about is somehow very simple, very, very, very clean, very much the kind of thing that you can imagine will show up over and over and over again, not something where you built this whole long description that went three pages long to say what you're studying. It's like, this is this very
Starting point is 01:27:31 simple thing and now there's a lot that comes out of that simple thing, but it is sort of based on this kind of foundational primitive. That's, at least, I wouldn't necessarily say that that is, I mean, if we talk about good science versus not very good science, you know, one thing is, so what would be my criteria? I mean, you know, I would say that science that nobody can understand isn't very good science, since the point of science is to have a narrative that we humans can understand. If you are producing something that nobody understands, that's, or nobody has the chance to understand, so to speak, that's not going to be a good thing. I think that there's also a lot of science that gets done that I would say is not, I don't know, it's not what it's advertised
Starting point is 01:28:19 to be, so to speak. This happens both in theory and an experiment. What do you mean? Well, science is hard, and unless you, you know, it's like, did it really work that way, or did you fudge something in the middle, so to speak? And, you know, there's a certain rate of fudging in the middle. I don't think we know what the rate is. In some fields, it's probably very high. And it's often not even like I'm nefariously fudging it in the middle. It's just, I knew it was going to come out this way. So the mouse that didn't do that, I'm going to, you know, that mouse must have been, you know, under the weather that day. So we'll ignore that mouse type thing. It's not, you know, it's not nefariously ignoring the mouse. It's just ignoring the mouse
Starting point is 01:29:03 because we're sure it isn't, you know, that mouse isn't the important mouse type thing. And I think the um that you know i know when i was doing particle physics back in the late 70s you know what a formative experience for me was a thing i calculation i did from kCD about some particular particle interaction charm particle production in proton proton collisions okay so i had worked out it will happen at this rate there was an experiment that said no such thing was observed at a rate i don't remember what it was five times below what i said it should be should happen And so if you are the official, you know, scientific method operative, you say, well, then my theory must be wrong.
Starting point is 01:29:46 Well, I didn't think the theory was likely to be wrong because it was based on sort of pretty foundational things. And, you know, I wrote some paper with a couple of other people and, you know, half the paper was his, the calculation. The other half of the paper was, well, this is an experiment and how come our, you know, how could our calculation possibly be wrong? Well, it turns out, as you might guess, you tell you this. story that, you know, the experiment was wrong. And, you know, that was for me an important
Starting point is 01:30:13 kind of formative realization. Now, you know, do I blame the experimentalists for the fact that it was wrong? No, experiments are hard. And, you know, they had a certain set of ideas about how it would work out. And, you know, those were not satisfied, and they missed it. Just like in doing computer experiments, if you don't measure the right thing, you might miss what you're looking for. Now, just a moment. How does that jive with earlier when you were talking about the experiments, you have to let the chips fall where they may and accept it? Fair point. I mean, you have to do a good experiment, and that's not a trivial thing. I mean, in other words, if you do a bad experiment, you'll come to the wrong conclusions. And one of the things that I suppose I've gotten so used to
Starting point is 01:30:53 in doing superior experiments is, you know, how do you make a very clean experiment? This is the typical problem. The typical problem with experiments is you do the experiment and you get a result and there was some effect that you didn't know that it mattered that the experiment was done, you know, not at sea level or something. But that was the critical thing. And you just didn't know that. And so what tends to happen with computer experiments is an awful lot easier than with physical experiments is like can you whittle it down to the point where you're doing a very, very minimal experiment where there's no, oh, there's some complicated thing and we don't know what it came from. I mean, back in the 80s when people were working on some, some, not all, but some of the
Starting point is 01:31:33 artificial life stuff, that was very much bitten by the experiments were just unbelievably complicated. They were like, well, I'm going to make this model of a fish and it's going to have 100 parameters in it. Well, then, you know, you can conclude almost nothing from such an experiment. And you say, well, you know, the fish wiggles in this way, but, you know, it could do anything with 100 parameters. And so I think the, you know, I would say, Perhaps the right thing to say is that... I've been using Claude to prep for theories of everything episodes, and it's fundamentally changed how I approach topics.
Starting point is 01:32:08 Something that's changed as of the past few weeks is that I used to rely on ChatGBTGPT for what's related to LLMs, but Claude converted me, especially since GPT 5. I don't remember the last time I've used ChatGPT consistently. Even something as simple as pasting large pieces of text, I love how Claude's interface has these as a tap. rather than embedding them as large walls of text. I also like how extremely fast it is to answer.
Starting point is 01:32:34 I like Claude's personality. I like its intelligence. I also use Claude code on a daily basis and have been for the past few months. It's a game changer for developers. It works directly in your terminal and understands your entire code base and handles complex engineering tasks. I also used Claude, the web version, Live, during this podcast with Eva Miranda here. That's actually a feature named Artifference. and no coding is required. I just describe what I want, and it spits out what I'm looking for.
Starting point is 01:33:04 It's the interactive version of, hey, you put words to what I was thinking. Instead, this one's more like you put implementation into what I was thinking. That's extremely powerful. Now, what sets Claude apart is its research depth. What I like about Claude is that it actually synthesizes across dozens of papers, finds connections between seemingly unrelated areas, and helps me spot philosophical assumptions there, buried within technical arguments. Use promo code, theories of everything, all one word capitalized, ready to tackle bigger problems, sign up for Claude today and get 50% off Claude Pro, which includes access to ClaudeCode when you use my link, clod.aI slash theories of everything.
Starting point is 01:33:49 And, you know, the most common cause of error, I suspect, in experiments, is prejudice about how it's going to come out. and then muddiness in the experiment. When the experiment is whittled down enough, you can't kind of hide, you can't let the prejudice, you can't distort the experiment by the prejudice, so to speak, because it is so simple. You can just see, this is what goes in, this is what comes out. There's nothing, you know, behind the curtain, so to speak.
Starting point is 01:34:17 So I think that's, but, you know, but a lot of experiments, physical experiments, you know, people who do physical experiments, have it much more, much worse than people like me who do computer experiments, because it's really hard, you know, to make sure you're not having any other effect you don't understand, et cetera, et cetera, et cetera. I mean, you know, the particular mistake that was made in this experiment I mentioned before was they were studying, they were looking for tracks of particles in some emulsion stack, and they, the particle tracks were a bit shorter than they were looking for
Starting point is 01:34:50 because they just didn't know about how those kinds of particles worked. Well, you know, That's something, I suppose, there are analogous mistakes one could make in a computer experiment, but it's a lot easier to not make those mistakes in a computer experiment, particularly if you do the visualization well, and you're really kind of seeing every bit, so to speak, at least at some level. Visualization is something I want to get back to, so I'm glad you brought it up, because you brought up three or four times now is something extremely important. So can you talk about how can visualization be done mediocrely, and then how is it done well?
Starting point is 01:35:22 Let's imagine you have several particles and you want to visualize them. Are you just talking about we visualize them as spheres instead of just 2D blocks? Are you talking about, well, we should be able to rotate? What are you talking about? No, the most common mistake is that there's a lot going on, and you choose to only look at some small slice of what's going on. So, for example, classic example, people studying snowflake growth, okay? They were studying, they said, well, the thing we really going to concentrate on
Starting point is 01:35:48 is the growth rate of the snowflake, okay, how fast the snowflake expands. Right? So they had worked out their growth rate and they'd carefully calibrated it and they found out it was correct for snowflakes and so on. But they never bothered to make a picture of what the actual snowflakes they were growing would their model look like. They were spherical. Okay. All right. So that's an example of kind of a visualization type mistake. So the trick is can you make a representation of things that is as faithful as possible, has as many of the details as possible, but you can you make a representation of things that is as faithful as possible, but you can you make a representation of things that is as faithful as possible, but you can you make a representation of you. yet is comprehensible to our visual system. And it's easier in some cases than others. Like cellular automata are a particularly easy case. They're particularly suitable for our visual system.
Starting point is 01:36:33 It's just at least the one-dimensional ones that I've always studied. You know, you just have this line of cells and they go down the page. You get this picture. Boom, there it is. Now, even there, with cellular automata, what had been done before my efforts were to look at two-dimensional cellular automata where you have kind of a video of what's going on. When you have that video, it is really hard to tell what's happening.
Starting point is 01:36:56 Some things like the Game of Life Cellular Automaton, people studied that at some length. And you can see these gliders moving across the screen, and they look very cool, and you can see glider guns doing their thing and all that kind of thing. But when you tip it on its side and you look at the kind of space-time picture of what's happening, it becomes much clearer of what's going on. And that's a case where our visual system, yes, we can see movies, but we don't get sort of in one gulp the whole story of what happened. And so that's another case where the more faithful visualization, if you can do it, it's not so trivial to do it because it's kind of a three-dimensional thing
Starting point is 01:37:32 and you can't see through all the layers and you have to figure out how you're going to deal with that. And it's complicated. For the things that I've studied, for example, in our models of physics, where we're dealing with hypergraphs. Hypergraphs are not on their face, trivial things to visualize.
Starting point is 01:37:48 And in fact, one of the things that made the physics project possible was that a decade earlier, in Wolfram Language, we had developed good visualization techniques for graphs, for networks, which was a completely independent effort that had nothing directly to do with my sort of project in physics, but it was something we did for other reasons, and that was something that was critically useful in doing physics project. I mean, back in the 90s, when I was doing graph layout, I actually found a young woman who was, who was spectacular good, on a piece of paper, laying out a graph to have the, you know, the lines not crossed and so on. Later on, she became a distinguished knitwear designer. So I think that was, I don't know
Starting point is 01:38:30 what was cause and what was effect, but it was kind of a unique skill of most people are not, you know, most people can't do it. You know, you give them a bunch of, a bunch of nodes in a graph and you say untangle this thing. It's really hard to do. I can't do it at all. But, you know, we had found algorithms for doing that, which I used a lot, you know, in doing. things with the physics project because that was a sort of a pre-existing thing. But it's still more difficult to visualize what's happening. And the things I've been doing very recently last few weeks on Lambda calculus, that's an area where the obvious visualizations are just horrendous. Our visual system doesn't figure out what the heck is going on. And, you know,
Starting point is 01:39:12 it's been possible. I found some reasonable ways to do that, which are very helpful in getting intuition about what's happening. But that's our highest bandwidth way of getting data into our brains is, you know, 10 megapixels of stuff that we get through our eyes. And so that's the best, and that's the question is, can you make a sort of a faithful representation of what's going on underneath that you can get into your brain so that your brain has the chance to get intuition about what's happening or to notice anomalies? I mean, you know, somehow I think the story of my life in science devolves every so often and like it did just a few days ago,
Starting point is 01:39:53 into a very large number of little pictures on the screen and going through screen after screen of these things looking for ones that are interesting. It's kind of a very natural history-like activity. It's kind of like, you know, when do I see the flightless bird or something like this? And, you know, why do I do that? Obviously, I've used machine learning techniques
Starting point is 01:40:14 to prune these things, et cetera, et cetera, et cetera. But in the end, you know, I'm looking for the unexpected. and, you know, the unexpected for me is what my brain registers is unexpected, and that's the thing I'm interested in, service plea. Okay, two questions here. One, if you're using machine learning to sift through, I remember there's a talk from Freeman Dyson.
Starting point is 01:40:36 He was saying that when the Higgs particle was discovered, he's happy that the Higgs was discovered, but he's not happy with how it was discovered because there was so much filtering out of data. And he says you want to be looking for the anomalies. I'm happy that the particle is finally discovered after many years of effort. But I'm unhappy with the way the particle was discovered. The large Hadron Collider is not a good machine for making discoveries.
Starting point is 01:41:01 The Higgs particle was only discovered with this machine because we told the machine what we expected it to discover. It is an unfortunate deficiency of the Hadron Collider that it cannot make unexpected discoveries. Big steps ahead in science usually come from unexpected discoveries. Do you have reservations about how some of these filtering techniques can be done? Absolutely. You get it wrong all the time. The less filtering you do the better.
Starting point is 01:41:29 The more, that's why I end up with, you know, looking at arrays of pictures, you know, screen after screen of arrays of pictures. It used to be on paper. But, you know, I would print these things out and go through sheafs of paper looking for things. Yeah, it's, I mean, there are things that are fairly safe to do, although, They bite you from time to time. Like I was just mentioning earlier, this very long-lived Lambda creature that I found, you know, the automated techniques that I had missed it. So that was some, and I noticed something was wrong
Starting point is 01:42:02 by looking at a visualization of what it was doing. Okay, then the other question was approximately an hour and a half ago or so, we were talking about bulk organization theory and natural selection. So bought. Pulk orchestration. Bulk orchestration, got it. Yeah, you made an acronym. Okay, that's a new acronym.
Starting point is 01:42:17 and then. Okay, the bot theory. And the question is whether it's critical to how bots themselves work. But let's skip that. So what I want to know is, you mentioned, if I recall correctly, that there was a recent visualization you did in order to make it easier to see the connection to biology. Not quite. I mean, I, no, not quite. That was related to this. I'm doing multiple projects right now. So that was about a different project, which actually happens to have some relevance to biology, but that relevance is more related to origin of life, and it's a slightly more circuitous route. So, different kind of thing.
Starting point is 01:42:56 But, you know... Let's end the conversation on many people who email you. They email you their theory. Their theory of everything, they'll say, I have a theory. You have a recent blog post about this, actually. I have several quotes from that. We can get to that, if you like, but how can someone, many of the people who watch this show, many people who are fans of yours,
Starting point is 01:43:14 many people who watch Sean Carroll's show or any science show at all, they want to contribute to science. And they may not have the tools to contribute to science. So they use LLMs, generally speaking. Or they just don't do anything, but they have the want. How can they productively contribute to science?
Starting point is 01:43:30 That's an interesting question. So I think there are a couple of... Okay, so first point is there are different areas of science, right? And at different times, people with different levels of training and expertise have been able to contribute in different ways. like there was a time when in natural history you could go and just find beetles and so on
Starting point is 01:43:46 and that was a contribution to science because every beetle you found was something which eventually there would be some systematic thing that came out from looking at that. You know, similarly in astronomy, you find a comet, you know, you find some astronomical phenomenon.
Starting point is 01:44:01 You know, that's going to become more difficult to do as an amateur because there are systematic, you know, high precision telescopes and so on that are doing these things. You're not going to find another continent. Not anymore, not anymore. Not, you know, in, you know, there have been times when this wasn't done so much by amateurs, but chemistry, for example, there was a thing where you just, you study another compound and you just keep on doing that.
Starting point is 01:44:26 And in the end, you build up this kind of collection of knowledge where somebody is going to pick up, you know, somebody studied, you know, lithium hydroxide back in the day for no particularly good reason. And then, you know, somebody at NASA realizes that's a way to scrub carbon dioxide or whatever and it gets used. in the polar spacecraft or whatever it is. You know, it's a, it's a, so I think there's this thing where there are things that you can kind of accumulate that maybe are not, you know, not necessarily in the nod of themselves. They don't require sort of integrating a lot of things to be able to make progress.
Starting point is 01:44:59 There are areas that are more difficult. So for example, right now physics is, as it has traditionally been done, is a more difficult area to contribute to. Because back, I would say, in the 1700s, not so difficult. But now there's a pretty tall tower of stuff that's known. I mean, stuff from mathematical physics and so on that's known. And if you say, well, I'm going to have a theory of how space time works,
Starting point is 01:45:29 if you don't know what's already known about space time, which is couched in quite sophisticated mathematical terms, And not capriciously, it's just that is, you know, our human, our everyday human experience doesn't happen to extend to how spacetime curves around a black hole. That's not everyday intuition. And so it's inevitable that it's going to be couched in terms that are not accessible from just having everyday intuition. And there are fields where not so much of that has happened.
Starting point is 01:46:02 Physics is one where there's a pretty tall tower of things that have been. figured out that get you to the sort of the description of physics as we know it today. Now, it turns out that, you know, the things that we've been able to do with our model of physics, one's a little bit closer to the ground again in some aspects of it, because the study of, you know, hypergraphy writing and so on, that's something, you know, pretty much anybody can understand the ideas of hypergraphy writing. It's not, doesn't require that you know, a whole bunch of stuff about, I don't know, you know, sophisticated things about partial differential equations
Starting point is 01:46:43 and, you know, function spaces and all this kind of thing, which are fairly complicated abstract concepts. It's something where, at least at the simplest level, it's like you've got this thing you could run under your computer, you can see what it does. Now, you know, connecting that to what's known in physics, that's more challenging. And knowing kind of how that relates to, I don't know,
Starting point is 01:47:05 some resulting quantum field theory or something is more challenging. The resulting quantum field theory is not a waste of time. The resulting quantum field theory is our best condensation of what we know about how the universe works. It's not something where it's like forget all that stuff. We can just go back to kind of the everyday intuition about physics. That was a good thing a few hundred years ago. It isn't a good thing anymore because we already learned a bunch of stuff. We already figured out a bunch of things. And if you say, well, just throw away all those things that's start from scratch, that's, you know, then you've got to recapitulate those few hundred years of discoveries, and that's a, you know, that's a heavy, that's a tall hill to climb, so to
Starting point is 01:47:44 speak. But I think the, you know, one of the areas where there's sort of a wonderful opportunity for people to contribute to science that is sort of high leverage science is in this field that I call ruleology, which is kind of studying simple rules and seeing what they do. And whether it's cellular automator or touring machines or lambda calculus or a hypergraph rewriting, these are things where, you know, you run it on your computer, okay, I built a bunch of tools for doing this, which, you know, make it really easy. But, you know, you do this. You're well organized. You kind of, you won't immediately have intuition about how these things work. At least I never have. You know, it takes actually doing it for a while to have the
Starting point is 01:48:26 intuition. And people usually don't at the beginning. They're just like, oh, it'll never do anything interesting. We'll just run it and see what it does. And then, you know, if you're well organized and kind of can develop intuition, you will eventually get to the point where you can say, okay, I see how this works. I can build this thing where I can, I can add some definite piece to knowledge. You know, I can, like, for example, we have this summer school every year for grownups, and we have another summer research program for high school students. So this year for the high school students, we had like, I don't know, 80 students or something, and I'm the one who gets to figure out projects for at least almost all of them.
Starting point is 01:49:09 Oh, you define them, or they come to you and you sift through them? No, no, I define them. Usually I, during the year, I kind of accumulate a list of ones I'd really like to see done. And so this year, several students, I gave them projects which I've been wanting to do for decades and which are just studying particular kinds of simple, simple systems. One was multi-way register machines was one of them. Another one, this year was games between programs. Another one was, this year was another, there was several, but, but anyway, these are things where, and, you know, I can say these high school kids, okay, they're very bright high school kids and they're using our tools and so on. In two weeks, we're able to make quite nice progress
Starting point is 01:49:56 and we're able to add something, you know, I'm sure those things will turn into academic papers and things like this. And they were able to, you know, starting from just being a bright high school student, so to speak, not knowing, you know, eight years of mathematical physics type thing. They don't know group theory. They don't know differential calculus or maybe some. They'd probably know basic calculus. They probably, but they wouldn't need to. I mean, They, you know, this is just a sort of be organized, be careful, and, you know, have the motivation and sort of, and sort of, and a little bit, think foundationally enough that you're kind of drilling down to say, what's the obvious experiment to do? Don't, don't invent this incredibly elaborate experiment where the conclusions won't tell you anything. You know, try and have the simplest experiment. You know, what is the simplest version of this that we can look at and so on? And that, you know, it's really neat because it means it's a, it's a, It's an area. Ruleology is an area.
Starting point is 01:50:52 It's a vast area. It's the whole computational universe. You know, I can, if you say, well, what's the thing that's never been studied before? I get out my computer, pick a random number, and I'm going to be able to give you something that I guarantee has never been studied before. And, you know, it'll have a lot of richness to it. If you study that thing, that randomly generated thing, the particulars of that thing may or may not show up as important in particular in the future, but certainly building up that
Starting point is 01:51:19 body of knowledge about things like that is something that is very high-leverage science. It's something that is very, it's something that you can kind of be sure is something that's a solid thing that people will be able to build on. You know, one of the things I find striking and a bit encouraging, I suppose, is, you know, you think about something like platonic solids, you know, the acosahedron, the dodecahedron, and so on. You say, well, that's, you know, you have an object, you know, a dodecahedron, you know, a piece of wooden dodecahedron or something. And you go back and you say, well, let's, you know,
Starting point is 01:51:53 we find a dodecahedron from ancient Egypt. It looks exactly the same as the dodecahedron we have today. This is a timeless object. It's a thing that, you know, the dodecahedron has been something that has been worth talking about from the time of ancient Egypt to today. And so similarly, these things in rheology have the same character.
Starting point is 01:52:13 They're very abstract, precise, you know, simple and they're sort of foundational. And it's something where, you know, this particular rule, it's not going to be the case that somebody's going to say, oh, we learned more about the immune system. So that model of the immune system is irrelevant now. You know, the things you measured about this are irrelevant. That's not going to happen because this is, you know, we're at the foundation, so to speak. This is an ultimate abstract thing. And so anything you build there is a permanent thing. And, you know, whether you happen to find, whether you do, you know, there were many naturalists before Darwin who went and, you know, collected lots of critters around the
Starting point is 01:52:53 world, you know, Darwin realized having collected lots of critters that there was a bigger picture that he could build. You know, it was still useful for people to have collected all those critters. And, you know, Darwin and everybody else doing evolutionary biology used a bunch of the information that had been collected by those people. It was, you know, it's a different thing to kind of integrate all those things, have the sort of philosophical integration to be able to come up with the bigger theory. But that's a much more difficult thing to do than to add these kind of solid bricks to science. And I would say that, you know, as a person who's done lots of Ruliology in my time, it's, you know, if you have a certain turn of mind, it's a lot of fun
Starting point is 01:53:35 because you just, you keep on finding stuff. You keep on discovering things. I mean, you know, I'm sure back in the day when, you know, the whole planet hadn't yet been explored, you would go to, you know, someplace in the center of Africa, and it's like, oh my gosh, there's a tree that does this or this. And this is exactly the same thing. You know, every day, you see lots of things, like in this stuff about Lambda calculus I've just been doing, there's all kinds of weird stuff. I've never seen it before. I don't think anybody's ever seen it before. And it's, I'm sure nobody's ever seen it before. And it's remarkable. And it's interesting. And it's kind of, It feels, you know, it's, it's an, it's an explorational kind of thing. It's kind of you get to see for the first time something nobody's ever seen before and something that is that, you know, you kind of know is going to be a permanent thing, that is going to be a thing that is never going to change. It's never going to be, oh, it doesn't work that way anymore. It is what it is, so to speak.
Starting point is 01:54:33 And I think that rheology is a great example of a place where there has been a fair amount of sort of quotes amateur ruleology that's been done over the years. It's not been as well organized as it should be, and I fault myself for that in large measure. I mean, back in the 80s, I got a bunch of people interested in this, and a bunch of people, both professional scientists and amateurs, started studying these kinds of things. And I started a journal that collected some of these things, complex systems. But I would say that the kind of the rhythm of how to present ruleology and so on, I didn't really develop as well as I should have done, and I'm now hoping to do that. And it's a question of, you know, for example, back in the day,
Starting point is 01:55:17 when people started having academic papers back in the 1600s, if you read those papers, they read like today's blog posts. They're much more kind of anecdotal. You know, I went to the top of this mountain. I saw this and this and this. And they have, they're more personal. They're actually, I would say, better communication than what one gets in sort of the very cold academic paper of today where you know particularly i would say math papers are one of my favorite non-favorite examples where it just starts you know let let g be a group it's like why are we looking at this group who knows it's just you know because it's like it is it is beneath us or it is not appropriate or it is kind of not professional enough for us to
Starting point is 01:56:01 describe why we're doing this but so you know the back in the 1600s when when academic papers kind of originated, they were like, you know, like the blog posts of today, so at least my blog posts today, where there's both the content and a certain amount of the kind of the wrapping of why we're doing this and kind of... You have the purpose in mind, and you're conveying it, and you're showing how what you're doing is connected to that.
Starting point is 01:56:27 Yeah, right, and it's not, and it's sort of telling a story more so than it's just saying fact, fact, fact. And, you know, I'm just filing the facts, so to speak. I think with Ruliology, one of the things that is interesting is it's a place where, you know, you discover something interesting, you want to just say, this is my discovery. And you, you know, you want a way to sort of accumulate lots and lots of discoveries without having to always feel like you have to wrap a whole kind of academic story around it. And, you know, the academic system is not well set up for that. the academic system it's like there's a there's this unit the sort of unit of academic achievement which is the paper so to speak and you know just there's this particular thing that I observed
Starting point is 01:57:19 in these particular characteristics that's not so much the kind of thing that one sees there it's the kind of thing that one should be accumulating lots of in ruleology and it's something that is very accessible to sort of well-organized people who you know, who want to sort of work cleanly on this without, you know, as amateurs, so to speak. I think that's a powerful thing, and I'm hoping to have the bandwidth to kind of put together a properly organized kind of ruley logical society or something where we can kind of accumulate this kind of information. I think it's a thing where, I mean, one of the things I do and the things that I write about science is every picture in everything I write, you can
Starting point is 01:58:04 click that picture, you'll get a piece of Wolfram Language code, and at least if our QA department didn't mess up, it will forever produce the picture that I said it produced, so to speak. And, you know, I mentioned the QA department, because it isn't actually trivial to have, you know, you make some piece of code, and you have got to make sure it keeps on working. We've been very good in Wolfram Language and maintaining compatibility for the last 38 years with the language, but if I used some weird undocumented feature that day, that might change. I don't. So that usually isn't a problem. But that fact that you kind of have it as a constraint for yourself, I like to have it as a constraint for myself, the things I write
Starting point is 01:58:46 should be understandable to anybody who puts the effort in to understand them. And reproducible. Yes, but also that they're understandable, not only do the humans, but also to the computers, so to speak, and that everything I do is like you can immediately reproduce it. That's, you know, turned out in practice to be a very powerful thing. Because people just take, you know, the code and all the visualizations and so on that I've made and they just go and start from there. They kind of start from that level of the tower, so to speak, rather than having to climb the tower themselves. And it's, you know, I think that's a powerful thing. It's very, it's pretty much undone in most of academic science, partly because academics, well, they don't tend to package sort of the kind of code in a form that will actually reproduceable and runnable.
Starting point is 01:59:34 they do more with our technology than anywhere else. But still, it's somewhat inadequate. And also, I think the motivations on the part of, you know, the typical academic scientist is kind of in the game of academia, so to speak, which involves, you know, I'm going to publish my thing, I'm going to publish another thing that leverages the thing I just published, and, you know, I'm going to get as many papers as possible and so on. For me, the calculation is actually rather different.
Starting point is 02:00:04 because for me, I'm trying to do a bunch of different things in my finite life, so to speak. And for me, I'll write something, and I really don't want to write something about the same topic again. It's kind of like it's a write-once-type activity. And so as much as possible, I'm like, I'm going to write this thing, and, okay, well, here it is.
Starting point is 02:00:25 I hope you can do something useful with it because I'm not going to come back to this particular thing. I mean, I find actually that the, you know, having, I do end up sort of building on the things I've done, but not kind of writing sort of an incremental version of the same document again. I find it, maybe it's just me, but I just, I can't bring myself to do that. I feel the same way. And it's kind of, it's actually very frustrating when I, you know, when I do a project, I like to kind of pick all the low-hanging fruit. And I, you know, I know that any fruit I don't pick first time around, I'm not going to
Starting point is 02:01:02 come back and pick. And it's just going to sit there. And it's going to be frustrating to me because it's kind of like, here's this thing. And I just figured out a little bit more, but I have no place to write that down. Actually, one thing I've been doing recently is in the NKS book, there were many notes at the back. And many 100-page things that I'm writing today are the expansion of three-paragraph notes in the NKS book. So, for example, the thing that I've been doing the last few weeks is in many ways an expansion of a small note in the NKS book that will turn into a hundred page document. And so that's that's sort of okay by my standards in the sense that I didn't really work it out in enough detail before and now I am and that's not kind of recapitulating
Starting point is 02:01:47 something I've already done. Anyway, Stephen, it's been wonderful. Thank you so much for spending almost three hours, two and a half or so. Oh my God, it's that long. I just I just start yacking in it. It keeps on going. Well, this will be released in approximately two weeks, and maybe by then people will be wondering, okay, I've heard about Ruraliology. I've heard that I can contribute by looking at Rule 2 million and 2. But specifically, I want more detail. So you mentioned there are 40 high school projects. Maybe within two weeks, there'll be a blog post out for people who want more. They can say, okay, well, we go to. I want to do it, but it won't be two weeks. Well, maybe whenever it's done, I can put a link on the description.
Starting point is 02:02:30 I'll update the description. So if you're watching this, whenever you're watching this, check the description. Or maybe I'll have something on my substack about how one can possibly use Ruliology, as Stephen would hopefully approve. Sounds good. Hi there. Kurt here. If you'd like more content from theories of everything and the very best listening experience, then be sure to check out my substack at Kurt Jymongle. Some of the top perks are that every week you get brand new episodes ahead of time, you also get bonus written content exclusively for our members. That's C-U-R-T-J-A-I-M-U-N-G-A-L.org. You can also
Starting point is 02:03:14 just search my name and the word substack on Google. Since I started that substack, it somehow already became number two in the science category. Now, substack for those who are unfamiliar, is like a newsletter, one that's beautifully formatted, there's zero spam. This is the best place to follow the content of this channel that isn't anywhere else. It's not on YouTube. It's not on Patreon. It's exclusive to the substack. It's free. There are ways for you to support me on substack if you want, and you'll get special bonuses if you do. Several people ask me like, hey, Kurt, you've spoken to so many people in the field of theoretical physics, a flaw. of consciousness. What are your thoughts, man? Well, while I remain impartial in interviews,
Starting point is 02:04:02 this substack is a way to peer into my present deliberations on these topics. And it's the perfect way to support me directly. Kurtjymongle.org or search Kurtjymongle substack on Google. Oh, and I've received several messages, emails, and comments from professors and researchers saying that they recommend theories of everything to their students. That's fantastic. If you're a professor or a lecturer or what have you and there's a particular standout episode that students can benefit from or your friends, please do share. And of course, a huge thank you to our advertising sponsor, The Economist. Visit Economist.com slash Toe to get a massive discount on their annual subscription. I subscribe to The Economist and you'll love it as well.
Starting point is 02:04:55 To is actually the only podcast that they currently partner with. So it's a huge honor for me, and for you, you're getting an exclusive discount. That's Economist.com slash Toe, T-O-E. And finally, you should know this podcast is on iTunes. It's on Spotify. It's on all the audio platforms. All you have to do is type in theories of everything, and you'll find it.
Starting point is 02:05:19 I know my last name is complicated, so maybe you don't want to type in Jymongle, but you can type in theories of everything. and you'll find it. Personally, I gain from re-watching lectures and podcasts. I also read in the comment that toe listeners also gain from replaying. So how about instead you relisten on one of those platforms like iTunes, Spotify, Google Podcasts? Whatever podcast catcher you use, I'm there with you. Thank you for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.