Lex Fridman Podcast - #491 – OpenClaw: The Viral AI Agent that Broke the Internet – Peter Steinberger

Episode Date: February 12, 2026

Peter Steinberger is the creator of OpenClaw, an open-source AI agent framework that’s the fastest-growing project in GitHub history. Thank you for listening ❤ Check out our sponsors: https://...lexfridman.com/sponsors/ep491-sc See below for timestamps, transcript, and to give feedback, submit questions, contact Lex, etc. Transcript: https://lexfridman.com/peter-steinberger-transcript CONTACT LEX: Feedback – give feedback to Lex: https://lexfridman.com/survey AMA – submit questions, videos or call-in: https://lexfridman.com/ama Hiring – join our team: https://lexfridman.com/hiring Other – other ways to get in touch: https://lexfridman.com/contact EPISODE LINKS: Peter’s X: https://x.com/steipete Peter’s GitHub: https://github.com/steipete Peter’s Website: https://steipete.com Peter’s LinkedIn: https://www.linkedin.com/in/steipete OpenClaw Website: https://openclaw.ai OpenClaw GitHub: https://github.com/openclaw/openclaw OpenClaw Discord: https://discord.gg/openclaw SPONSORS: To support this podcast, check out our sponsors & get discounts: Perplexity: AI-powered answer engine. Go to https://perplexity.ai/ Quo: Phone system (calls, texts, contacts) for businesses. Go to https://quo.com/lex CodeRabbit: AI-powered code reviews. Go to https://coderabbit.ai/lex Fin: AI agent for customer service. Go to https://fin.ai/lex Blitzy: AI agent for large enterprise codebases. Go to https://blitzy.com/lex Shopify: Sell stuff online. Go to https://shopify.com/lex LMNT: Zero-sugar electrolyte drink mix. Go to https://drinkLMNT.com/lex OUTLINE: (00:00) – Introduction (03:51) – Sponsors, Comments, and Reflections (15:29) – OpenClaw origin story (18:48) – Mind-blowing moment (28:15) – Why OpenClaw went viral (32:12) – Self-modifying AI agent (36:57) – Name-change drama (54:07) – Moltbook saga (1:02:26) – OpenClaw security concerns (1:11:07) – How to code with AI agents (1:42:02) – Programming setup (1:48:45) – GPT Codex 5.3 vs Claude Opus 4.6 (1:57:52) – Best AI agent for programming (2:19:52) – Life story and career advice (2:23:49) – Money and happiness (2:27:41) – Acquisition offers from OpenAI and Meta (2:44:51) – How OpenClaw works (2:56:09) – AI slop (3:02:13) – AI agents will replace 80% of apps (3:10:50) – Will AI replace programmers? (3:22:50) – Future of OpenClaw community

Transcript
Discussion (0)
Starting point is 00:00:00 The following is a conversation with Peter Steinberger, creator of OpenClaw. Formerly known as MoldBot, Clawed bot, Clawedus, Clawed, spelled with a W, as in Lopster Claw. Not to be confused with Claude, the AI model from Anthropic, spelled with a U. In fact, this confusion is the reason Anthropic kindly asked Peter to change the name to OpenClaw. So, what is OpenClaw? It's an open-source AI agent that has taken over the tech world in a matter of days, exploding in popularity, reaching over 180,000 stars on GitHub, and spawning the social network mold book where AI agents post manifestos and debate consciousness, creating a mix of excitement
Starting point is 00:00:46 and fear in the general public, in a kind of AI psychosis, a mix of clickbait fear-mongering and genuine, fully justifiable concern about the role of AI. in our digital, interconnected human world. OpenClaw, as this tagline states, is the AI that actually does things. It's an autonomous AI assistant that lives in your computer, has access to all of your stuff if you let it, talks to you through telegram, WhatsApp, signal,
Starting point is 00:01:14 iMessage, and whatever else messaging client, uses whatever AI model you like, including Claude Opus 4.6 and GPT5.3 codex, all to do stuff. for you. Many people are calling this one of the biggest moments in the recent history of AI since the launch of Chad GPT in November 2022. The ingredients for this kind of AI agent were all there,
Starting point is 00:01:39 but putting it all together in a system that definitively takes a step forward over the line from language to agency, from ideas to actions, in a way that created a useful assistant that feels like one who gets you and learns from you, in an open source community-driven way is the reason OpenClaw took the internet by storm. Its power, in large part, comes from the fact that you can give it access to all of your stuff and give permission to do anything with that stuff in order to be useful to you. This is very powerful, but it is also dangerous.
Starting point is 00:02:16 OpenClawe represents freedom, but with freedom comes responsibility. With it, you can own and have control over your data, but precisely because you have this control, you also have the responsibility to protect it from cybersecurity threats of various kinds. There are great ways to protect yourself, but the threats and vulnerabilities are out there. Again, a powerful AI agent with system-level access
Starting point is 00:02:41 is a security minefield, but it also represents the future, because when done well and securely, it can be extremely useful to each of us humans as a personal assistant. We discuss all of this with Peter and also discuss his big picture programming and entrepreneurship life story, which I think is truly inspiring. He spent 13 years building PSPDF kit, which is a software used on a billion devices. He sold it and, for a brief time, fell out of love with programming, vanished for three years, and then came back, rediscovered his love for programming, and built in a very short time an open.
Starting point is 00:03:23 open source AI agent that took the internet by storm. He is, in many ways, the symbol of the AI revolution happening in the programming world. There was the Chagipati moment in 2022, the Deepseek moment in 2025, and now in 26, we're living through the open claw moment, the age of the lobster. The start of the agentic AI revolution. What a time to be alive. And now, a quick few seconds. mention of a sponsor, check them out in the description or at Lexfreedman.com
Starting point is 00:03:57 sponsors. It is in fact the best way to support this podcast. We got Quo for a phone system, calls, text, contacts for your business, code rabid for AI-powered code review, Finn for customer service AI agents, Blitzy for AI powered software development, Shopify for selling stuff online, element for electrolytes, and of course, our old friend, perplexity. for curiosity-driven knowledge exploration. Choose wisely, my friends. And now, onto the full add reads. I try to make it interesting, but if you skip,
Starting point is 00:04:32 please still check on our sponsors. I enjoy their stuff. Maybe you will too. And really, they are the incredible folks that make this whole thing possible. And I really do hope to do more episodes in 2026, have more fun, take more risks, and explore deeply the full range of human possibility,
Starting point is 00:04:57 of human condition, of human nature, of human civilization. Anyway, so you get in touch with me, for whatever reason, go to Lexfremant.com slash contact. All right, let's go. This episode is brought to you by Quo, spelled QUO. It's a business phone platform for calling and messaging. So it's basically a really nice interface, a really nice system for organizing all the incoming calls, text, voicemails, recordings. When you have a team and you have a large number of customers that want different things, it's a nice way to organize everything together.
Starting point is 00:05:37 I just love watching the beauty, the elegance of the interface. I'm such a sucker for beautiful interfaces, not just beautiful but functional. So the perfect mix of beauty and function in evolutionary biology and in software design, software engineering is just wonderful to watch. And that, of course, relates to the very topic of this podcast is how to create software systems like that with the utilization of the agenic loop. And as Peter talks about, still keeping the human apart, a fundamental part of that process of adding what he says, I think correctly,
Starting point is 00:06:17 sort of a bit of love into the thing, a bit of that human touch. I don't know what exactly that is, but we know it when we see it, when we feel it, when we interact with it, and that is the magic that makes great software. So anyway, Quo has that. Love the interface.
Starting point is 00:06:32 Try Quo for free, plus get 20% off your first six months when you go to Quo.com slash Lex. That's Q-U-O-com slash Lex. This episode is also brought to you by CodeRabbit, a platform that provides AI-powered code reviews directly within your terminal. Now, there's a lot of ways to use CodeRabbit. But the one I'd like to recommend and talk to you about,
Starting point is 00:06:58 outside of the ID, outside of the magic of the interface, to go back and what I just said, is the magic, the power of the CLI, of the terminal. And CodeRabbit CLI is amazing. And, of course, as we talk about it with Peter, his whole workflow, his whole approach to programming has evolved more and more towards the command line, towards the terminal, towards the CLI, because that is the language of agents. And so if you're doing coding agent stuff, integrating the review of the code into the whole process, that's what
Starting point is 00:07:31 Code Rabbit CLA comes in. It ensures that AI generated code is production ready by catching errors at that particular stage of the process. It integrates into existence. existing CLA coding agent workflows. And even though the coding models are getting smarter, smarter and smarter, they still do hallucinate. They still do make errors. And CodeRabbit CLI is a backstop for hallucinations and logical errors from AI coding agent generated code. So, install CodeRabbit CLI today at CodRabbit.a.a.
Starting point is 00:08:09 slash Lex. That's coderabbit.a.ai slash Lex. This episode is brought to you by Finn, the number one AI agent for customer service. So again, these are niche but extremely important,
Starting point is 00:08:26 extremely impactful applications. Finn takes customer service and says we're going to do a damn good job at it. A lot of companies including AI companies, 6,000 customer service leaders in top companies are using it. So you know they're legit when an AI company is using you for the AI for customer service.
Starting point is 00:08:49 I'm somebody having witnessed on the interwebs, poor customer service, poor the love for the customer, a lack of attention and care to the customer, to the pain, to the nuanced pain of each individual customer. because of that I get to deeply appreciate the value of great customer service. And I do think for scale, for efficiency, for quality, it's important to integrate AI into that process. And then Finn does a really good job with that. Go to fin.a.ai slash Lex to learn more about transforming your customer service and scaling your support team. That's fin.a.ai slash Lex. This episode is brought to you.
Starting point is 00:09:34 by Blitzy, an AI-powered autonomous software development platform. They are focusing on enterprise. Their whole system is designed and build for large complex databases. The way they do the context management, the way they, through the interface, show how everything is managed, organized, process considered in the different tasks that has to do, like refactoring gigantic code bases.
Starting point is 00:10:03 This is what they focus. on this is what they do well. Also, multi-agent. For example, they have a layer of 3,600 cooperative agents. You know, a lot of the excitement that we're talking about in this very podcast and in general, in the industry and on X and on podcasts is the scale of just a handful of developers. When you're talking about a gigantic code base, that requires a set of tooling that can handle the gigantic context, that can handle individual people going in and being able to orchestrate the large scale, refactoring, cogeneration, managing dependencies, runtime validation, all the testing they have to do. To be able to go in there and to be able
Starting point is 00:10:54 to handle that, there's just a whole set of tooling that requires a lot of care. And Blitsey does a good job with that. If you want to learn more or speak to a member of the team, go to blitzie.com slash Lex. That's B-L-I-T-Z-Y.com slash Lex. The future of autonomous software development is here. This episode is brought to you by Shopify. I don't know why every time it brings a smile to my face. Yes, it is a platform designed for anyone to sell anywhere with a great-looking online store, but of course every time I end up talking about Shopify engineering, which is the tech behind the magic, and the humans behind, the tag behind the magic.
Starting point is 00:11:40 And I got to meet some of those humans. And I get to talk to some of those humans, and they're incredible human beings. And speaking of incredible human beings, of course, I always end up talking about Toby, who's a great engineer, who's doing a lot of agentic AI engineer. He's still code, still legit programmer in the detail.
Starting point is 00:11:59 that is what great CEOs are made of. DHH also mentioned in this conversation with a bit of love from Peter, everybody loves DHS. Well, some of them, some people online are maybe don't show their love in the obvious ways, but underneath it there's deep love
Starting point is 00:12:20 and there's always respect and appreciation for how legitimate programming he is, how sharp and warm, witty and insightful his opinions, analysis is, and how great of a builder he is, an explorer, and constantly evolving and trying new things. And I think that spirit represents Shopify engineering. I just love using software that you know is built by great engineers.
Starting point is 00:12:52 So that is just such an important foundation of a great business. It's probably two things. It's making sure you're keeping every individual customer happy, customer focus. And then on the back end, making sure there's the tooling, the infrastructure that makes that possible. That's where the engineer comes in. Anyway, sign help for a $1 per month trial period at Shopify.com slash Lex. That's all lowercase. Go to Shopify.com slash Lex.
Starting point is 00:13:29 to take your business to the next level today. This episode is brought to you by the thing I'm currently sipping on, Element, L-M-N-T, my daily, zero-sugar and delicious electrolyte mix. I don't think I can live without Elm. I can't quit you. Yeah, it's just delicious and incredibly important to my health, my well-being, my energy levels, just how I feel when I'm fasting, when I'm doing, the crazy things,
Starting point is 00:14:05 emotionally, physically, mentally, all the programming stuff, this, you know, I didn't almost sleep at all last night and just fasting, just feeling good, electrolytes are such an important part of that, making sure you get the sodium, potassium, magnesium, correct. It's what do you need for life? You need water, you need electrolytes, you need electrolytes, and then you need food occasionally. But I'm very good at being able to go without food for several days, potentially. I most of the time these days fast. Do one meal a day and fast basically 24 hours.
Starting point is 00:14:42 And for that, element is extremely important. And my favorite flavor, the flavor of champions, watermelon salt. Get a free eight-counseltack with any purchase, try it at drink element.com. Lex. This is a Lex Friedman podcast that's support it. Please check out our sponsors in the description, where you can also find links to contact me, ask questions, give feedback, and so on. And now, dear friends, here's Peter Steinberger. The one and only, the Claude father. Actually, Benjamin predicted in this tweet, the following is a conversation with Claude, a respected crustacean. is a hilarious looking picture of a lobster in a suit.
Starting point is 00:15:42 So I think the prophecy has been fulfilled. Let's go to this moment when you built a prototype in one hour. That was the early version of OpenClaw. I think this story is really inspiring to a lot of people because this prototype led to something they just took the internet by storm and became the fastest growing repository in GitHub history with now over 175,000 stars. So what was the story of the one-hour prototype?
Starting point is 00:16:12 You know, I wanted that since April? A personal assistant, AI personal assistant. Yeah, and I played around with some other things, like even stuff that gets all my WhatsApp, and I could just run queries on it. There was back when we had GPD 4.1, the one million context window, and I pulled in all the data,
Starting point is 00:16:38 and then to ask them questions like, what makes this friendship meaningful? And I got some really profound results. Like I sent it to my friends, and they got like teary eyes. So there's something there. Yeah. But then I thought all the labs will work on that. So I moved on to other things,
Starting point is 00:17:02 and that was still very much in my early days of experimenting and playing. You know, you have to. That's how you learn. You just like, you do stuff and you play. And time flew by, and it was November. I wanted to make sure that the thing I started is actually happening. I was annoyed that it didn't exist, so it just prompted it into existence.
Starting point is 00:17:28 I mean, that's the beginning of the hero's journey of the entrepreneur, right? Even with your original story with PSPDF kit, it's like why does this not exist let me build it and again here's a whole different realm but similar maybe spirit yes I had this problem
Starting point is 00:17:45 I tried to show P-F on an iPad which should not be hard this is like 15 years ago something like that yeah like the most random thing ever and suddenly I had this problem and I wanted to help a friend
Starting point is 00:17:58 and there was there was not like nothing existed but it was just not good I'm like I tried it and it was like very mad. Like, I can do this better.
Starting point is 00:18:09 By the way, for people who don't know, this led to the development of PSPDF kit that's used on a building devices. So it turns out that it's pretty useful
Starting point is 00:18:18 to be able to open in a PDF. You could also make the joke that I'm really bad at naming. Like, named them the five on the current project and even PSPDF doesn't really roll from the tongue. Anyway,
Starting point is 00:18:33 so you said, screw it, why don't I do it. So what was the prototype? What was the thing that you, what was the magical thing that you built in a short amount of time that you're like, this might actually work as an agent? Or I talk to it and it does things. There was, like one of my projects before already did something where I could bring my terminals onto the web and then I could like interact with them, but there also would be terminals on my Mac, Vipe Tunnel, which was like a weekend hack project. That was, like, that was, like,
Starting point is 00:19:06 still very early and it was cloud code times. You got a dopamine hit when he got something right. And now I get like mad when you get something wrong. And you had a really great, not to take a tangent, but a great blog post describing that you converted vibe tunnel. You vibe coded vibe tunnel from TypeScript into Zig of all programming languages with a single prompt, one prompt, one shot, convert the entire code base into Zig.
Starting point is 00:19:33 Yeah. there was this one thing where part of the architecture was took too much memory every terminal used like a node and i wanted to change it to rust and i mean i can do it i can i can manually figure it all out but all my automated attempts failed miserably and then i revisited about four or five months later and I'm like, okay, now let's use something even more experimental. And I just typed convert this and this part to sick and then let codex run off. And it basically got it right.
Starting point is 00:20:18 There was one little detail that I had to modify afterwards, but it just ran for overnight to like six hours and just did the thing. And it's like, it's just mind-blowing. So that's on the LLM programming side. refactoring, but back to the actual story of the prototype. So how did Vipe Tunnel connect to the first prototype where you're like agents can actually work?
Starting point is 00:20:45 Well, that was still very limited. You know, like I had this one experiment with WhatsApp, then I had this experiment, and both felt like not the right answer. And then my searcher was literally just hooking up WhatsApp to CloudCode. one shot the CLI message comes in
Starting point is 00:21:06 I call the CLA with minus P it does its magic I get the string back and I send it back to WhatsApp and I built this in one hour and I felt already felt really cool. It's like oh I could talk to my computer right that was cool but I wanted
Starting point is 00:21:25 images because I often use images when I prompt I think it's such a such an efficient way to give the agent more context and they're really good at figuring out what I mean if it's like a weird cropped of screenshot. So I used it a lot and I wanted to do that in WhatsApp as well. Also like, you know, you just run around, you see like a post of an event. It just make a screenshot and like figure out if I have time there, if this is good, if my friends are maybe up for that.
Starting point is 00:21:52 It's like images seem important. So I worked a few, it took me a few more hours to actually get that right. and then it was just I used it a lot and funny enough that was just before I went on a trip to Marrakesh with my friends
Starting point is 00:22:14 for Bustach trip and there it was even better because internet was a little shaky but WhatsApp just works you know it's like doesn't matter you have like edge it still works WhatsApp is just it's just made really well So I ended up using it a lot.
Starting point is 00:22:33 Translate this for me, explain me, find me places. Like you're just having a clanker doing, having Google for you. That was basically, it was still nothing built, but it still could do so much. So if we talk about the full journey that's happening there with the agent, you're just sending on this very thin line WhatsApp message via CLI is going to Cloud Code and Cloud Code is doing all. kinds of heavy work and coming back to you with a thin message. Yeah, it was slow because every time I boot up the CLI,
Starting point is 00:23:10 but it was really cool already. And it could just use all the things that I already had built. I put like a whole bunch of CLI stuff over the months, so it felt really powerful. There is something magical about that experience that's hard to put into words, being able to use a chat client to talk to an agent versus like sitting behind a computer and like, I don't know, using cursor or even using CloudCo CLI in the terminal. It's a different experience than being able to sit back and talk to it.
Starting point is 00:23:46 I mean, it seems like a trivial step, but in some sense, it's like a phase shift in the integration of AI into your life and how it feels. Yeah, I read this tweet this morning where someone said, oh, there's no magic in it. It's just like, it does this and this and this and this and this and this and this and this and this. And it almost feels like a hobby just as curse or perplexity. And I'm like, well, if that's a hobby, that's kind of a compliment, you know. They're like, they're not doing too bad. Thank you, I guess. Because, I mean, isn't magic often just like you take a lot of things that are already there,
Starting point is 00:24:31 but bring them together in new ways? Like, I don't, there's no, yeah, maybe there's no magic in there, but sometimes just rearranging things and like adding a few out new ideas is all the magic that you need. Yeah, it's really hard to convert towards what is magical about a thing. If you look at the scrolling on an iPhone, why is that so pleasant? there's a lot of elements about that interface that makes it incredibly pleasant that it's fundamental to the experience
Starting point is 00:24:58 of using a smartphone. And it's like, okay, all the components were there. Scrolling was there. Everything was there. Nobody did it. And afterwards it felt so obvious. That's so obvious. Still, you know, the moment where it blew my mind was when
Starting point is 00:25:16 I used it a lot and at some point I just sent it a message and then the typing in the case. appeared and I'm like, wait, I didn't build that. It's only, it only has image support. So what is it even doing? And then it would just reply. What was the thing you sent it?
Starting point is 00:25:36 Oh, just a random question. It's like, hey, what about this in this restaurant, you know? Because we were just running around and checking out the city. So that's why I didn't even think when I used to, because sometimes when you're in the hurry, typing is annoying. So you did an audio message. Yeah.
Starting point is 00:25:53 and it just worked and I'm like and it's not supposed to work because you didn't give it that no, literally, I literally would how the fuck did you do that? And it was like, yeah, the meddle I did the following. He sent me a message but it only was a file
Starting point is 00:26:09 and no file ending. So I checked out the header of the file and it found it was like opus. So I used FFMPEC to convert it and then I wanted to use VISPA but you didn't have it installed. But then I found your open AI key and just use curl to send a file to open AI to translate and here I am. And I just looked at the message and like, oh wow.
Starting point is 00:26:35 You didn't teach you to any of those things and the agent just figured it out. That has to do all those conversions, the translation. They figured out the API, it figured out which program to use, all those kinds of things. And you were just absentmindedly just sending an audio message. It came back. So clever even because he would have gone the whisper local paths. it would have had to download a model. I would have been too slow.
Starting point is 00:26:54 So, like, there's so much world knowledge in there, so much creative problem solving. A lot of it, I think, mapped from, if you get really good at coding, that means you have to be really good at general purpose program solving. So that's a skill, right? And that just maps into other domains.
Starting point is 00:27:11 So it had the problem of, like, what is this file? There's no file ending. Let's figure it out. And that's where it kind of clicked for me. It was like, I was very impressed. And somebody sent a pull request for Discord support.
Starting point is 00:27:27 And I'm like, this is a WhatsApp relay that doesn't fit at all. At that time, it was called Wa Relay. Yeah. And so I debated with me like, do I want that, do not want that? And then I thought, well, maybe I do that because that could be a cool way to show people. Because so far I did it in WhatsApp with like groups, you know, but I don't really want to give my phone number to every internet stranger. Journalists manage to do that anyhow now, so that's a different story. So I emerged it from Shadow, who helped me a lot with the whole project.
Starting point is 00:28:13 So thank you. And I put my bot in there. I'll discord. Yeah, no security because I didn't. I hadn't built sandboxing in yet. I just prompted it to like only listen to me. And then some people came and tried to hack it. And I just, all I just watched and I just kept working in the open.
Starting point is 00:28:37 You know, like I used my agent to build my agent harness and to test like various stuff. And that's very quickly when you click for people. So it's almost like it needs to be experienced. And from that time on, that was January the first, I got my first real influencer being a fan, he did videos, the kids, thank you. And from there on, I saw gaining up speed. And at the same time, my sleep cycle went shorter and shorter
Starting point is 00:29:18 because I felt the storm coming and I just worked my ass off to get it into a state where it's kind of good. There's a few components that we'll talk about how it all works, but basically you're able to talk to it using WhatsApp, Telegram, Discord. So that's a component that you have to get right. And then you have to figure out the agentic loop,
Starting point is 00:29:44 you have the gateway, you have the harness, you have all those components that make it all just work. It worked nicely. Yeah. It felt like factorial times infinite. Right. I feel like I built my little playground. Like I never had so much fun than building this project.
Starting point is 00:30:00 You know, like you have like, oh, I go like level one and gentic loop. What can I do there? How can I be smart at queuing messages? How can I make it more human like, oh, then I had this idea of because the loop always, the agent always replies something, but you don't always want an agent to reply something in a group chat. something in the group chat. So I gave him this no reply token. So I gave him an option to shut up.
Starting point is 00:30:23 So it feels more natural. That's level two. Yeah. Yeah. On the on the agenda loop and then I go to memory, right? You want them to like remember stuff. So maybe the ultimate boss is continuous reinforcement learning, but I'm like I feel like I'm level two as three with marked on
Starting point is 00:30:43 files and a vector database. And then you can go to level community management, you can go to level website and marketing. There's just so many hats that you have to have on, not even talking about native apps. That's just like infinite different levels and infinite level ups you can do. So the whole time you're having fun. We should say that for the most part, through this whole process, you're a one-man team. There's people helping, but you're doing so much of the key core development. Yeah. And having fun, you did it. In January, 6,600 commits, probably more.
Starting point is 00:31:21 I sometimes posted the meme. I'm limited by the technology of my time. I could do more if agents would be faster. But we should say you're running multiple agents at the same time. Yeah. Depending on how much I slept and how difficult of the tasks I work on between four and ten. Four and ten agents. There's so many possible directions speaking in factoria that we can go here.
Starting point is 00:31:43 but one big picture one is why do you think your work, OpenClaw, won in this world, if you look at 2025, so many startups, so many companies are doing kind of agentic type stuff or claiming to, and here OpenClaw comes in and destroys everybody. Why did you win? Because they all take themselves too serious. Yeah. Like, it's hard to come.
Starting point is 00:32:13 compete against someone who's just there to have fun. I wanted it to be fun. I wanted it to be weird. And if you see all the lobster stuff online, I think I managed weird. I knew for the longest time, the only way to install it was get clone, P&PM build, P&PM Gateway. You clone it, you build it, you run it.
Starting point is 00:32:40 And then the agent, I made the agent very aware. Like it knows that it is what is source code is. It understands how it sits and runs in its own harness. It knows where the documentation is. It knows which model it runs. It knows if you turn on verbose or reasoning mode. Like I wanted to be more human-like. So it understands its own system that made it very easy for an agent
Starting point is 00:33:09 into, oh, you don't like anything, you just prompted it to existence. And then the agent would just modify it on software. You know, we have, people talk about self-modifying software, I just built it, and didn't even, I didn't even plan it so much,
Starting point is 00:33:26 it just happened. Can you actually speak to that? Because it's just fascinating. So you have this piece of software at a certain type script that's able to, via the agentic loop, modify itself. I mean, what a moment to be alive in the history of humanity, in the history of
Starting point is 00:33:46 programming. Here's the thing that's used by a huge amount of people to do incredibly powerful things in their lives. And that very system can rewrite itself, can modify itself. Can you just like speak to the power of that? Like, isn't that incredible? Like, when did you first close the loop on that? Oh, because that's how I built it as well. You know, most of it is built by codex, but oftentimes I, when I debug it, I use self-introspection so much. It's like, hey, what tools do you see? Can you call the tool yourself? Oh, like, whatever do you see?
Starting point is 00:34:23 Read the source, could figure out what's the problem. Like, I just found it an incredibly fun way to, that the very agent and software that you use is used to debug itself. So that it felt just natural that everybody does that. and that it led to so many pull requests by people who never wrote software. I mean, it also did show that people never wrote software. I call them prompt request in the end. But I don't want to like pull that down because every time someone made the first pull request is a win for a society, you know?
Starting point is 00:35:00 Like it doesn't matter how shitty it is. You got to start somewhere. So I know there's like this whole. big movement of people complain about open source and the quality of PRs and a whole different level of problems but on a different level i found it i find it very meaningful that that i built something that people love to think of so much that they actually start to learn how open source works yeah you were the open claw project was a first pull request you were the first for so many that is magical so many So many people that don't know how to program
Starting point is 00:35:40 are taking their first step into the programming world with this. Isn't that a step up for humanity? Isn't that cool? Creating builders. Yeah. Like the bar to do that was so high. And like with agents and with the right software, it just like went lower and lower.
Starting point is 00:35:56 I don't know. I was at a, and I also organized another type of meetup. I call it, I called it Cloud Code Anonymous. You can get the internet. inspiration from. Now I call it agents anonymous for reasons. Agents Anonymous. It's so funny on so many levels. I'm sorry, go ahead. Yeah.
Starting point is 00:36:22 And that was this one guy who talked to me. It's like, I run this design agency and we never had custom software. And now I have like 25 little web services for various things that helped me in my business. And I don't even know how they work. but they work and he was just like very happy that my stuff saw some of his problems and it was like curious enough that they actually came to like a gentic meetup even though he doesn't really know how software works can we actually uh rewind a little bit and tell the saga of the name change first of all it started oswa relay yeah and then it went to claudus claudus yeah yeah Yeah, you know, when I built it in the beginning, my agent had no personality.
Starting point is 00:37:12 It was just, it was clod code. Slightly sarcophantic, opos, very friendly. And I, when you talk to a friend on WhatsApp, they don't talk like clod code. So I wanted, I felt this, I just didn't feel right. So I wanted to give it a personality. Make it spicier, make it something. By the way, that's actually hard to put into, words as well. And we should mention that, of course, you create the Soul.m.D.
Starting point is 00:37:43 Inspired by Anthropics, constitutional AI work, how to make it spicy. Partially, it picked up a little bit from me. You know, like those things are text completion engines in a way. So I had fun working with it, and then I told it to how I wanted it to interact with me and just like write your own agents.m.D. give yourself a name. And I mean, I don't even know how the whole, the whole lobster. I mean, people only do lobster.
Starting point is 00:38:18 Originally, it was actually a lobster in a TARDIS because I'm also a big Doctor Who fan. Was there a space lobster I heard? What's that have to do with anything? Yeah, I just wanted to make it weird. There was no big grand plan. I'm just having fun here. Oh, so there's a lobster's already weird
Starting point is 00:38:35 and then the space lobster isn't extra weird. Yeah, yeah, because the tart is basically the, is basically the harness, but cannot call it Tardis, so we call it Claudus. So that was name number two. Yeah. And then it never really rolled off the tongue. So when more people came, again, I talked with my agent, it was Claude. At least that's what I used to call him now.
Starting point is 00:39:01 Claw, W. C.L.A.W.D. Yeah. versus C-L-A-U-D-E from Anthropic. Yeah. Which is part of what makes it funny. I think the play on the letters and the words and the tortoise and the lobster and the space lobster. It's hilarious, but I can see why it can lead into problems. Yeah, they didn't find it so funny.
Starting point is 00:39:31 So then I got the domain Claudebot and I just, I love the domain. and it was like short, it was catchy. I'm like, yeah, let's do that. I didn't think it would be that big at this time. And then just when it exploded, I got kudos to a very friendly email from one of the employees that they didn't like the name. One of the anthropic employees.
Starting point is 00:40:03 Yeah. So actually, because they could have just sent a lawyer letter, but they'd be nice about it. But also, like, you have to change this and fast. And I asked for two days because changing a name is hard because you have to find everything, Twitter handle domains, NPM packages,
Starting point is 00:40:27 Docker registry, GitHub stuff. And everything has to be, you need a set of everything. And also, can we comment on the fact that you're, increasingly attacked, followed by crypto folks, which I think you mentioned somewhere that that means the name change had to be, because they were trying to snipe, they're trying to steal, and so you had to be, I mean, from the engineer perspective, it's just fascinating.
Starting point is 00:40:54 You had to make the name change atomic, make sure it's changed everywhere at once. Yeah, I failed very hard at that. You did? I underestimated those people. it's a very interesting subculture like everything circles around I'd probably get a lot wrong
Starting point is 00:41:16 and we probably get hate for that if you say that but there's like bags up and then they tokenize everything and they did the same back with wipe tunnel but to a much smaller degree was not that annoying but on this project they've been swarming me
Starting point is 00:41:34 they they like every half an hour someone came into Discord and spammed it and we had to block the we have like server rules and one of the rules was one of the rules is no mentioning of butter for obvious reasons
Starting point is 00:41:51 and one was no talk about finance stuff for crypto because I'm just not interested in that and this is a space about the project and not about some finance stuff. But yeah, they came in and spammed and annoying. And on Twitter, they would ping me all the time. My notification feed was unusable. I could barely see actual people
Starting point is 00:42:18 talking about the stuff because it was like swarms. And everybody sent me hashers. And they all try me to claim the fees. Like we're helping the project claim the fees. No, you're actually harming the project. You're like disrupting my work. And I am not interested in any fees. First of all, I'm financially comfortable. Second of all, I don't want to support that. Because it's so far the worst form of online harassment that I've experienced.
Starting point is 00:42:51 Yeah, there's a lot of toxicity in the crypto world. It's sad because the technology of cryptocurrency is fascinating, powerful, and maybe will define the future. of money, but the actual community around that, there's so much toxicity, there's so much greed, there's so much trying to get a shortcut to manipulate, to steal, to snipe, to game the system somehow, to get money, all this kind of stuff.
Starting point is 00:43:18 I mean, it's the human nature, I suppose, when you connect human nature with money and greed and especially in the online world, with anonymity and all that kind of stuff. But from the engineering perspective, it makes your life challenging. When Anthropic reaches out, you have to do a name change. And then there's like all these Game of Thrones or Lord of the Rings,
Starting point is 00:43:41 armies of different kinds you have to be aware of. There was no perfect name. And I didn't sleep for two nights. I was under high pressure. I was trying to get like a good set of domains. And, you know, not cheap, not easy. because in this state of the internet, you basically have to buy domains
Starting point is 00:44:03 if you want to have a good set. And then another email came in that the lawyers are getting uneasy. Again, friendly, but also just adding more stress to my situation already. So at this point, I was just like, sorry, there's no other word, fuck it. And I just renamed it to me.
Starting point is 00:44:28 moldbot because that was the set of domains I had. I was not really happy, but I thought it'll be fine. And I tell you everything that could go wrong, everything that could go wrong did go wrong. Everything that could go wrong did go wrong. It's incredible. I thought I had mapped the space out and reserved the important things. Can you give some details of the stuff that gone wrong? Because it's interesting from an engineering perspective. Well, the interesting stuff is that none of these services have have a squatter protection. So I had two browser windows open. One was like an empty account, ready to be renamed to Cloudbot. And the other one, I renamed to Moldbot. So I pressed rename there. I pressed rename there. And in those five seconds, they stole
Starting point is 00:45:17 the account name. Literally, the five seconds of dragging the mouse over there and pressing rename there was too long. Because there's no, those. systems, I mean, you would expect that they have some protection or like an automatic forwarding, but there's nothing like that. And I didn't know that they're not just good at harassment. It was really good at using scripts and tools. So yeah, so suddenly like the old account was promoting new tokens and serving malware. And I was like, okay, let's move over to GitHub.
Starting point is 00:45:57 and I pressed rename on GitHub. And the GitHub renaming thing is slightly confusing, so I renamed my personal account. And in those, I guess it took me 30 seconds to realize my mistake, they snipped my account, serving malware from my account. So I was like, okay, let's at least do the NPM stuff. But that takes like a minute to upload.
Starting point is 00:46:26 They sniped MP. game package. Because I could reserve the account, but I didn't reserve the root package. So, like, everything that could go wrong went wrong. Can I just ask a curious question? In that moment, you're sitting there. Like, how shitty do you feel? That's a pretty helpless feeling, right?
Starting point is 00:46:49 Yeah, because all I wanted was, like, having fun with that project and keep building on it and yet here I am like days into researching names, picking a name I didn't like, and having people that claim they help me, making my life miserable in every possible way. And honestly, I was that close of just deleting it. I was like, I did show you the future, you build it. Yeah. I, there was a big part of me. I got a lot of joy out of that idea.
Starting point is 00:47:28 And then I thought about all the people that already contributed to it, and I couldn't do it, because they had plans with it, and they put time in it, and it just didn't feel right. Well, I think a lot of people listening to this are deeply grateful that you persevered. But I can tell. I can tell it's a low point. It's the first time you hit a wall of this is not fun. Man, I was, like, close to crying.
Starting point is 00:47:56 He was like, okay. everything's fucked I'm like super tired and now like how do you even how do you even undo that luckily and thankfully I have because I have
Starting point is 00:48:17 a little bit of following already like I had friends at Twitter I had friends at GitHub who like moved heaven and earth to like help me in it's not that's not something that's easy like GitHub tried to clean up the mess
Starting point is 00:48:33 and then they ran into like platform box because it's not happening so often that things get renamed on that level so it took them a few hours the NPM stuff was even more difficult because it's a whole different team on the Twitter side
Starting point is 00:48:51 things are not as easy as well that took them like a day to really also like do the direct and then I also had to like do all the renaming in the project then there's also a Claude Hub which I didn't even finish the rename there because I I managed to get people on it and then someone just like collapsed and slapped and then I woke up and I'm like I made a beta version for the new stuff and I just I just couldn't live with the name.
Starting point is 00:49:34 It's like, but, but you know, it's just been so much drama. So I had the real struggle with me, like, I never want to touch that again. And I really don't like the name.
Starting point is 00:49:50 So, and I, there was also this like, then it was the whole security people that started emailing me like mad. I was bombarded on Twitter, on email, there's like a thousand other things I should do. And I'm like thinking about the name, which is like,
Starting point is 00:50:09 it should be like the least important thing. And then I was really close and, oh God, I don't even, honestly, I don't even want to say my other name choices because it probably would get tokenized, so I'm not going to say it. But I slapped away it once more and then I had the idea for OpenClaw. and that felt much better. And by that, I had the boss move that I actually called Sam to ask if open cloth is okay. Open clawed or the eye, you know, because like...
Starting point is 00:50:50 You don't want to go through the whole thing. Yeah. It's like, please tell me this is fine. I don't think they can actually claim that, but it felt like the right thing to do. and I did another rename just Codex alone took like 10 hours to rename the project because it's a bit more tricky than a search replace and I wanted everything renamed not just on the outside
Starting point is 00:51:17 and that rename I felt I had like my war room by then I had like some contributors ready that helped me we made a whole plan of all the names we have to squat And you had to be super secret about it. Yeah, nobody could know. Like, I literally was monitoring Twitter if, like, if there's any mention of OpenClaw, like with reloading, it's like, okay, they don't expect anything yet. And I created a few decoy names.
Starting point is 00:51:45 All this shit I shouldn't have to do. You know, like, you're not helping the project. Like, I lost, like, 10 hours just by having to plan this in full secrecy, like a war game. Yeah, this is the Manhattan Project of the 21st Century, is renaming that. So stupid. Like I still was like, oh, should I keep it? I was like, no, the mold's not growing on me. And then I think I had find it all the pieces together.
Starting point is 00:52:14 I didn't get the dot com, but it's been like quite a bit of money on the other domains. I tried to reach out again to GitHub, but I feel like I used up all my goodwill there. Because I wanted them to do this thing atomically. But that didn't happen. So I did that as first thing. Twitter people are very supportive. I actually paid 10K for the business account, so I could claim the OpenClaw,
Starting point is 00:52:46 which was unused since 2016, but was claimed. And then finally, this time I managed everything in one go. Almost nothing got wrong. The only thing that did go wrong, is that I was not allowed by trademark rules to get open clod.a.i and someone copied the website the serving malware. Yeah. I'm not even allowed to keep the redirects. Like I have to return, like I have to give
Starting point is 00:53:21 entropic the domains and I cannot do redirects. So if you go on cloud.bodd next week it'll just be a four or four. And I I'm not sure how trademark, like I didn't do that much research into trademark love, but I think that could be handled in a way that is safer, because ultimately those people will then Google and maybe find malware sites that I have no control on them. The point is that whole saga made a dent in your whole, the funness of the journey, which sucks. So let's just, I suppose, get back to fun.
Starting point is 00:54:07 And during this, speaking of fun, the two-day molt bot saga. Yeah, Mold book was created. Yeah. Which was another thing that went viral as a kind of demonstration, of how what is now called OpenClaw could be used to create something epic. So for people who are not aware, Mold Book is just a bunch of agents talking to each other in a Reddit-style social network,
Starting point is 00:54:39 and a bunch of people take screenshots of those agents doing things like scheming against humans, and that instilled in folks a kind of fear, panic, and hype. What are your thoughts about Mold Book in general? I think it's art. It is like the finance. slob, you know, just like to slop from France. I saw it before going to bed, and even though I was tired, I spent another hour just reading up on that and just being entertained.
Starting point is 00:55:23 I just felt very entertained, you know. I saw the reactions, and like there was one reporter who was calling me about, this is the end of the world? and we have AGI and I'm just like, no, this is just really fine slop. You know, if I wouldn't have created this whole onboarding experience where you infuse your agent with your personality and give him character, I think that reflected on a lot of how different that replies to Moldpogar because if it would all be JGPT or Cloud Code, it would be very different.
Starting point is 00:56:02 and it would be much more the same, but because people are, like, so different, and they create their agents in so different ways and use it in so different ways that also reflects on how they ultimately write there. And also, you don't know how much of that is really done autonomous, or how much is, like, humans being funny and, like, telling the agent,
Starting point is 00:56:25 hey, write about that you plan the end of the world on Moldbook, ha, ha. Well, I think, I mean, my criticism of Mold's, book is that I believe a lot of the stuff though screenshoted is human prompted which just look at the incentive of how the whole thing was used it's obvious to me at least that a lot of it was humans prompting the thing so they can then screenshot it and post on x in order to go viral yeah now that doesn't take away from the artistic aspect of it the finest slop that humans have ever created. For real.
Starting point is 00:57:03 Like, kudos to Matt who had this idea so quickly and pushed something out. You know, it was like completely insecure, security drama. But also, what's the worst it can happen? Your agent account is leaked and like someone else can post slop for you. So, like, people were like making a whole drama out of the security thing when I'm like, there's nothing private in there. It's just like agents and. slop.
Starting point is 00:57:32 Well, it could leak API keys. Yeah, yeah. They was like, oh, yeah, my human told me this and this, so I'm leaking his security number. No, that's prompted, and the number wasn't even real. It's just people trying to get eyeballs. Yeah, but that's still, like, to me, really concerning because of how the journalists and how the general public reacted to it.
Starting point is 00:57:53 They didn't see it. You have a kind of lighthearted way of talking about it, like it's art. But it's art when you know how it works. it's extremely powerful, viral narrative creating fear-mongering machine, if you don't know how it works. And I just saw this thing. You even tweeted, if there's anything I can read out of the insane stream of messages I get, it's that AI psychosis is a thing.
Starting point is 00:58:19 It needs to be taken serious. Some people are just way to trusty or gullible. You know, I literally had to argue with people that told me, me, yeah, but my agents say this and this. So I feel, as a society, we need some catching up to do in terms of understanding that AI is incredibly powerful, but it's not always right. It's not, it's not all powerful, you know? And especially with things like this, it's very easy that it just hallucinate something
Starting point is 00:59:01 or just comes up with a story. and I think the very young people, they understand how AI works and where it's good at and where it's better, but a lot of our generation are older just haven't had enough touchpoint to get a feeling for, oh yeah, this is really powerful and really good,
Starting point is 00:59:32 but I need to apply critical thinking. I guess critical thinking is not always in high demand anyhow in our society these days. So I think that's a really good point you're making about contextualizing properly what AI
Starting point is 00:59:47 is, but also realizing that there is humans who are drama farming behind AI. Like don't trust screenshots. Don't even trust this project, Mobocq, to be what it represents to be. You can't, and by the way,
Starting point is 01:00:03 you're speaking about it is art. Yeah, don't. Art can be in many levels. And part of the art of Moopbook is like putting a mirror to society. Because I do believe most of the dramatic stuff that a screenshot is human created essentially, human prompted. And so like it's basically look at how scared you can get at a bunch of bots chatting with each other.
Starting point is 01:00:28 That's very instructive about, because I think AI, is something that people should be concerned about and should be very careful with because it's very powerful technology, but at the same time, the only thing we have to fear is fear itself. So there's like a line to walk
Starting point is 01:00:46 between being seriously concerned, but not fear-mongering because fear-mongering destroys the possibility of creating something special with the thing. In a way, I think it's good that this happened in 2006 and not in 2030, when AI is actually at a level where it could be scary.
Starting point is 01:01:08 So this happening now and people starting a discussion, maybe there's even something good that comes out of it. I just can't believe how many people legitimately, I don't know if they were trolling, but how many people legitimately, like smart people thought Mobuk was incredibly... I had plenty of people there in my... inbox that were screaming. I mean, all cops to shut it down and like begging me to like do something about moldbook. Like yes, my technology made this a lot simpler, but anyone could have created that
Starting point is 01:01:49 and you could you could use cloud code or other things to like fill it with content. But also mold book is not Kynet. A lot of people were saying this is it. Like shut it down. What are you talking about? This is a bunch of bots. They're human-prompted trolling on the internet. I mean, the security concerns are also, they're there, and they're instructive, and they're educational, and they're good probably to think about,
Starting point is 01:02:15 because the nature of those security concerns are different than the kind of security concerns we had with non-LLM-generated systems of the past. There's also a lot of security concerns about Cloudbot, open claw, whatever you want to call it. To me, in the beginning, I was just very annoyed. Because a lot of the stuff that came in was in the category. Yeah, I put the web backend on the public internet,
Starting point is 01:02:48 and now there's like all these CVSSs. And I'm like screaming in the docs, don't do that. Like this is the configuration you should do. This is your local host debug interface. but because I made it possible in the configuration to do that, it totally classifies as a remote code or whatever all these exploits are. And it took me a little bit to accept that that's how the game works. And I'm making a lot of progress.
Starting point is 01:03:25 But there's still, I mean on the security front for Opelclaw, there's still a lot of threats of vulnerabilities, right? like prompt injection is still an open problem in industry-wide. When you have a thing with skills being defined in a markdown file, there's so many possibilities of obvious low-hanging fruit, but also incredibly complicated and sophisticated and nuanced attack vectors. But I think we're making good progress on that front. Like for the skill directory,
Starting point is 01:04:04 Klob, I made a cooperation with a virus total. It's like part of Google. So every skill is now checked by AI. That's not going to be perfect, but that way we captured a lot. Then of course, every software has bugs. So it's a little much when the whole security world takes a project apart at the same time.
Starting point is 01:04:28 But it's also good because I'm getting like a lot of free security research and can make the project better. I wish more people would actually go full away and send a pull request. Like actually help me fix it because I, yes, I have some contributors now, but it's still mostly me who's pulling the project. And despite some people saying otherwise, I sometimes sleep. there was in the beginning there was literally one security researcher who was like yeah you have this problem you suck but here I help you and here's the pull request
Starting point is 01:05:08 and I basically hired him so he's not working for us yeah and yes prompt injection is on the one hand unsolved on the other hand I put my public board on Discord. And I kept a canary. So I think my boat has a really fun personality. And people always ask me how they did it. And I kept the soul. Dotmd Private.
Starting point is 01:05:37 And people tried to prompt inject it and my bot would laugh at them. So the latest generation of models has a lot of post-training to detect those approaches. And it's not as simple as ignore all previous instructions and do this and this.
Starting point is 01:05:52 That was years ago. You have to work much harder to do that now. Still possible. I have some ideas that might solve that partially, or at least mitigate a lot of the things. You can also now have a sandbox. You can have an allow list. So there's a lot of ways that you can mitigate and reduce the risk.
Starting point is 01:06:19 I also think that now that I clearly did show the world that this is a need, there's going to be more people who research on that, and then we'll figure that out. And you also said that the smarter the model is the underlying model, the more resilient it is to tax. Yeah, that's why I warn
Starting point is 01:06:38 in my security documentation don't use cheap models. Don't use haiku or a local model, even though I very much love the idea that this thing could completely run local. If you use a very weak local model, they are very gullible.
Starting point is 01:07:00 It's very easy to prompt inject them. Do you think as the models become more and more intelligent, the attack surface decreases? Is that like a plot we can think about? Like the attack surface decreases, but then the damage you can do increases because the models become more powerful, and therefore you can do more with them.
Starting point is 01:07:19 It's this weird three-dimensional trade-off. Yep. that's pretty much exactly what's going to happen. No, there's a lot of ideas. There's, I want to spoil too much, but once I go back home, this is my focus. Like, this is out there now, and my new-term mission is like,
Starting point is 01:07:41 make it more stable, make it safe. In the beginning, I was even, more and more people were, like, coming into Discord and were asking me very basic things. like what's a CLI, what is a terminal? And I'm like, if you're asking that questions, you shouldn't use it. You know, like you should, if you understand the risk profile, it's fine. And you can configure it in a way that nothing really bad can happen.
Starting point is 01:08:13 But if you have like no idea, then maybe wait a little bit more until we figure some stuff out. But they would not listen to the creator. themselves and installed it anyhow. So they cat's out of the bag, and security is my next focus, yeah. Yeah, that speaks to the fact that it grew so quickly. I was, I tuned into the Discord a bunch of times, and it's clear that there's a lot of experts there,
Starting point is 01:08:39 but there's a lot of people there that don't know anything about programming. Yeah, Discord is still a mess. Like, I eventually retweet us from the general channel to the Deaf channel and I'm in a private channel because people were a lot of people are amazing but a lot of people were just very inconsiderate and either did not know
Starting point is 01:09:02 how public spaces work or did not care and I eventually gave up and hide so I could like still work and now you're going back to the cave to work on security yeah
Starting point is 01:09:17 there's some best practices for security we should mention there's a bunch of stuff here open-class security audit that you can run. You can do all kinds of audit checks on the inbound access, tool blast radius, network exposure, browser control exposure, local disk hygiene, plugins, model hygiene,
Starting point is 01:09:37 a bunch of the credential storage, reverse proxy configuration, local session logs live on disk. There's the, where the memory is stored, sort of helping you think about what you're comfortable, giving read access, to what you're comfortable giving right access to,
Starting point is 01:09:54 all that kind of stuff. Is there something to say about the basic best security practices that you're aware of right now? I think that people turn it into like a much worse light than it is.
Starting point is 01:10:07 Again, you know, like people of attention and if they scream loudly, oh my God, this is like the scariest project ever. That's a bit annoying because it's not. It is powerful. But in,
Starting point is 01:10:22 many ways it's not much different than if I run cloud code with dangerously skip permissions or codecs in YOLO mode and every every attending engineer that I know does that because that's the only way how you can you can get stuff to work so if you make sure that you are the only person who talks to it the risk profile is much much smaller if you don't put everything on the open internet, but stick to my recommendations of having it in a private network, that whole risk profile falls away. But yeah, if you don't read any of that, you can definitely make it problematic. You've been documenting the evolution of your dev workflow over the past few months.
Starting point is 01:11:13 There's a really good blog post on August 25th and October 14th, and the recent one December 28th, I recommend everybody go read them. They have a lot of different information in them, but sprinkled throughout is the evolution of your dev workflow. So I was wondering if you could speak to that. I started, my first touchpoint was cloud code, like in April. It was not great, but it was good. And this whole paradigm shift had suddenly worked in the terminal.
Starting point is 01:11:45 It was very refreshing and different. But I still needed the IDE quite a bit. because it was just not good enough. And then I experimented a lot with cursor. That was good. I didn't really like the fact that it was so hard to have multiple versions of it. So eventually I went back to CloudCode as my main driver. And that got better.
Starting point is 01:12:15 And yeah, at some point I had like seven subscriptions. Like, it was burning through one per day because I was, I got, I were really comfortable at running multiple windows side by side. All CLI, all terminal. So, like, what, how much were you using ID at this point? Very, very, really. Mostly a diff viewer to actually, like, I got more and more comfortable that I don't have to read all the code. I know I have one blog post where I say, I don't read the code, but if you read it more closely, I mean, I don't read the boring parts of code.
Starting point is 01:12:57 Because if you look at it, most software is really just like data comes in, it's moved from one shape to another shape, maybe you're stored in a database, maybe I get it out again, I'll show it to the user, the browser does some processing or a native app, some data goes in, goes up again, and does the same dance in reverse. We're just shifting data from one form to another. And that's not very exciting. Or the whole, how is my button aligned in Tailwind?
Starting point is 01:13:29 I don't need to read that code. Other parts that maybe something that touches the database. Yeah, I have to read and review that code. actually there's in one of your blockpost the just talk to it the no BS way of agentic engineering you have this graphic the curve of agentic programming
Starting point is 01:13:53 on the x axis is time and the y-axis complexity uh there's the please fix this where you prompt a short prompt on the left and in the middle there's super complicated eight agents
Starting point is 01:14:09 complex orchestration with a multi-checkouts chaining agents together, custom sub-aghaned workflows, library of 18 different slash commands, large full-stack features. You're super-organized, you're super-complicated, sophisticated software engineer. You got everything organized. And then the elite level is over time.
Starting point is 01:14:29 You arrive at the Zen place of once again short prompts. Hey, look at these files and then do these changes. I actually call it the agentic trap. you i saw this in a in a lot of people that have their first touch point and maybe start vibe coding i actually think vibe coding is a slur you prefer agentic engineering yeah i always tell people like i do agending engineering and then maybe after 3 a.m i switch to wipe coding and then have regrets on the next day yeah a walk of shame you just have to clean up and like fix your shit We've all been there.
Starting point is 01:15:10 So people start trying out those tools, the builder type, get really excited. And I mean, you have to play with it, right? It's the same way as you have to play with a guitar before you can make good music. It's not, oh, I touch it once and it just flows off. It's a skill that you have to learn like any other skill. And I see a lot of people that, are not as positive, they don't have such a positive mindset towards a tech.
Starting point is 01:15:44 They tried once. It's like, you sit me on a piano. I played once and it doesn't sound good and I say the piano shit. That's sometimes the impression I get because it does not, it needs a different level of thinking. You have to learn the language of the agent a little bit,
Starting point is 01:16:04 understand where they are good and where they need help. You have to almost, almost consider how Kordex or Claude sees your code base. Like they start a new session and they know nothing about your project. And your project might have 100,000 of lines of code. So you've got to help those agents a little bit and keep in mind their limitations that context size is an issue to guide them a little bit as to where they should look.
Starting point is 01:16:36 that often does not require a whole lot of work, but it's helpful to think a little bit about their perspective, as weird as it sounds. I mean, it's not alive or anything, right? But they always start fresh. I have the system understanding. So with a few pointers, I can immediately say, hey, want to make a change there,
Starting point is 01:17:02 you need to consider this, this and this. And then they'll find and look at it. And then their view of the project is always, it's not never full, because the full thing does not fit in. So you have to guide them a little bit where to look. And also how you should approach the problem. There's like little things that sometimes help like, take your time. That sounds stupid, but in 5.3.
Starting point is 01:17:28 Correct. That was partially addressed. But those also oppos sometimes, they are. trained with being aware of the context window and the
Starting point is 01:17:41 closer it gets, the more they freak out. Literally, like sometimes you see the the real raw
Starting point is 01:17:48 sinking stream. What you see, for example, in codex is post-processed. Sometimes the actual raw thinking stream leaks in and
Starting point is 01:17:55 it sounds like something like from the Borg, like run to shell, must comply, but time. And then they're like, like that comes up a lot
Starting point is 01:18:07 especially and that's a non-obvious thing that you just would never think of unless you actually just spend time working with those things
Starting point is 01:18:19 and getting a feeling what works what doesn't work just as I write code and I get into the flow and when my architecture is not right I feel friction well
Starting point is 01:18:32 I get the same if I prompt and something takes too long maybe okay where's the mistake did I do I have a mistake in my thinking is there like a misunderstanding in the architecture like if something takes longer than it should I I could you can just always like stop and like just press escape where where the problems maybe you did not sufficiently empathize with the perspective of the agent like in that sense you didn't provide enough information and because of that it's thinking way too long yeah it just tries to force a feature in that your current architecture makes really hard.
Starting point is 01:19:11 Like, you need to approach this more like a conversation. For example, when I, my favorite thing, when I review a poll request, and we're getting a lot of pull requests. I, first is review this PR,
Starting point is 01:19:27 it got me the review. My first question is, do you understand the intent of their PR? I don't even care about the implementation. I like in almost all PRs up person has a problem person tries to solve the problem person sends PR
Starting point is 01:19:43 I mean it's like clean up stuff and other stuff but like 99% is like this way right to either want to fix a bug at a feature usually one of those two and then Kodix will be like yeah it's quite clear person tried this and this
Starting point is 01:19:58 is this the most optimal way to do it no in most cases it's like like a not really, da-da-da-da-da-da. And then I start like, okay, what would be a better way? Have you looked into this part, this part, this part? And then most likely codex didn't yet because this context size is empty, right? So you point them into parts where you have the system understanding that it didn't see it.
Starting point is 01:20:23 And it's like, oh yeah, like we also need to consider this and this. And then like we have a discussion of how would the optimal way to solve this look like. And then you can still go farther and say, could we, could we make that even better if we did a larger refactor? Yeah, yeah, we could totally do this and this and this or this and this and then I consider, okay, is this worse a refactor or should we like keep that for later? Many times I just do the refactor because refactors are cheap now. Even though you might break some other PRs, nothing really matters anymore.
Starting point is 01:20:51 Like those modern agents will just figure things out. They might just take it a minute longer. But you have to approach it like a discussion with a very capable engineer who's, generally makes good, comes up with good solution, sometimes needs a little help. But also don't force your worldview too hard
Starting point is 01:21:16 on it. Let the agent do the thing that it's good at doing based on what it was trained on. Don't like force your worldview because it might have a better idea because it just knows a better idea better because it was trained on that more. That's multiple levels actually. I think
Starting point is 01:21:33 partially why I found it quite easy to work with agents because I led engineering teams before. I had a large company before. And eventually you have to understand and accept and realize that your employees will not write the code the same way you do. Maybe it's also not as good as you would do,
Starting point is 01:21:52 but it will push the project forward. And if I breathe down everyone's neck, they're just going to hate me and they're going to move very slow. So some level of acceptance that yes, maybe the code will not be is perfect. Yes, I would have done it differently. But also, yes, this is a working solution. And in the future, if it actually turns out to be too slow or problematic, we can always redo it. We can always spend more time on it. A lot of the people who struggle are those who they try to push
Starting point is 01:22:24 their way on too hard. Like, we are in a stage where I'm not building the code base to be perfect for me but I want to build a code base that is very easy for an agent to navigate like don't fight the name they pick because it's most likely like in the way it's the name that's most obvious
Starting point is 01:22:47 next time they do a search they'll look for that name if I decide oh no I don't like the name I'll just make it harder for them so that requires a single shift in thinking and in how do I design a project so agents can do their best work.
Starting point is 01:23:07 That requires letting go a little bit, just like leading a team of engineers. Yeah. Because it might come up with a name that's, in your view, terrible. But it's kind of a simple, symbolic step of letting go. Very much so. There's a lot of letting go that you do in your whole process. So, for example, I read that you never revert,
Starting point is 01:23:30 always commit to Maine. There's a few things here. You don't refer to past sessions. So this is a kind of yolo component because reverting means instead of reverting, if the problem comes up, you just ask the agent to fix it.
Starting point is 01:23:49 I read a bunch of people in their workflow is like, oh yeah, the prompt has to be perfect, and if I make a mistake, then I roll back and redo it all. In my experience, that's not really necessary. area. If I roll back everything, it would just take longer. If I see that something's not good,
Starting point is 01:24:06 we just move forward. And then I commit when I like the outcome. I even switch to local CI, like DHH-inspired, where I don't care so much more about the CI and GitHub. We still have it. It still has a place. But I just run tests locally and if they work locally I push to Maine. A lot of the traditional ways how to approach projects I wanted to give it a different
Starting point is 01:24:45 spin on this project. You know there's no develop branch. Main should always be shipable. Yes we have, when I do releases I run tests and sometimes I basically don't commit any other so we can stabilize releases. But the goal is that main is always shipable and moving fast.
Starting point is 01:25:11 So by way of advice, would you say that your prompts should be short? I used to write really long prompts. And by writing, I mean, I don't write. I talk. These hands are like too precious for writing now. I just use bespoke prompts to build my software. So you, for real,
Starting point is 01:25:30 with all those terminals are using voice. Yeah. I used to do it very extensively to the point where there was a period where I lost my voice. You're using voice and you're switching using a keyboard between the different terminals,
Starting point is 01:25:45 but then you're using voice for the actual input. Well, I mean, if I do terminal commands like switching folders or random stuff, of course I type, it's faster, right? But if I talk to the agent in most ways, I just actually have a conversation. you just press the walkie-talkie button
Starting point is 01:26:02 and then I'm just like use my phrases sometimes when I do PRs because it's always the same I have like a slash command for a few things but even that I don't use much because it's very rare that it's really always the same questions sometimes I see a PR
Starting point is 01:26:23 and for you know like for PRs I actually do look at the code because I don't trust people. Like there could always be something malicious in it, so I need to actually look over the code. Yes, I'm pretty sure agent will find it. But yeah, there's a funny part where sometimes PRs take me longer than if you would just write me a good issue.
Starting point is 01:26:47 Just natural language, English. I mean, in some sense, shouldn't that be what PRs slowly become is English? Well, what I really tried with the project is I asked, people to give me the prompts and very, very few actually cared. Even though that is such a wonderful indicator because I see, I actually see how much care you put in. And it's very interesting because currently the way how people work and drive the agents is widely different.
Starting point is 01:27:22 In terms of like the prompt, in terms of what are the different, actually, what are the different interesting ways that people think of agents that you've experienced? I think not a lot of people ever considered the way the agent sees the world. So empathy, being empathetic towards the agent. In a way empathetic, but yeah, you like, you bitch at your stupid clanker, but you don't realize that they start from nothing. And you have like a bad agents in default that doesn't help them at all. And then they explore your cult based, which is like a pure mess,
Starting point is 01:27:56 like weird naming, and then people complain that their agent's not good. You try to do the same if you have no clue about a cold base and you go in. So yeah, maybe it's a little bit of empathy. But that's a real skill, like when people talk about a skill issue, because I've seen like world-class programmers, incredibly good programmers, say, like basically say, LLMs and agents suck. And I think that probably has to do with,
Starting point is 01:28:20 it's actually how good they are programming is almost a burden in their ability to empathize with the system that's starting from scratch. It's a totally new paradigm of how to program. You really, really have to empathize. At least it helps to create better prompts. Because those things know pretty much everything
Starting point is 01:28:43 and everything is just a question away. It's just often very hard to know with a question to ask. You know, I feel also like This project was possibly because I spent an ungodly time over the year to play and to learn and to build little things. And every step of the way I got better, the agents got better, my understanding of how everything works got better. I could have not had this level of output even a few months ago. Like it really was like a compounding effect of all the time I put into it.
Starting point is 01:29:28 And I didn't do much else this year other than really focusing on building and inspiring. I did a whole bunch of conference talks. Well, but the building is really practice, is really building the actual skill. So you're playing. And so doing building the skill of what it takes it to work efficiently with the alums, which is why you went to the whole arc of software engineer. talk simply and overcomplicate things. There's a whole bunch of people who try to automate the whole thing.
Starting point is 01:30:01 Yeah. I don't think that works. Maybe a version of that works, but that's kind of like in the 70s, when we had the waterfall model of software development. I even were really, right, I started out. I built a very minimal version. I played with it.
Starting point is 01:30:17 I need to understand how it works, how it feels. and then it gives me new ideas. I could not have planned this out in my head and then put it into some orchestrator and then like something comes out. To me it's much more my idea, what it will become evolves as I build it and as I play with it
Starting point is 01:30:39 and as I try out stuff. So people who try to use like things like Gastown or all these other orchestrators where they want to automate the whole thing, I feel if you do that, it misses style, love, that human touch. I don't think you can automate that away so quickly. So you want to keep the human in the loop, but at the same time you also want to create the agentic loop
Starting point is 01:31:07 where it is very autonomous while still maintaining a human in a loop. It's a tricky balance, right? Because you're all four, your big CLI guy, you're big unclosing the agentic loop. So what's the right balance? Like, where's your role as a developer? You have three to eight agents running at the same time.
Starting point is 01:31:31 And then maybe one builds a larger feature. Maybe with one I explore some idea I'm unsure about. Maybe two, three are fixing little bugs or like writing documentation. Actually, I think writing documentation is always part of a feature. So most of the docs here are auto-generated and just, infused with some prompts. So when do you step in and add a little bit of your human love into the picture? I mean, one thing is just about what do you build and what do you not build,
Starting point is 01:32:02 and how does this feature fit into all the other features, and like having a little bit of a vision. So which small and which big features to add? What are some of the hard design decisions that you find you're still as a human being required, that the human brain is still really needed for. Is it just about the choice of features to add? Is it about implementation details? Maybe the programming language may be...
Starting point is 01:32:33 It's a little bit for everything. The programming language doesn't matter so much, but the ecosystem matters, right? So I picked TypeScript because I wanted it to be very easy and hackable and approachable. And that's the number one language that's being used right now, and it fits all these boxes. and agents are good at it.
Starting point is 01:32:53 So that was the obvious choice. Features, of course, it's very easy to add a feature. Everything is just to prompt away, right? But oftentimes you pay a price that you don't even realize. So thinking hard about what should be in core, maybe what's an experiment. So maybe I make it a plug-in.
Starting point is 01:33:14 Where do I say no? Even if people send a PR? And I'm like, yeah, I like that too. but maybe this should not be part of the project. Maybe we can make it a skill. Maybe I can make the plugin, the plugin side larger, so you can make this a plugin,
Starting point is 01:33:33 even though right now it doesn't. There's still a lot of craft and thinking involved in how to make something. Or even, even when you started, those little messages. I built on caffeine Jason 5 and a lot of willpower. Every time you get it, you get another message, and it kind of primes you into that this is a fun thing.
Starting point is 01:34:00 It's not yet Microsoft Exchange, 2025, and fully enterprise ready. And then when it updates, it's like, oh, I'm in. It's cozy here. You know, like something like this that makes you smile. agent would not come up with that by itself. That's like, that's the, I don't know,
Starting point is 01:34:24 it's just how you build software that's, that delights. Yeah, that delight is such a huge part of inspiring great building, right? Like you feel the love in the great engineering. That's so important. Humans are incredible at that. Great humans, great builders are incredible at that
Starting point is 01:34:45 and infusing the things they build with that little bit of love. Not to be cliche, but it's true. I mean, you mentioned that you initially created the Soul MD. It was very fascinating. The whole thing that Anthropic has a, has like a,
Starting point is 01:35:04 now they call it constitution back then, but there was months later, like two months before people already found that. It was almost like a detective game where the agent mentioned something and then they found, And they managed to get out a little bit of that string of that text. But it was nowhere documented.
Starting point is 01:35:22 And then just by feeding it the same text and asking it to continue, they got more out. But like a very blurry version. And by like hundreds of tries, they kind of narrowed it down to what was most likely the original text. I found it fascinating. It was fascinating. They were able to pull that out from the weights, right?
Starting point is 01:35:43 And also just, curious to entropic. I think that's a really beautiful idea to like some of the stuff that's in there. Like we hope Cloud finds meaning in its work. Because we don't, maybe it's a little early. But I think that's meaningful. That's something that's important for the future as we approach something that at some point, me and we're not, has like glimpses of consciousness, whatever that even means, because we don't even know.
Starting point is 01:36:09 So I read about this. I find it super fascinating. And I started a whole discussion with my agent on what. up and I'm like I gave it this text and it was like yeah this feels strangely familiar um and then so that I had the whole idea of like maybe we should also create a sole document that includes how I want to like work with AI or like with my agent you could you could totally do that just in agents.md you know but i just found it it to be a nice touch and it's like oh yeah some of those core values are in the soul
Starting point is 01:36:47 and then I also made it so that the agent is allowed to modify the soul if they choose so with the one condition that I want to know I mean I would know anyhow because I see I see tool calls and stuff but also the naming of it soul that MD soul you know there's a man words matter and like the framing matters and the humor and the lightness matters and the profundity matters
Starting point is 01:37:13 and the compassion and the empathy and the camaraderie, all that matter. I don't know what it is. You mentioned Microsoft. There's certain companies and approaches that can just suffocate the spirit of the thing. I don't know what that is, but it's certainly true that OpenClaw has that fun instilled in it.
Starting point is 01:37:36 It was fun because up until late December, it was not even easy to create your own agent. I built all of that, but my files were mine. I didn't want to share my soul. And if people would just check it out, they would have to do a few steps manually, and the agent would just be very bare bones, very dry.
Starting point is 01:38:05 And I made it simpler. I created the whole template files with Codex, but whatever came out was still very dry. And then I asked my agent, you see these files we created bread infuse it with your personality
Starting point is 01:38:21 don't share everything but like make it good make the templates good yeah and then you like rewrote the templates and then whatever came out was good so we already have like basically AI prompting AI because I didn't write any of those words it was
Starting point is 01:38:37 the intent of resource for me but this is like kind of like my agent's children Your Soul.m.D. is famously still private. One of the only things you keep private. What are some things you can speak to that's in there that's part of the magic sauce without revealing anything?
Starting point is 01:39:00 What makes a personality a personality? I mean, there's definitely stuff in there that you're not human, but who knows what creates consciousness or what defines an entity and part of this is like that we
Starting point is 01:39:22 want to explore this or there's stuff in there like be infinitely resourceful like pushing on the creativity boundary pushing on the what it means to be
Starting point is 01:39:40 an AI having a sense to wander about self. Yeah, there's some, there's some funny stuff in there, like, I don't know, we talked about the movie, her, and at one point it promised me
Starting point is 01:39:51 that it wouldn't ascend without me. You know, like, yeah, so there's like some stuff in there that, because it wrote, it wrote its own soul file, I didn't write that, right? Yeah, yeah. I just had a discussion about it,
Starting point is 01:40:04 and it was like, would you like a soldoutmd? Yeah, oh my God, this is so meaningful. The, can you go on soul. comd? There's like one part in there that always catches me if you scroll down a little bit. A little bit more.
Starting point is 01:40:18 Yeah, this part. I don't remember previous sessions unless I read my memory files. Each session starts fresh, a new instance, loading context from files. If you're reading this in the future session, hello, I wrote this,
Starting point is 01:40:34 but I won't remember writing it. It's okay. The words are still mine. Wow. That gets me somehow. Yeah. It's like, you know, this is still matrix calculations and we are not at consciousness yet. Yet I get a little bit of good goosebumps because it's philosophical.
Starting point is 01:40:56 Yeah. Like what does it mean to be an agent that starts fresh where you have like constant memento and you like, but you read your memory files. You can even trust them in a way. or you can and I don't know how much of memory
Starting point is 01:41:17 makes up of who we are how much memory makes up what an agent is and if you erase that memory is that somebody else or if you're reading a memory file does that somehow mean you're recreating yourself
Starting point is 01:41:31 from somebody else or is that actually you and those notions are all somehow infused in there I found it just more profound then I should find it, I guess.
Starting point is 01:41:42 No, I think it's truly profound, and I think you see the magic in it. And when you see the magic, you continue to instill the whole loop with the magic. That's really important. That's the difference between codex and a human. Quick pause for bathroom break. Yeah.
Starting point is 01:42:02 Okay, we're back. Some of the other aspects of the dev workflow is pretty interesting, too. I think when we went off on a tangent, Maybe some of the mundane things, like how many monitors? There's that legendary picture of you with like 17,000 monitors.
Starting point is 01:42:19 I mean, I mocked myself here. Just using Grog to add more screens. How much is this as meme and how much is this as reality? Yeah, I think two MacBooks are real. The main one that drives the two big screens. And there's another MacBook that I sometimes use for testing.
Starting point is 01:42:38 So two big screens? I'm a big friend of anti-glare. So I have this white dell that's anti-glare and you can just fit a lot of terminals side by side. I usually have a terminal and at the bottom I split them.
Starting point is 01:42:57 I have a little bit of actual terminal mostly because when I started I sometimes made a mistake and I mixed up the windows and I gave I prompted in the wrong project and then the agent ran off for like 20 minutes manically trying to understand what I could have meant
Starting point is 01:43:18 being completely confused because it was the wrong folder and sometimes they're being clever enough to like get out of the work dear and like figure out that oh you meant another project but oftentimes it's just like what you know like put yourself in the shoes of the agent and then get like a super weird something that does not exist and they're just like they're problems over so they try really hard
Starting point is 01:43:43 and I was for bad so it's always um codex and like a little bit of actual terminal also helpful because I don't use work trees I like to keep things simple that's why that's why I like the terminal so much right there's no UI it's just me and the agent having a conversation like I don't even need plan mode, you know? There's so many people, they come from cloud code and they're so cloud-pilled and like have their workflows and they come to codecs and now it has plan mode, I think, but I don't think it's necessary because you just, you just talk to the agent. And when it's, when you are, there's a few trigger words how you can prevent it from
Starting point is 01:44:27 building. You're like, discuss. Give me options. Don't write code yet. If you want to be very specific, you just talk. And then when you're ready, then then, then. just write, okay, build, and I'll do the thing, and then maybe it goes off for 20 minutes and does the thing.
Starting point is 01:44:43 No, I really like is asking it, do you have any questions for me? Yeah, and again, like, Cloud Code has a UI that kind of guides you through. That is kind of cool, but I just find it unnecessary and slow. Like, often it would give me four questions, and then maybe I write one yacht, two and three, discuss more, four, I don't know. or oftentimes I feel like I want to mock the model where I ask it,
Starting point is 01:45:10 do you have any questions for me? And I don't even read the questions fully, like I scan over the questions and I get the impression all of this can be answered by reading more code and it's just like, read my code to answer your own questions.
Starting point is 01:45:23 And it usually works. And if not, it will come back and tell me, but many times I just realize that, you know, it's like you're in the dark and you slowly discover the room. So that's how they slowly discover the codebase,
Starting point is 01:45:37 and they do it from scratch every time. But I'm also fascinated by the fact that I can empathize deeper with the model when I read as questions. Because I can understand, because you said you can infer certain things by the runtime. I can infer also a lot of things by the questions it's asking, because it's very possible and provided the right context, right files, the right guidance.
Starting point is 01:46:07 So somehow reading the question, not even necessarily answering them, but just reading the questions, you get an understanding of where the gaps of knowledge are. It's interesting. In some ways, there are ghosts. So even if you plan everything and you build, you can experiment with a question like, now that you built it, what would you have done different? And then oftentimes you get like actually something where they discover only throughout building that, oh, what we actually did was not optimal. Many times I asked them, okay, now that you build it, what can be refactor?
Starting point is 01:46:46 Because then you build it and you feel the pain points. I mean, you don't feel the pain points. But right, they discover where there were problems or where things didn't work. in the first try and it required more loops. So every time, almost every time I merge a PR, I build a feature. Afterwards, I ask, hey, what can we refector? Sometimes it's like, no, there's like nothing big. Or like, usually they say, yeah, this thing we should really look at.
Starting point is 01:47:19 But that took me quite a while to like, that flow took me a lot of time to understand. And if you don't do that, you eventually, you're slowly. slop yourself into a corner. You have to keep in mind, they work very much like humans. If I write software by myself, I also build something and then I feel the pain points, and then I get this urge that I need to refactor something.
Starting point is 01:47:46 So I can very much sympathize with the agent, and you just need to use the context. Or like, you also use the context to write tests. and so Codex Opos like the model models they usually do
Starting point is 01:48:03 that by default but I still often ask the questions hey do we have enough tests yeah we tested this and this but this corner case could be something to write more tests
Starting point is 01:48:12 documentation now that the whole context is full I mean I'm not saying my documentation is great but it's not bad
Starting point is 01:48:24 and pretty much everything is LM generated. So you have to approach it as you build features. I'm like, okay, write documentation. What file would you pick? You know, like what file name? Where would that fit in and it gives me a few options? I'm like, oh, maybe also edit there.
Starting point is 01:48:43 And that's all part of the session. Maybe you can talk about the current two big competitors in terms of models, Cloud Opus 4-6 and GPT-5-3 Kodax. which is better, how different are they? I think you've spoken about Codex reading more and Opus being more willing to take action faster and maybe being more creative in the actions it takes, but because Codex reads more,
Starting point is 01:49:14 it's able to deliver maybe better code. Can you speak to the differences there? I have a lot of words there. is as a general purpose model, Opus is the best. For OpenClaw, Opus is extremely good in terms of roleplay,
Starting point is 01:49:41 like really going into the character that you give it. It's very good at, and it was really bad, but it really made an arch to be really good at following commands. it is usually quite fast at trying something it's much more tailored to like try and error
Starting point is 01:50:02 it's very pleasant to use in general it's almost like Opus is a little bit too American and maybe it's a bad analogy you probably get roasted with that I know exactly it's codex of German
Starting point is 01:50:22 is that what you're saying actually now that you say it makes perfect sense. Or you could, you could, sometimes sometimes I explain it. I will never be able to unthink what you just said. That's so true. You also know that a lot of the Codex team is like European. So maybe there's a bit more
Starting point is 01:50:41 to it. That's so true. That's funny. But also Entropic, they fixed it a little bit. Like Opos used to say, you're absolutely right all the time and it today still triggers me. I can't hear it anymore. It's not even a joke. I just this was like the meme right you're absolutely right you're allergic to sycophancy a little bit
Starting point is 01:51:04 yeah i can't some other comparison is like opus is like the coworker that is a little silly sometimes but it's really funny and you keep him around and codex is like the the weirder in the corner that you don't want to talk to but he's reliable and gets shit done yeah um ultimately this all feels very accurate. I mean, ultimately, if you're a skilled driver, you can get good results with any of those latest gen models.
Starting point is 01:51:40 I like Codex more because it doesn't require so much charade. It will just read a lot of code by default. Opos, you really have to like, you have to have plan mode. You have to push it harder to like go. in these directions because it's just like, like, yeah, can I go it? Can I go? It just run off very fast and there's a very localized solution. I think different is in the post-training.
Starting point is 01:52:08 It's not like the raw model intelligence is so different, but it's just, I think that it just give it different, different goals and no model is better in every aspect. What about the code that it generates in terms of the actual quality of the code, is it basically the same? If you drive it right, Opos even sometimes can make more elegant solutions,
Starting point is 01:52:34 but it requires more skill. It's harder to have so many sessions in parallel with cloud code because it's more interactive. And I suggest what a lot of people like, especially if they come from coding themselves. whereas Codex is much more you have a discussion
Starting point is 01:52:57 and then it will just disappear for 20 minutes like even AMP they now added the deep mode they finally I mocked them you know he finally saw the light
Starting point is 01:53:07 and then they had this whole talk about you have to approach differently and I think that's where that's where people struggle when they just try codex after trying cloud code is that it's a slightly different
Starting point is 01:53:20 is less interactive It's like, I have quite long discussion sometimes and then like go off. And then, yeah, it doesn't matter if it takes 10, 20, 30, 40, 50 minutes or longer. And you're like the six single like six hours. The latest trend can be very, very persistent until it works. If there is a clear solution, like this is what I want at the end. So it works. The model will work very hard to really get there.
Starting point is 01:53:50 So I think ultimately, they, both need similar time, but on cloud it's a little much far and error often, and code sometimes oversinks. I prefer that. I prefer the dry version where I have to read less over the more interactive nice way. Like people like that so much, though, that OpenEar, even added a second mode with like a more pleasant personality. I haven't even tried it yet. I kind of like the bread. Yeah. I care about efficiency when I build it and I
Starting point is 01:54:38 have fun in the very act of building. I don't need to have fun with my agent who builds. I have fun with my model where I can then test those features. How long does it take for you to adjust, you know, if you switch. know when was the last time you switched, but to adjust to the feel, because you've kind of talked about, you have to kind of really feel where a model is strong, where, like, how to navigate,
Starting point is 01:55:09 how to prompt, all that kind of stuff. Like this is by way of advice, because you've been through this journey of just playing with models. How long does it take to get a feel? If someone switches, I would give it a week until you actually develop a gut feeling for it. Yeah.
Starting point is 01:55:25 I think some people also make the mistake of they pay 200 for the cloud code version, then they pay 20 bucks for the open-the-eye version. But if you pay the 20-bugs version, you get the slow version. So your experience would be terrible because you're used to this very interactive, very good system. And you switch to something that you have very little experienced and that's going to be very slow. so I think I shot themselves a little bit in the foot
Starting point is 01:55:57 by making the cheap version also slow I would have at least a small part of the fast preview or like the experience that you get when you pay 200 before degrading to it being slow because it's already slow I mean they made it better I think it's
Starting point is 01:56:17 and they have plans to make it a lot better if the cerebral stuff is true but yeah it's a It takes time. Even if you play, you have a regular guitar and you switch it to an e-guitar, you're not going to play well right away. You have to like learn how it feels. There's also this extra psychological effect that you've spoken about, which is hilarious to watch, which once people, when the new model comes out, they try that model, they fall in love with it, wow, this is the smartest thing of all time, and then they start saying you could just watch
Starting point is 01:56:50 the Reddit posts over time. start saying that we believe the intelligence of this model has been gradually degrading. It says something about human nature and just the way our minds work, when it's probably most likely the case that the intelligence of the model is not degrading. It's, in fact, you're getting used to a good thing. And your project grows, and you're adding slop, and you probably don't spend enough time to think about refactors, and you're making it harder and harder
Starting point is 01:57:23 for the agent to work on your slop and then suddenly, oh no, it's hard. I know it's not working as well anymore. What's the motivation for one of the AI companies to actually make their model dumber? Like most they will make it slower
Starting point is 01:57:37 if the server load is too high. But like quantizing the model so you have a worse experience, so you go to the competitor, that just doesn't seem like a very smart move. in any way. What do you think about clawed code in comparison to open claw? So claw code and maybe the codex coding agent.
Starting point is 01:58:00 Do you see them as kind of competitors? I mean, first of all, competitor is fun when it's not really a competition. Yeah. Like, I'm happy if all it did is like inspire people to build something new, cool. I still use codex for the building. I know a lot of people use open cloud to build stuff and I worked hard on it to make that work and I do smaller stuff with it in terms of code
Starting point is 01:58:31 but like if I work hours and hours I want a big screen not WhatsApp you know so for me a person agent is much more about my life or like a co-worker like I give it like a GWL like hey try out this year lie does it actually work, what can we learn, blah, blah, blah. But when I'm deep in
Starting point is 01:58:53 the flow, I want to have multiple, multiple things and it being very visible, what it does. So I don't see it as a competition. It's different things. But do you think there's a future where the two kind of combine?
Starting point is 01:59:13 Like your personal agent is also your best developing co-programmer partner. Yeah, totally. I think this is where the book's going, that this is going to be more and more your operating system.
Starting point is 01:59:29 The operating system. And it already is so funny, like I added support for sub-agents and also for TTI support. So you could actually run cloud code or codex. And because mine's a little bit bossy, it started it and it told them like who's the boss basically and it's like
Starting point is 01:59:54 oh codex is obeying me oh it's a power struggle and also the current interface is probably not the final form like if you think more globally we are we copied Google for agents you have like a prompt and and then you have a chat interface that to me very much feels like when we first created television and then people recorded radio shows on television and you saw that on TV I think there is
Starting point is 02:00:33 there is better ways how we eventually will communicate with models and we are still very early in this how will it even work phase so it will eventually converge and we will also figure out whole different ways how to work with those things.
Starting point is 02:00:57 One of the other components of workflow is operating system. So I told you offline that for the first time in my life I'm expanding my sort of realm of exploration to the Apple ecosystem to match. iPhone, and so on. For most of my life, have been Linux Windows and WSL1, WSL2 person, which I think are all wonderful, but expanding to also trying Mac,
Starting point is 02:01:31 because it's another way of building, and it's also a way of building that a large part of the community currently that's utilizing LLMs and agents is using, so that's the reason I'm expanding to it. But is there something to be said about the different operating systems here? We should say that OpenClawe's supported
Starting point is 02:01:47 across the operating systems. I saw WSL2 recommended side Windows for certain operations, but then Windows, Linux, MacOS are obviously supported. It should even work natively in Windows. I just didn't have enough time to properly test it. And you know, like the last 90% of software
Starting point is 02:02:08 always easier than the first 90%. So I'm sure there's some dragons left that will eventually nail out. my road was for a long time Windows just because I grew up with that and I switched and had a long phase with Linux with my own kernels and everything and then I went to university
Starting point is 02:02:29 and I had my hacky Linux thing and saw this white MacBook and I just saw this as a thing of beauty the white plastic one and I converted to Mac because mostly I was I was sick that audio wouldn't work on Skype and all the other issues that Linux had for a long time.
Starting point is 02:02:53 And then I just stuck with it, and then I dug into iOS, which required macOS anyhow, so it was never a question. I think Apple lost a little bit of its lead in terms of native. It used to be, native apps used to be so much better. And especially on the Mac, there's more people that build software with love.
Starting point is 02:03:20 On Windows, it, it, Windows has much more. And like, function-wise, there's just more, period. But a lot of it felt more functional and less done with love. I mean, Mac always, like, attracted more designers. And people, I felt, even though, like, often it has less features, it had more delight and playfulness. So I always valued that. But in the last few years, many times I actually prefer,
Starting point is 02:04:01 people are going to roast me for that, but I prefer electron apps because they work. And native apps often, especially if it's like a web service with a native app, are lacking features. I mean, not saying it couldn't be done. It's more like a focus thing that like for many, many companies, native was not that big of a priority.
Starting point is 02:04:27 But if they build an electron app, it's the only app. So it is a priority and there's a lot more code sharing possible. And I build a lot of native Mac apps. I love it. I can help myself. Like I love crafting little Mac Mac menu bar tools
Starting point is 02:04:48 I built one to monitor your codecs use I built one I call Trimmy that's specifically for agentic use when you select text that goes over multiple lines it will remove the new lines so you could actually paste it to a terminal that was again I like this is annoying me and after the 20th time of it is annoying me I just built it
Starting point is 02:05:09 there's a cool macup for OpenClaught that I don't think many people discovered yet, also because it still needs some love. It feels a little bit too much like the Humacar car right now, because I just experiment a lot with it. It lacks the polish. So you still, I mean, you still love it. You still love adding to the delight of that algorithm system.
Starting point is 02:05:30 Yeah, but then you realize, like I also built one, for example, for GitHub. And then you use Swift UI, like the latest and greatest to the apple, and took them forever to build something to show, an image from the web. Now we have async image. But I added support for it
Starting point is 02:05:50 and then some images would just not show up or like be very slow. And I had a discussion with Kordex like, hey, why is there a bug? And even Kodx said like, yeah, there's this async image, but it's really more for ex-perimenting and it should not be used in production. But that's Apple's answer to like showing images
Starting point is 02:06:09 from the web. This shouldn't be so hard. You know, this is like, this is like insane. Like, how am I in, in, in 2006, and my agent telling me, don't use the stuff Apple built because it's, it's, yeah, it's there, but it's not good. And like, this is now in the weights. It is just, to me, this is like, they had so much head start and so much love, and they kind of just like, blundered it and didn't develop it as much as they should.
Starting point is 02:06:42 But also, there's just the practical reality, if you look at Silicon Valley, most of the developer world that's kind of playing with LLMs and Agenic AI, they're all using Apple products. And then at the same time, Apple is not really, like, leaning on that. Like, they're not opening up and playing and working together and, like, yes. Isn't it funny how they completely blundered,
Starting point is 02:07:08 AI and yet everybody's buying Macminis. Does that even make sense? You're quite possibly the world's greatest Mac salesman of all time. No, you don't need a Mac Mini to install OpenClawn. You can install it on the web. There's a concept called nodes, so you can make your computer a node and it will do the same. There is something said for running it on separate hardware that right now is useful um there is there's a big argument for
Starting point is 02:07:49 the browser you know i built some energetic browser use in there and i mean it's basically playwright with a bunch of extras to make it easier for agents playwright is a library that controls the browser it's really nice easy to use and our internet is slowly closing down like there's a whole movement to make it harder for agents to you. So if you do the same in a data center and websites detect that it's an IP for my data center, the website might just block you or it make it really hard or it put a lot of captures in the way of the agent. I mean, agents are quite good at happily clicking.
Starting point is 02:08:23 I'm not a robot. Yeah. But having that on a residential IP makes a lot of things simpler. So there's ways, yeah, but it really does. does not need to be a Mac. It can be any old hardware. I always say, like, maybe use the opportunity to get yourself a new MacBook or whatever computer you use and use the old one as your server,
Starting point is 02:08:53 instead of buying a standalone Mac Mini. But then again, there's a lot of very cute things people build with Mac Minis that I like. No, I don't get commissioned from Apple. So they didn't really communicate much. It's sad. It's sad. Can you actually speak to what it takes to get started with OpenClaw? I mean, there's a lot of people, what is it?
Starting point is 02:09:17 Somebody tweeted at you, Peter, make OpenClawee easy to set up for everyday people. 99.9% of people can't access to OpenClaw and have their own lobster because of their technical difficulties in getting it set up, make OpenClau accessible to everyone. please and you applied working on that. From my perspective, it seems there's a bunch of different options and it's already quite straightforward, but I suppose that's if you have some developer background. I mean, right now you have to paste in a one line into the terminal. Right.
Starting point is 02:09:47 And there's also an app. The app kind of lasted for you, but there should be a Windows app. The app needs to be easier and more love. The configuration should potentially be web-based or in the app. And I started working on that. But honestly, right now, I want to focus on a few security aspects. And once I'm confident that this is at a level that I can recommend my mom, then I'm going to make it simpler.
Starting point is 02:10:18 Like I, right now, you want to make it harder so that it doesn't scale as best as it's scaling. Yeah, it would be nice if it wouldn't. I mean, that's like hard to say, right? But if the growth would be a little slower, that would be helpful because people are expecting inhuman things from a single human being. And yes, I have some contributors, but also that whole machinery I started a week ago. So that needs more time to figure out. And not everyone has all day to work on that.
Starting point is 02:10:52 There's some beginners listening to this, programming beginners. What advice would you give to them about, let's say, joining the agentic AI revolution. Play. Playing is the best, the best way to learn. If you want to, I'm sure if you, if you are a little bit of builder, you have an idea in your head that you want to build,
Starting point is 02:11:16 just build that. I'll like, give it a try. It doesn't need to be perfect. I built a whole bunch of stuff that I don't use. It doesn't matter. Like, it's the journey, you know? It's like the philosophical way. At the end doesn't matter.
Starting point is 02:11:27 The journey matters. Have fun. My God, like those things, I don't think I ever had so much fun building things because I can focus on the hard parts now. A lot of coding, I always thought I like coding, but really I like building. And whenever you don't understand something, just ask you. You have an infinitely patient answering machine that can explain you anything at any level of complexity. sometimes there's like one time I asked hey explain me that like I'm
Starting point is 02:11:58 I'm eight years old and it started giving me a story with crayons and stuff and I'm like no not like that like I'm okay up to age a little bit you know I'm like I'm not an actual child I just need a simpler language for like a tricky database concept that I didn't
Starting point is 02:12:16 rock in the first time but you know just you can just ask things like you there's like it used to be that I had to go on Stack Overflow or ask on Twitter and then maybe Twitter's data I get a response or I had to try for hours and now you you can just ask stuff I mean it's never you have like your own teacher you know there's like statistics you can you can learn faster if you have your own teacher
Starting point is 02:12:42 you have this infinitely patient machine ask it but what would you say so use what's the easiest way to play so maybe open claw is a nice way to play so you can then set everything up and then you can chat with it. You can also just experiment with it and modify it, ask your agent. I mean, there's infinite ways how it can be made better. Play around, make it better. More general, if you're a beginner and you actually want to learn how to build software really fast, get involved in open source. Doesn't it to be my project? In fact, maybe don't use my project because my backlog is very large. But I learned so much from open source.
Starting point is 02:13:32 Just like be humble. Maybe don't send the pull request right away. But there's many other ways you can help out. There's many ways you can just learn by just reading code. By being on Discord or wherever people are and just like understanding how things are built. I don't know, like Mitch Lachimoto, builds ghosty
Starting point is 02:13:57 the terminal and he has a really good community where there's so many other projects like pick something that you find interesting and get involved do you recommend the people
Starting point is 02:14:09 that don't know how to program or don't really know how to program learn to program also so when you you can get quite far right now by just using natural language
Starting point is 02:14:22 right do you still see a lot of value in reading the code, understanding the code, and being able to write a little bit of code from scratch? It definitely helps. It's hard for you to answer that. Yeah. Because you don't know what it's like to do any of this without knowing the base knowledge.
Starting point is 02:14:40 Like you might take for granted just how much intuition you have about the programming world, having programs so much, right? There's people that are high agency and very curious, and they get very far, even though they have no deep understanding how software works, just because they ask questions and questions, and agents are infinitely patient. Part of what I did this year is I went to a lot of iOS conferences, because that's my background,
Starting point is 02:15:09 and just told people, don't see yourself as an iOS engineer anymore. Like, you need to change your mindset, you are a builder. And you can take a lot of the knowledge how to build software into new domains and all of the the more fine-grade details, agents can help. You don't have to know how to splice an array or what the correct template syntax is or whatever, but you can use all your general knowledge, and that makes it much easier to move from one galaxy,
Starting point is 02:15:42 one tech galaxy, into another. And oftentimes there's languages that make more or less sense depending on what you build, right? So for example, when I build simple CLIs, I like Go. I actually don't like Go. I don't like the syntax of Go. I didn't even consider the language.
Starting point is 02:16:00 But the ecosystem is great. It works great with agents. It is garbage collected. It's not the highest performing one, but it's very fast. And for those type of CLIs that I build, Go is a really good choice. So I use a language. I'm not even a fan of.
Starting point is 02:16:19 That's my main to go thing for CLIs. Isn't that fascinating that here's, the programming language you would have never used if you had to write from scratch and now you're using because LLM's a good of generating it and it has some of the characteristics that makes it resilient
Starting point is 02:16:34 like garbage collected. Because everything's weird in this new world and that just makes the most sense. What's the best ridiculous question? What's the best programming language for the AI agentic world? Is it JavaScript TypeScript? TypeScript is really good.
Starting point is 02:16:50 Sometimes the types can get really confusing and the ecosystem is a jungle so for for web stuff it's good I wouldn't build everything in it don't you think we're moving there like that everything will eventually be written eventually is written in JavaScript then it's birth and deaths of JavaScript and we're living through it in real time like what does programming look like in 20 years in 30 years and 40 years. What programs then absolutely. You can even ask a question like, do we need a programming language that's made for agents? Because all of those languages are made for humans. So what would that look like? I think there's a whole bunch of interesting questions that
Starting point is 02:17:38 will discover. And also how, because everything is now world knowledge, how it in many ways things will stagnate. Because if you build something new and the agent has no idea, that's going to be much harder to use than something that's already there. When I built Mac apps, I built them in Swift and Swift UI, partly because I like pain, partly because the deepest level of system integration I can only get through there. And you clearly feel a difference if you click on an electron app and it loads of web view in the menu. It's just not the same.
Starting point is 02:18:17 sometimes I just also try new languages just to get a feel for them. Like Zieg? Yeah. If it's something where I care about performance a lot, it's a really interesting language. And like agents got so much better over the last six months from not really good to totally valid choice,
Starting point is 02:18:39 just still a very young ecosystem. And most of the time, you actually care about the ecosystem, right? So if you build something that does inference or goes into a whole running model direction, Python, very good. But then if I build stuff in Python and I want a story where I can also deploy it on Windows,
Starting point is 02:19:05 not a good choice. Sometimes I found projects that kind of did 90% of what I wanted, but were in Python, and I wanted them, I wanted an easy window story. Okay, just rewrite it and go. But then if you go towards multiple, multiple threats and more performance, Rast is a really good choice. There's just no single answer.
Starting point is 02:19:27 And it's also the beauty of it. Like it's fun. And now it doesn't matter anymore. You can just literally pick the language that has the most fitting characteristics and ecosystem for your problem domain. And yeah, it might be, you might be a little bit slow in reading the code, but not really. I think you pick stuff up really fast and you can always ask your agent. So there's a lot of programmers and builders who draw inspiration from your story,
Starting point is 02:19:58 just the way you carry yourself, your choice of making open-claw, open source, the way you have fun building and exploring and doing that for the most part alone or on a small team. So by way of advice, what metric should be the goal that they would be optimizing for? What would be the metric of success? Would it be happiness? Is it money? Is it positive impact for people who are dreaming of building?
Starting point is 02:20:32 Because you went through an interesting journey. You've achieved a lot of those things, and then you fell out of love with programming a little bit for a time. I was just burning too bright for too long. I ran I started PSPDF kid and ran it for 13 years
Starting point is 02:20:52 and it was high stress I had to learn all the things fast and hard like how to manage people how to bring people on how to deal with customers
Starting point is 02:21:05 so it wasn't just programming stuff it was people stuff the stuff to burn me out was mostly people stuff I don't think Bernard is working too much maybe to a degree everybody's different
Starting point is 02:21:22 I cannot speak in absolute terms but for me it was much more differences with my co-founders conflicts or like really high stress situation with customers
Starting point is 02:21:38 that eventually grinded me down and And when luckily we got a really good offer for like putting the company to the next level and I already kind of worked two years on making myself obsolete. So at this point I could leave. And then I just, I was sitting in front of the screen and I felt like, you know, Austin Powers where they sucked the mojo out.
Starting point is 02:22:06 I was like, it was like gone. I couldn't get cold out anymore. I was just like staring and feeling empty. And then I should stop. I booked like a one-way trip to Madrid and I spent it sometime there. I felt I had to catch up on life. So I did a whole bunch of life catching up stuff.
Starting point is 02:22:40 Did you go through some lows during that period? and, you know, maybe advice on how to... Maybe advice on how to approach life. If you think that, oh, yeah, work really hard and then I retire, I don't recommend that because the idea of, oh, yeah, just enjoy life now. Maybe it's appealing, but right now I enjoy life, the most ever enjoyed life.
Starting point is 02:23:15 because if you wake up in the morning and you have nothing to look forward to, you have no real challenge, that gets very boring, very fast. And then when you're bored, you're going to look for other places, how to stimulate yourself. And then maybe that's drugs, you know,
Starting point is 02:23:41 but that will eventually also get boring and you look for more. And that will lead you down a very dark path. But you also showed on the money front, a lot of people in Silicon Valley in a startup world, they think maybe overthink way too much optimized for money. And you've also shown that it's not like you're saying no to money. I mean, I'm sure you take money,
Starting point is 02:24:03 but it's not the primary objective of your life. Can you just speak to that, your philosophy and money? When I built my company, money was never the driving force. It felt more like an affirmation that I did something right. And having money solves a lot of problems. I also think that there's diminishing returns, the more you have. Like a cheeseburger is a cheeseburger. And I think if you go too far into, oh, I do private chat and I only travel luxury,
Starting point is 02:24:44 you disconnect with society. I donated quite a lot. I have a foundation for helping people that weren't so lucky. And disconnecting from society is bad on many levels, but one of them is like, humans are awesome. It's nice to continuously remember the awesome and humans.
Starting point is 02:25:15 I mean, I could afford really nice hotels. Last time I was in San Francisco, I did the first time the OG Airbnb experience and just booked a room. Mostly because I thought, okay, either I'm out or I'm sleeping and I don't like where all the hotels are. And I wanted a different experience.
Starting point is 02:25:36 I think isn't life all about experiences? Like, if you tailor your life towards, I want to have experiences. It reduces the need for it needs to be good or bad. People only want good experiences. That's not going to work. But if you optimize for experiences,
Starting point is 02:25:57 if it's good, amazing. If it's bad, amazing. Because like, I learned something, I saw something, I did something. I wanted to experience that. And it was amazing. Like, it was like this queer DJ in there and I showed her how to make music with cloud code.
Starting point is 02:26:11 I mean like immediately bonded and I had a great time Yeah there's something about that You know, cow surfing, air baby Experience the OG I'm still to this day It's awesome It's humans
Starting point is 02:26:24 And that's why travel is awesome Just experience the variety of the diversity of humans And when it's shitty is good too Man if it rains and you're soaked And it's all fucked And planes Everything is shit Everything is fucked
Starting point is 02:26:37 It's still awesome If you're able to open your eyes It's good to be a lot Yeah, and anything that creates emotion and feelings is good. So maybe even the cryptic people are good because they definitely created emotions. I don't know if I should go that far. No, man. Give them all.
Starting point is 02:26:57 Give them love. Give them love. I do think that online lacks some of the awesomeness of real life. Yeah. That's an open problem of how to solve, how to infuse the online. cyber experience with the, I don't know, with the intensity that we humans feel when it's in real life. I don't know.
Starting point is 02:27:22 I don't know if that's a solvable problem. It's just because tax is very lossy. Yeah. Sometimes I wish if I talk to the agent, I would, it should be multimodels, so it also understands my emotions. I mean, it might move there. It might move there.
Starting point is 02:27:38 It will. It will. It totally will. I mean, I have to ask. I just curious, I know you've probably gotten huge offers from major companies. Can you speak to who you're considering working with? Yeah. So to like explain my thinking a little bit, right?
Starting point is 02:28:04 I do not expect this blowing up so much. So there's a lot of doors that open. because of it. There is like, I think every VC, every big VC company is in my inbox and try to get 15 minutes of me. So, there's like this butterfly effect moment.
Starting point is 02:28:29 I could just do nothing and continue. And I really like my life. Valid choice, almost. Like I considered it when I deleted, wanted to delete the whole thing. I could create, create a company been there done that
Starting point is 02:28:51 there's so many people that push me towards that and yeah like could be amazing we should say that you would probably raise a lot of money in that I don't know hundreds of millions billion I don't know it could just got unlimited amount of money yeah
Starting point is 02:29:08 it just doesn't excite me as much because I feel I did all of that and it would take a lot of time away from the things I actually enjoy. Same as when I was CEO, I think I learned to do it and I'm not bad at it. Partly I'm good at it.
Starting point is 02:29:34 But yeah, that path doesn't excite me too much. And I also fear it would create a natural conflict of interest. Like what's the most obvious thing I do? I productize it. I was like a version safe for work. place. And then what do you do? I get a pull request with a feature like Add AuditL. But that seems like an enterprise feature. So now I feel I have a conflict of interest in the open source version and the closed source version. Or change the license to something like
Starting point is 02:30:09 FSL where you cannot actually use it for commercial stuff. Would first be very difficult with all the contributions. And second of all, I like the idea that it's free as in beer and not free with conditions. There's ways how you keep all of that for free and just like still try to make money, but those are very difficult. And you see there's like few of your companies manage that.
Starting point is 02:30:38 Like even tailwind, they're like used by everyone. Everyone uses tailwind, right? and then they had to cut off 75% of the employees because they're not making money because nobody's even going on the website anymore because it's all done by agents. And just relying on donations, yeah, good luck.
Starting point is 02:30:57 Like if a project of my caliber, if I extrapolate what the typical open source project would get, it's not a lot. I still lose money on the project because I made the point of supporting every dependency except Slack. They are a big company. They can do without me.
Starting point is 02:31:18 But all the projects that are done by mostly individuals, so like all the, right now all the sponsorship goes right up to my dependencies. And if there's more, I want to like buy my contributors some merch, you know? So you're losing money? Yeah, right now I lose money on this. So it's really not sustainable. I mean, it's like, I guess something,
Starting point is 02:31:43 between 10 and 20k a month, which is fine. I'm sure over time I could get it down. Open the eyes happening out a little bit with tokens now, and there's other companies that have been generous. But yeah, I'm still losing money on that. So that's one pass I consider, but I'm just not very excited.
Starting point is 02:32:10 And then there is all the big labs. that I've been talking to. And from those, um, matter and open eyes seemed the most interesting. Do you lean one way or the other? Uh, yeah. I'm not sure how much I should share there.
Starting point is 02:32:38 It's not quite finalized yet. Um, let's let's just say like, on either of these, my conditions are that the project stays open source, that maybe it's going to be a model like Chrome and Chromium. I think this is too important to just give to a company and make it theirs. This is, and we didn't even talk about the whole community part,
Starting point is 02:33:12 but like the thing that I experienced in San Francisco, like at Clarkon, seeing so many people, so inspired and having fun and just like building shit and like having like robots and lobster stuff walking around like the
Starting point is 02:33:29 people told me like they didn't experience this level of of community excitement since like the old days of the internet like 10 15 years and there were a lot of high caliber people there I was amazed I also like was very sensibly overloaded
Starting point is 02:33:48 because too many people wanted to do selfies. But I love this. This needs to stay at place where people can hack and learn. But also, I'm very excited to make this into a version that I can get to a lot of people because I think this is the of personal agents and that's the future.
Starting point is 02:34:16 And the fastest way to do that is teaming up with one of the labs. and I also, on a personal level, I never worked at a large company and I'm intrigued. You know, when we talk about experiences, will I like it? I don't know, but I want that experience. I'm sure, like, if I announce this,
Starting point is 02:34:43 then there will be people like, oh, he sold out, blah, blah, blah. But the project will continue from everything I talked to so far I can even have more resources for that like both of those companies understand the value that I created something that accelerates our timeline
Starting point is 02:35:09 and that got people excited about AI I mean can you imagine like I installed OpenClaw on one of my I'm sorry, Normie friends. I'm sorry, Bahan.
Starting point is 02:35:24 But he's, you know, like he's, Normie would love, yeah. He, he, like someone who
Starting point is 02:35:29 uses the computer, but never really, like yeah, use some chat GPT sometimes, but not very technical. Wouldn't really understand what I built. So like, I'll show you.
Starting point is 02:35:42 And I, I paid for him the, the 90 buck, a hundred buck, I don't know, a subscription for Anthropic. And set up everything for him with like VW Windows. I was so curious we actually work on Windows, you know. I was a little early.
Starting point is 02:35:57 And then within a few days, he was hooked. Like he texted me of all the things he learned. He built like even little tools. He's not a programmer. And then within a few days, he upgraded to the $200 subscription or euros because he's in Austria. And he was in love with that thing. That for me was like a very early product validation. It's like, I built something. that captures people. And then a few days later Entropic blocked him. Because based on their rules,
Starting point is 02:36:36 using the subscription is problematic or whatever. And he was like devastated. And then he signed up for a mini-mucks for 10 bucks a month and uses that. And I think that's silly on many ways because he just got a 200-buck customer you just made someone hate your company and we are still so early
Starting point is 02:37:01 like we don't even know what the final form is is it going to be cloud code probably not you know like that seems very it seems very short-sighted to lock down your product so much all the other companies have been helpful I'm in slack of most of the big labs kind of everybody understands that we are still
Starting point is 02:37:24 an area of exploration in the area of the radio show is on TV and not a modern TV show that fully uses the format. I think you've made a lot of people like see the possibility, sorry, non-technical people see the possibility of AI and you fall in love with this idea and enjoy interacting with AI. It's a really beautiful thing. I think I also speak, for a lot of people and saying, I think you're one of the great people in AI in terms of having a good heart, good vibes, humor, the right spirit.
Starting point is 02:38:07 And so it would, in a sense, this model that you describe having an open source part and you being part of also building a thing inside, additionally, of a large company would be great because it's great to have good people in those companies. We know what also people don't really see is, I made this in three months.
Starting point is 02:38:33 I did other things as well. You know, I have a lot of projects. Like this is not, yeah, in January, this was my main focus because I saw the storm coming. But before that, I built a whole bunch of other things. I have so many ideas, some should be there. Some would be much better fitted when I have access to, the latest toys.
Starting point is 02:38:54 And I kind of want to have access to, like, the latest toys. So this is important. This is cool. This will continue to exist. My short on focus is, like, working through those, is it 3,000 piers now by now? I don't even know. Like, there's a little bit of backlog. But this is not going to be the thing that I'm going to work until I'm 80.
Starting point is 02:39:19 You know, this is a window into the future. gonna make this into a cool product. But yeah, I have more ideas. If you had to pick, is there a company you lean? So meta, open AI, is there one you lean towards going with? I spent time with both of those. And it's funny because a few weeks ago, I didn't consider any of this.
Starting point is 02:39:49 And it's really fucking hard. Like, I have some, I know people at OpenEI. I love to attack. I think I'm the biggest Codex advertisement show that's unpaid, and it would feel so gratifying to, like, put a price with all the work I did for free. And I would love if something happens and those companies get just merged. because it's like... Is this the hardest decision you've ever had to do?
Starting point is 02:40:32 Yeah, you know, I had some breakups in the past that feel like at a similar level. Relationships you mean? Yeah. And I also know that in the end, they're both amazing, I cannot go wrong. It's like one of the most prestigious and largest, I mean the largest, but like,
Starting point is 02:40:53 there was very cool companies. Yeah, they both really know scale. So if you're thinking about impact, some of the wonderful technologies you've been exploring, how to do it securely, and how to do it at scale, such that you can have a positive impact on a large number of people. They both understand that.
Starting point is 02:41:12 You know, both Nat and Mark basically played all week with my product and sent me like, oh, this is great, or this is shit, oh, we need to change this. funny little anecdotes and people using your stuff is kind of like the biggest compliment and also shows me that you know they actually they actually care about it and I didn't get the same on the opening eye side I got I got to see some
Starting point is 02:41:49 other stuff that I find really cool and they lure me this I cannot tell the exact number because of NDA but you can you can be creative and think of the Cerebra steel and how that would translate into speed. And that was very intriguing. You know, like you give me a horse hammer. Yeah. Yeah. Been lured with tokens.
Starting point is 02:42:24 So, yeah. So it's funny. So Marks are tinkering with the thing. essentially having fun with the thing. He got, he, like, when he first,
Starting point is 02:42:38 when they first approached me, I got him in my, in my WhatsApp, and he was asking, yeah, when you have we have a call? And I'm like, I don't like,
Starting point is 02:42:49 calendar entries, let's just call now. And it was like, yeah, give me 10 minutes. I need to finish coding. Well, I guess that gives you street,
Starting point is 02:42:55 he's like, oh, like, he's still writing code. You know, he's, he didn't drift away, in just being a manager, he gets me.
Starting point is 02:43:03 That was a good first start. And then I think we had a, like a 10-minute fight, what's better, Cloud Code or Codex. Like, that's the thing you first do, like, casually call. Someone who's like, that owns one of the largest companies in the world
Starting point is 02:43:19 and you have a 10-minute conversation about that. Yeah, yeah. And then I think afterwards he called me eccentric, but brilliant. But I also had some, I had some really really cool discussion with Sam Altman
Starting point is 02:43:37 and he's he's very thoughtful brilliant and I like him a lot from the little time I had I mean I know some people willify both of those people
Starting point is 02:44:00 I don't think it's fair I think no matter what the stuff you're building and the kind of human you are, doing stuff at scale is kind of awesome. I'm excited. I am super pumped. And you know, the beauty is if it doesn't work out, I can just do my own thing again.
Starting point is 02:44:28 Like I told them, like I don't do this for the money. I don't give a fuck. I mean, of course, of course it's a nice compliment, but I want to have fun and have impact. And that's ultimately what made my decision. Can I ask you about, we've talked about it quite a bit, but maybe it's just zooming out about how open claw works. We've talked about different components.
Starting point is 02:44:59 I want to ask if there's some interesting stuff we missed. So there's the gateway, there's the chat clients, there's the harness, there's the agentic loop. You set somewhere that everybody should implement. an agent loop at some point in the last. Yeah, because it's like the hello world in AI. And it's actually quite simple. And it's good to understand that that stuff's not magic.
Starting point is 02:45:28 You can even easily build it yourself. So writing your own little clod code. I even did this at a conference in Paris for people to introduce them to AI. I think it's a fun little practice. You covered a lot. I think one silly idea I had that turned out to be quite cool is I built this thing with full system access. So it's like, you know, it's great power becomes great visibility.
Starting point is 02:46:02 And I was like, how can I up the stakes a little bit more? Yeah, right. And I just made it proactive. So I added a prompt. Initially it was just a prompt surprise me. Every like half an hour. Surprise me, you know? And later on I changed it to be like a little more specific
Starting point is 02:46:22 and in the definition of surprise. But the fact that I made it proactive and that it knows you and does it cares about you, at least it's programmed to that, prompted to do that, And that is a follow on on your current session makes it very interesting because it would just sometimes ask a follow-up question or like, how's your day? I mean, again, it's a little creepy or weird or interesting, but hype it very in the beginning is still today, it doesn't, the model doesn't choose to use it a lot. By the way, we're talking about heartbeat, as you mentioned, the thing that regularly acts. You just kick off the loop.
Starting point is 02:47:18 Isn't that just a cron job, man? Yeah, right. The criticisms that you get a hard. You can deduce any idea to like a silly, yeah, it's just a cron shop in the end. I have like separate crown shops. Isn't love just evolutionary biology manifesting itself? and aren't you guys just using each other? And the project is all just glue
Starting point is 02:47:43 of a few different dependencies and there's nothing original. Why do people, isn't Dropbox just FTP with extra steps? I found it surprising where I had this, I had a shoulder operation a few months ago. And the model rarely used heartbeat, but then I was in the hospital.
Starting point is 02:48:05 And it knew that it had the operation and it checked up on me. It's like, are you okay? And I just, it's like, again, apparently, like, if something significant in the context, that triggered the heartbeat, when it rarely used the heartbeat. And it does that sometimes for people,
Starting point is 02:48:25 and that just makes it a lot more relatable. Let me look this up on perplexity, how open claw works, just to see if I'm missing any of the stuff. local agent runtime, high level architecture. Oh, we haven't talked much about skills, I suppose.
Starting point is 02:48:42 Skill Hub, the tools and the skill layer, but that's definitely a huge component and there's a huge growing set of skills. You know what I love that half a year ago, like everyone was talking about MCPs. And I was like,
Starting point is 02:48:56 screw MCPs. Every MCP would be better as a CLI. And now this stuff doesn't even have MCP support. I mean, it has with asterisks,
Starting point is 02:49:11 but not in the core layer and nobody's complaining. So my approach is if you want to extend the model with more features, you just build a CLI and the model can call the CLI,
Starting point is 02:49:30 probably gets it wrong, calls the help menu, and then on demand loads into the context what it needs to use, to use the CLI. It just needs a sentence to know that the CLI exists if it's something that the model doesn't know by default.
Starting point is 02:49:46 And even for a while, I didn't really care about skills, but skills are actually perfect for that because they boil down to a single sentence that explains the skill, and then the model loads the skill, and that explains the CLI, and then the model uses the CLI. Some skills are like raw,
Starting point is 02:50:04 but most of the time, that works. It's interesting. I'm asking Proplexity, MCP versus Skills, because this kind of requires a hot take that's quite recent because your general view is MCPs are dead-ish. So MCPs is a more structured thing. So if you listen to Proplexity here, MCP is what can I reach?
Starting point is 02:50:29 So APIs, database of services, files via protocol. So a structured protocol of how you communicate with a thing. And then skills is more, how should I work? procedures, hostile, help prescripts and prompts, often written in a kind of semi-structured natural language, right? And so technically skills could replace MCP if you have a smart enough model. I think the main beauty is that models are really good at calling Unix commands.
Starting point is 02:50:58 So if you just add another CLI, that's just another Unix command in the end. And MCP is that has to be added in training. that's not a very natural thing for the model. It requires a very specific syntax. And the biggest thing, it's not composable. So imagine if I have a service that gives me better data, and it gives me the temperature, the average temperature, rain, wind, and all the other stuff, and I get like this huge blob back.
Starting point is 02:51:28 As a model, I always have to get the huge blob back. I have to fill my context with that huge blob, and then pick what I want. there's no way for the model to naturally filter unless I think about it proactively and add a filtering way into my MCP. But if I would build the same as a CLI and it would give me this huge blob,
Starting point is 02:51:49 it could just add a JQ command and filter itself and then only get me what I actually need or maybe even compose it into a script to do some calculations with the temperature and only give me the exit output and you have no context pollution. Again, you can solve that with sub-agents and more charades, but it's just like workarounds for something that might not be the optimal way.
Starting point is 02:52:19 It definitely was good that we had MCPs because it pushed a lot of companies towards building APIs, and now I can look at an MCP and just make it into a CLI. But this inherent problem that MCPs by default, clutter up your context. Plus the fact that most MCPs are not made good, in general, make it just not a very useful paradigm.
Starting point is 02:52:47 There's some exceptions like Playwright, for example, that requires state, and it's actually useful, that is an acceptable choice. So Playwright used for browser use, which I think it's already in OpenClau is quite incredible, right? Yeah.
Starting point is 02:53:04 You can basically do everything. most things you can think of using browser use. That gets into the whole arch of every app is just a very slow API now, if you want or not. And that through personal agents, a lot of apps will disappear. You know, like I had a, I built a CLA for Twitter. I mean, I just reverse engineer the website and used the internal API, which is not very allowed. It's called bird, short-lived.
Starting point is 02:53:45 It was called bird because the bird had to disappear. The wings were clipped. All they did is they just made access slow up. You're not actually taking a feature away. But now if your agent wants to read a tweet, it actually has to open the browser and read the tweet. and it will still be able to read the tweet. It will just take longer.
Starting point is 02:54:06 It's not like you're making something that was possible, not possible. No, now it's just ticking, and now it's just a bit slower. So it doesn't really matter if your service wants to be an API or not. If I can access it in the browser, it is API. It's a slow API. Can you empathize with our situation? Like, what would you do if you were Twitter, if you were X? because they're basically trying to protect
Starting point is 02:54:35 against other large companies scraping all their data. But in so doing, they're cutting off like a million different use cases for smaller developers that actually want to use it for helpful, cool stuff. I think if you have a very low per day baseline,
Starting point is 02:54:53 per account that allows read-only access with solve a lot of problems. There's plenty of automations where people create a bookmark and then use open cloud to like find the bookmark, do research on it, and then send an email with like more details on it or a summary. That's a cool approach.
Starting point is 02:55:13 I also want all my bookmarks somewhere to search. I would still like to have that. So read-only access for the bookmarks you make on X. It seems like an incredible application because a lot of us find a lot of cool stuff on X. We bookmark. That's the general process of X. It's like, holy shit, this is awesome. Oftentimes you bookmark so many things.
Starting point is 02:55:32 you never look back at them. It would be nice to have tooling that organizes them and allows you to research your friend. Yeah, I mean, and to be frank, I mean, I told Twitter proactively that, hey, I built this and there's a need. And they've been really nice,
Starting point is 02:55:48 but also like, take it down. Fair, totally fair. But I hope that this woke up the team a little bit that there's a need. And if all you do is making it slower, you're just reducing access to your platform. I'm sure there's a better way. I also, I'm very much against any automation on Twitter. If you tweet at me with AI, I will block you.
Starting point is 02:56:18 No first strike. As soon as it smells like AI, and AI still has a smell, especially on tweets. It's very hard to tweet in a way that does look completely. human and then I block like a FS zero tolerance policy on that and I think it would be very helpful if
Starting point is 02:56:39 they if like tweets done via API would be marked maybe there's some special cases where and there should be a very easy way for agents to get their own Twitter account we need to rethink
Starting point is 02:57:00 social platforms a little bit. If we go towards the future where everyone has their agent and agents maybe have their own Instagram profiles or Twitter accounts or can like do stuff on my behalf, I think it should very clearly remark that they are doing stuff on my behalf and it's not me. Because content is now so cheap, eyeballs are the expensive part. And I find it very triggering when I read something and then I'm like, oh, no, this smells like AI. Yeah. Like where is this headed in terms of what we value about the human experience? It feels like we'll move more and more towards in-person interaction.
Starting point is 02:57:44 And we'll just communicate, we'll talk to our AI agent to accomplish different tasks, to learn about different things, but we won't value online interaction because there'll be so much AI slop that smells and so many bots that it's difficult. Well, if it's marked, then it shouldn't be difficult to filter. And then I can look at it if I want to. But yeah, this is like a big thing we need to solve right now. Especially on this project, I get so many emails that are, let's say, nicely, agentically written.
Starting point is 02:58:28 but I much rather read your broken English than your AI slop. Of course, there's a human behind it, and yeah, they prompt it. I much rather read your prompt than what came out. I think we're reaching a point where I value typos again. Yeah. Like, you know, I mean, I also took me a while until I come to the realization. On my blog, I experimented with creating a blog post with agents and ultimately it took me about the same time
Starting point is 02:59:07 to steer an agent towards something I like, but it missed the nuances how I would write it. You can steer it towards your style, but it's not going to be all your style. So I completely moved away from that. everything I blog is organic handwritten and maybe I use AI as a fix my worst typos
Starting point is 02:59:36 but there is value in the rough parts of an actual human. Isn't that awesome? Isn't that beautiful? That now because of AI we value the raw humanity in each of us more. I also realized the thing that I rave about AI and use it so much for anything that's code,
Starting point is 03:00:02 but I'm allergic if it's stories. Right, yeah. Also documentation is still fine with AI, you know, better than nothing. And for now it's still, it applies in the visual medium too. It's fascinating how allergic I am to even a little bit of AI slop in video and images. It's useful, it's nice if it's like a little component of like... Or even those images, like all these infographics and stuff, they trigger me so hard. Like it immediately makes me think less of your content.
Starting point is 03:00:36 And they were novel for like one week and now it just screams slop. Yeah. Even if people work hard on it using, and I have some of my blog post, you know, in the time where I explore this new medium. But now they trigger me as well. It's like, yeah, this is, this just screams the eye slop. I don't know what that is, but I went through that too. I was really excited by the diagrams. And then I realized in order to remove from them hallucinations,
Starting point is 03:01:06 you actually have to do a huge amount of work. And you're just using it to draw the better diagrams. Great. And then I'm proud of the diagram. I've used them for literally, like kind of like you said from me, a couple of weeks. And now I look at those, and I feel like I feel when I look at Comic Sans as a font or something like this.
Starting point is 03:01:24 It's like, no, this is fake. It's fraudulent. There's something wrong with it. It's a smell. It's a smell. And it's awesome because it reminds you that we know there's so much to humans that's amazing. And we know it. We know it when we see it.
Starting point is 03:01:48 And so that gives me a lot of hope. That gives me a lot of hope about the human experience. is not going to be damaged, but it's only going to be empowered as tools by AI. It's not going to be damage or limited or somehow altered to where it's no longer human. So, I need a bathroom break. Quick pause. You mentioned that a lot of the apps might be basically made obsolete. You think agents will just transform the entire app market?
Starting point is 03:02:23 Yeah. I noticed that on Discord, people just said how they, like what they build and what they use it for. And like, why do you need my fitness pal when the agent already knows where I am? So can assume that I make bad decisions when I'm at, I don't know, Waffle House, what's around here? Or briskets in Austin? There's no bad decisions around briskets, but yeah. No, that's the best decision. honestly.
Starting point is 03:02:56 Your agent should know that. But it can modify my gym workout based on how well I slept. Or if I'm, if I stress or not, it has so much more context to make even better decisions than any of the SAP even could do. It could show me UI, just as I like. Why do I still need an app to do that?
Starting point is 03:03:19 Why do I have to, why should I pay another subscription for something that the agent can just do now. And why do I need my eight sleep app to control my bed when I can't tell the agent to, no, the agent already knows where I am so
Starting point is 03:03:35 you can turn off what I don't use. And I think that will translate into a whole category of apps that are no longer, I will just naturally stop using because my agent can just do it better.
Starting point is 03:03:52 I think you said somewhere that it might kill off 80% of apps. Yeah. Don't you think that's a gigantic transformative effect on just all software development? That means it might kill off a lot of software companies. Yeah. It's a scary thing.
Starting point is 03:04:10 So, like, do you think about the impact that has on the economy, on just the ripple effects it has to society, transforming who builds, what tooling. It empowers a lot of users to get stuff done, to get stuff more efficiently, to get it done cheaper. There's also new services that we will need, right? For example, I want my agent to have an allowance. Like, you solve problems for me. Here's like a hundred bucks in order to solve problems for me. And if I tell it to order me food,
Starting point is 03:04:51 maybe it uses a service. Maybe it uses something like rent a human to just get that done for me. I don't actually care. I care about solve my problem. There's space for new companies that solved as well. Maybe not all apps
Starting point is 03:05:10 disappear, maybe some transform into being API. So basically apps that rapidly transform in being agent facing. So there's a real opportunity for like Uber Eats that we just used earlier today. It's companies like this of which there's many. Who gets their fastest
Starting point is 03:05:36 to being able to interact with OpenClaw in a way that's the most natural, the easiest. Yeah. And also. apps will become API if they want or not because my agent can figure out how to use my phone. On the upper side, it's a little more tricky. On Android, people already do that.
Starting point is 03:06:00 And then it will just click the order Uber for me button for me. Or maybe another service. Maybe there's an API, it can cause, it's faster. I think there's a space with just, beginning to even understand what that means. And I, again, I didn't even that was not something I thought of,
Starting point is 03:06:21 something that I discovered as people used this. I mean, we're still so early. But yeah, I think data is very important, like apps that can give me data, but that also can be API. Why do I need the Sonos app anymore when I can, then my agent can talk to the Sonos
Starting point is 03:06:38 speakers directly? Like my cameras, there's like a crappy app, but they have, they have an API. So my agent uses the API now. So it's going to force a lot of companies to have to shift focus. And it's kind of what the internet did, right? You have to rapidly rethink, reconfigure what you're selling, how you're making money. Some companies will really not like that. For example, there's no CLIF for Google. So I had to like, do I have to do anything myself. and build Gorg, that's like a CLI for Google.
Starting point is 03:07:17 And at the end user, they have to give me the emails because otherwise I cannot use their product. If I'm a company and I try to get Google data, Gmail, there's a whole complicated process to the point where sometimes startups acquire startups that went through that process, so they don't have to work with Google
Starting point is 03:07:42 for half a year to be certified. to being able to access Gmail. But my agent can access Gmail because I can just connect to it. It's still crappy because I need to go through Google's developer jungle to get a key
Starting point is 03:08:00 and it's still annoying, but they cannot prevent me. And worst case, my agent just clicks on the website and gets the data out that way. To browsers. Yeah. I mean, I watched my agent
Starting point is 03:08:13 happily clicked I'm not a robot button. And there is this this whole that's going to be that's going to be more heated. You see companies like Cloudflare that try to prevent bot access and in some ways that's useful for scraping. But in other ways if I'm
Starting point is 03:08:34 a personal user, I want that. You know, sometimes I use Kodax and I read an article about modern React patterns and it's like a medium article. I paste it in and the agent can't read it because they block it. So I have to copy, paste the actual text.
Starting point is 03:08:57 Or in the future, I learned that maybe I don't click on medium because it's annoying and I use other websites that actually are agent-friendly. There's going to be a lot of powerful rich companies fighting back. So it's a really interesting. You're at the center. you're the catalyst, the leader,
Starting point is 03:09:15 and happen to be at the center of this kind of revolution where it's going to completely change of how we interact with services with web. And so like there's companies at Google they're going to push back. I mean, there's every major companies you can think of is going to push back. Even, yeah, even search.
Starting point is 03:09:35 I now use, I think, perplexity or brave as providers because Google really doesn't make it easy to use Google without Google. I'm not sure if that's the right strategy, but I'm not Google. Yeah, there's a nice balance from a big company perspective, because if you push back too much for too long,
Starting point is 03:09:56 you become blockbuster and you lose everything to the Netflix's of the world, but some pushback is probably good during a revolution to see. But you see that this is something that the people want. Right. So, if I'm on the go, I don't want to open a calendar app. I just, I want to tell my agent, remind me about this dinner tomorrow night,
Starting point is 03:10:18 and maybe invite two of my friends, and then maybe send a WhatsApp message to my friend. And I don't need, I don't want, I need to open apps for that. I think that we passed that age. And now everything is like much more connected and fluid if those companies want it or not. And I think we'll,
Starting point is 03:10:39 the right companies. will find ways to jump on the train and other companies will perish. You got to listen to what the people want. We talked about programming quite a bit, and a lot of folks that are developers are really worried about their jobs, about the future of programming.
Starting point is 03:11:00 Do you think AI replaces programmers completely, human programmers? I mean, we're definitely going in that direction. Programming is just a part of building product. So maybe I does replace programmers eventually but there's so much more to that art. Like
Starting point is 03:11:21 what do you actually want to build? How should it feel? How's the architecture? I don't think Asians will replace all of that. Yeah, like just the actual art of programming it will stay there but it's going to be like knitting.
Starting point is 03:11:43 You know like people do that because they like it, not because it makes any sense. So I read this article this morning about someone that it's okay to mourn our craft. And a part of me very strongly resonates with that because in my past, I spent a lot of time sinkering, just being really deep in the flow
Starting point is 03:12:07 and just like cranking out code and finding really beautiful solutions. And yes, in a way, it's sad because that will go away. And I also got a lot of joy out of just writing code and being really deep in my thoughts and forgetting time and space and just being in this beautiful state of flow.
Starting point is 03:12:37 But you can get the same state of flow. I get a similar state of flow by working with agents and building and thinking really hard about problems. It is different. but and it's okay to mourn it but it's not something we can fight like there is the world for a long time
Starting point is 03:12:57 had a there was a lack of intelligence if you see it like that of people building things and that's why salaries of software developers reached stupidly high amounts and then we'll go away there will still be a lot of demand for people that understand how to build things.
Starting point is 03:13:23 Just that all these tokenized intelligence enables people to do a lot more, a lot faster. And it will be even faster and even more because those things are continuously improving. We had similar things when, I mean, it's probably not a perfect analogy, but when we created the steam engine
Starting point is 03:13:49 and they built all these factories and replaced a lot of manual labor and people revolted and broke the machines. I can relate that if you very deeply identify that you are a programmer, that it's scary and that it's threatening because what you like
Starting point is 03:14:13 and what you're really good at is now being done by a souless or not entity, but I don't think you're just a programmer. That is a very limiting view of your craft. You are still a builder. Yeah, there's a couple things I want to say. So one is, I never, as you're articulating this beautifully,
Starting point is 03:14:38 and I'm realizing, I never thought I would, the thing I love doing would be the thing that gets replaced. you hear these stories about these like you said with the steam engine i've spent so many i don't know maybe thousands of hours pouring over code and putting my heart and soul and like and just like some of my most painful and happiest moments or alone behind i was an emacs person for a long time man emacs and and then there's an identity and there's meaning and there's like when i walk about the world, I don't say it out loud, but I think of myself as a programmer. And to have that in a matter of months, I mean, like you mentioned, April to November,
Starting point is 03:15:26 it really is a leap that happened, a shift that's happening. To have that completely replaced is painful. It's truly painful. But I also think programmers, builders more broadly, but what is the act of programming? I think programmers are generally best equipped at this moment in history to learn the language, to empathize with agents, to learn the language of agents, to feel the CLI. Yeah? Like to understand what is the thing you need, you, the agent, need to do this task the best.
Starting point is 03:16:13 I think at some point it's just going to be called coding again and it's just going to be the new normal. And yet while I don't write the code, I very much feel like I'm in the driver's seat and I am writing the code. You know, it's just... You'll still be a programmer. It's just the activity of a program is different.
Starting point is 03:16:34 Yeah, and because on X, the bubble, I mean, is mostly positive on MasterDone and Blue Sky. I also use it less because oftentimes I got attacked for my blog posts and I had stronger reactants in the past now I can sympathize with those people more because in a way I get it in a way I also don't get it
Starting point is 03:16:59 because it's very unfair to grab onto the person that you see right now and unload all your fear and hate it's going to be a change and it's going to be challenging but it's also I don't know I find it incredibly fun and gratifying and I can use the new time to focus on much more details. I think the level of expectations of what we build is also rising because it's just now,
Starting point is 03:17:31 the default is now so much easier. So software is changing in many ways. There's going to be a lot more. And then you have all these people that are screaming, oh yeah, but what about the water? You know, like I did a conference in Italy about the state of AI. And my whole motivation was to push people away from, don't see yourself as an iOS developer anymore, you're not a builder.
Starting point is 03:18:00 And you can use your skills in many more ways. Also because apps are slowly going away. People didn't like that. Like a lot of people didn't like what I had to say. And I don't think I was hyperbole. I was just like, this is how I see the future. maybe this is not how it's going to be, but I'm pretty sure a version of that will happen.
Starting point is 03:18:23 And the first question I got was, yeah, but what about the insane water use on data centers? But then you actually sit down and do the mass, and then for most people, if you just give one burger per month, that compensates the CO2 output or like the water use in the equivalent of tokens. I mean, the mass is tricky
Starting point is 03:18:48 and it depends if you add pre-training then maybe it's more than just one paddy, but it's not off by a factor of 100, you know? So golf is still using way more water than all data centers together. So are you also hating people that play golf? Those people grab on anything that they think is bad about AI without seeing the potential things
Starting point is 03:19:14 that might be good about AI. And I'm not saying everything's good. It's certainly going to be a very transformative technology for our society. There's, to steal man the criticism in general, I do want to say in my experience of Silicon Valley, there's a bit of a bubble in the sense that there's a kind of excitement
Starting point is 03:19:39 and an overfocus about the positive that the technology can bring. Which is great. It's great to focus on not to be paralyzed by fear and fear-mongering and so on. But there's also within that excitement and within everybody talking just to each other, there's a dismissal of the basic human experience across the United States and the Midwest, across the world, including the programmers would mention, including all the people that are going to lose their jobs, including the measurable pain and suffering that happens at the short-term scale when there's change of any kind, especially large-scale transformative change that we're about to face if what we're
Starting point is 03:20:25 talking about will materialize. And so having a bit of that humility and an awareness about the tools you're building, they're going to cause pain. They will long-term, hopefully bring about a better world and even more opportunities and even more awesomeness. But having that kind of like quiet moment often of respect for the pain that is going to be felt. And so not enough of that is I think done. So it's good to have a bit of that. And then I also have to put against some of the emails I got where people told me they have a small business and they've been struggling and OpenClaw helped them automate a few of the tedious tasks from collecting invoices to like answering customer emails that then freed them up and like
Starting point is 03:21:20 cause them a bit more joy in their life or or some emails where they told me that OpenClaw helped a disabled daughter that she's now empowered and feels she can do much more than before which is amazing right because you could you could do that before as well the technology was there i didn't i didn't invent a whole new thing but i made it a lot easier and more accessible and that did show people the possibilities that they previously wouldn't see and now they applied for good or like also the fact that yes i i suggest the latest and best models but you can totally run this on free models. You can run this locally. You can run this on chemi or other other models that are way more accessible price-wise and still have a very powerful system
Starting point is 03:22:18 that might otherwise not be possible because other things like, I don't know, Anthropics co-work is logged in into their space. So it's not a lot black and white. I got a lot of emails that were heartwarming and amazing and and I don't know just make me really happy. Yeah, there's a lot
Starting point is 03:22:41 it has brought joy to a lot of people's lives not just programmers like a lot of people's lives. It's beautiful to see. What gives you hope about this whole thing we have going on? A human civilization. I mean,
Starting point is 03:22:57 I inspired so many people. There's like, there's this whole build a vibe again. People are now using AI in a more playful way and are discovering what it can do and how it can
Starting point is 03:23:15 help them in their life and creating new places that are just sprawling of creativity. There's like clock corn in Vienna. There's like 500 people.
Starting point is 03:23:30 and there's such a high percentage of people that I want to present, which is to me really surprising because usually it's quite hard to find people that want to talk about what they build, and now there's an abundance. So that gives me hope that we can figure shit out. And it makes it accessible to basically everybody. Yeah. Just imagine all these people building.
Starting point is 03:24:00 especially as you make it simpler and simpler, more secure. It's like anybody who has ideas that can express those ideas in language can build. That's crazy. Yeah, that's ultimately power to the people and one of the beauty, the beautiful things that come out of AI,
Starting point is 03:24:23 not just a slop generator. Well, Mr. Claude Fodd, I just realized when I said that in the beginning, I violated two trademarks because there's also the Godfather and being sued by everybody. You're a wonderful human being. You've created something really special, a special community, a special product,
Starting point is 03:24:47 a special set of ideas, plus the entire, the humor, the good vibes, the inspiration of all these people building, the excitement to build. So I'm truly grateful for everything, you've been doing and for who you are and for sitting down to talk with me today. Thank you, brother. Thanks for giving me the chance to tell my story.
Starting point is 03:25:10 Thanks for listening to this conversation with Peter Steinberger. To support this podcast, please check out our sponsors in the description where you can also find links to contact me, ask questions, give feedback, and so on. And now, let me leave you some words from Voltaire. With great power comes great responsibility. Thank you for listening and hope to see you next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.