Software Huddle - Seattle Startups, AI’s Future & Big Acquisitions with Yujian Tang

Episode Date: March 13, 2025

Today on the show, we talked with Yujian Tang. He was on the show previously when he worked at Zilliz, when we talked about vector databases and RAG. He's since branched out on his own, building the t...ech startup scene in Seattle and organizing AI events all over the place. We talk about his latest venture, the Seattle Startup Summit, coming up on March 28th. They're still Early Bird Tickets available if you're interested. We also talk about AI models, the impact AI is having on programming, including our own programming projects and share our takes on some of the recent acquisitions that have happened in tech, including Voyage AI.

Transcript
Discussion (0)
Starting point is 00:00:00 What do each of you use sort of day to day or like which models or products or things you are using in your day to day? So I mostly like when I do like LLM like work, I mostly just use GPT-4. For programming, I primarily use GPT-4 and quad depending on what the project is. It's very interesting whenever I think about like IBM because it's just like... They're making a lot of moves. It's just like big giant old company that like, you know, they're just, it's like, what is, what is this company doing?
Starting point is 00:00:28 I think our generation like doesn't think about them, right? Like we'd never think they like buy something from IBM, but yet they're, they're buying all these other companies. Yeah. Well, I mean, they're such a huge conglomerate. They're called international business machine, you know, like that's what they are. What a name.
Starting point is 00:00:43 Yeah. That's amazing. When's the summit happening? March 28th. Get your tickets, people. Hey everyone, welcome to Software Huddle. Today on the show, Alex and I talk with Eugene Tang. He was on the show previously when he worked at Zillow's,
Starting point is 00:00:57 when we talked vector databases and RAG. He's since branched out on his own, building the tech startup scene in Seattle and organizing AI events all over the place. We talk about his latest venture, the Seattle Startup Summit, coming up on March 28th. There's still early bird tickets available if you're interested. We also talk AI models, the impact AI is having on programming, including our own programming projects, and share our takes on some of the recent acquisitions that have happened in tech including Voyage AI as always if you have ideas or suggestions for the show, please let us know and with that
Starting point is 00:01:30 Let's get you over the episode. Hello and welcome to software huddle. I'm here with my forever co-host Alex debris Alex. How are you? Sean, I'm doing well. It's good to be back new year new me Yeah, good to see ya. Yeah, we kind of we we Kind of put the podcast a little bit on hold there for for most basically December and January. Wife kind of took over. I switched jobs. You had stuff going on. We kind of had to slow down for a little bit, but we're back in full force now. Yep. Yep. Always always good to get back. Yeah. Definitely got jammed up with like more work stuff than I anticipated for a while.
Starting point is 00:02:03 So it was a nice break. And I feel like, you know, that holiday time always ends up being kind of like a crapshoot anyway. Like, yeah, so good, good to take a break, but good to be back as well. Yeah, people aren't listening to podcasts as much anyway, during December. So it's a good time to take maybe take a break. Yep. Yep. All right. And we have with us back on the show, Eugene. Welcome back. Thanks for joining us. Thank you, Sean. Glad to be back. Yeah. So last time you were on, you were still, you were working at Zillas when we're talking vector databases. So, you know, you've moved on from Zillas. That was quite a while ago. You know, what's going on now?
Starting point is 00:02:41 Yeah. Yeah. So I was working on the Vector Database stuff, working on some Rack stuff, working on some agent stuff back in 2023, early 2024 as well. Basically, what I was doing in Zillis is I was doing these developer communities, right? So I was helping build the community there, trying to get people to come together to talk about AI, talk about what's going on in the space.
Starting point is 00:03:03 And one thing I found about running these events is you're a little bit limited on your topic set, right? You're basically like, hey, we're going to talk about Vector's databases, we're going to talk about RAG. And then that's pretty much it. And so I was like, hey, what if I had a community where we could talk
Starting point is 00:03:22 about a lot of different things, where it's like a whole spectrum? Because as a dev, and I imagine that's why people listen to this as well, is you get to just hear so many different things when it's more third-party-ish. So I started my own community last year. We started with hackathons. We do some tech talks in SF. We do hackathons in Seattle. we grew from about zero to seven thousand members. We have 11,000 people attend our events and this year I basically turned that into something called Seattle Startup Summit which is going to be like a hopefully, hopefully 800 person event. We'll see. We're at about 460 some signups now. So we'll see if we can hit
Starting point is 00:04:06 that 800 mark. Yeah, very cool. When's the summit happening? March 28. Get your tickets people. Is this live? Is this live? This is not live. Oh, okay. Okay. Okay. Yeah. It's March 28. So it's a couple of weeks. Yeah.
Starting point is 00:04:22 When you're doing an event like this, like, you know, say you want 800 people to be there, do you have to like oversubscribe? Like, what is sort of the attrition rate with this sort of stuff? Yeah, 100% you have to oversubscribe, right? So with free events, you're pretty much looking at it actually kind of depends on the location too, because it's like the culture of the place, right? So in SF, you get about a third of people show up. If you're lucky and you have a really good event, you get maybe 40, 45%.
Starting point is 00:04:49 If you're like Databricks, you probably get 50%, you know? Or, you know, one of those big summits. With a paid event, or in Seattle, you get about 50%. With a paid event, so like Seattle Startup Summit, I'm charging for tickets, right? I do that primarily because I want people who actually sign up to show up. But even with charging for tickets,
Starting point is 00:05:10 even with charging for tickets, I found that, you know, you still have maybe like a 10 to 20% of people that show up. And so like the tickets I'm charging for the summit are like $250, like full price tickets. But a lot of people buy like, you know, half off tickets if they can find a code or, you know, they get one of the early price tickets. But a lot of people buy half off tickets if they can find a code or they get one of the early bird tickets.
Starting point is 00:05:29 And then I'm like, later you sign up for a conference or later you sign up for an event, the more likely you are to show up. So I don't know what the interest rate is going to be. I hope it's going to be like I'm aiming at 1,000 signups. So that's what I'm aiming for. We'll see if we get there. Yeah.
Starting point is 00:05:46 Cool. And what kind of person are you targeting for devs, ad startups, people that are interested in startups? What are you looking for here? Yeah, yeah. So my primary goal with this community that I build is, hey, AI is changing everything. It's not just devs and software.
Starting point is 00:06:06 It's like healthcare and finance and all this kind of stuff. But I think on the front line, devs are your target audience. Before you can have your like AI finance apps, AI healthcare apps, you're going to need devs that are literate in this space. And so my goal to community is really to get devs all the information, all the resources, all the tutorials, all the networking, all that they need to become literate in this space. And so my goal of the community is really to get devs all the information, all the resources, all the tutorials, all the networking, all that they need to become literate in this space.
Starting point is 00:06:30 So my primary audience is devs. And then, you know, like secondary audience, like tech leadership, of course you have to have, you know, tech leadership there. They got to buy in founders, people like that. Actually, one of the things with Seattle Startup Summit that I feel very strongly about is, oh, so Sean, I know you're in the Bay,
Starting point is 00:06:50 Alex, where are you based? I'm in Omaha, Nebraska, so I'm always away, yeah. Okay, okay. Yeah. Okay, so. Not a tech hub, that's for sure. Not a tech hub, yes. But so Seattle is like one of those weird places
Starting point is 00:07:04 that's like, it's a tech hub in the sense that there's a lot of technical talent here. There's Amazon, there's Microsoft, there's Snowflake, there's Databricks, everybody's here basically. There's actually not a lot of startup activity. If you go and you look at what people are talking about, you can see that when you hear people talking about tech hubs and AI,
Starting point is 00:07:24 you always hear people say something about Seattle. But then if you hear people talking about tech hubs and AI, you always hear people say something about Seattle. But then if you hear people talk about startups, you'll never hear people say anything about Seattle. Unless they're talking about, here's a tier two startup city. Why is that? Yeah. Yeah, that's actually one of the things that I was like, what is up here? What is up with this? My opinion is, one of the things that I want to do is I want
Starting point is 00:07:45 to nurture that developer to start a founder pipeline. And I think part of it, and once again, I'm not entirely sure this is just like, you know, you're just taking shots in the dark until something sticks, right? Like my opinion is like, there just needs to be more activity around it. You know, there's not as much activation. If you go to the Bay, like Sean, you've probably see this. There's events every day of the week.
Starting point is 00:08:11 Yeah. I mean, I think that's why the attrition rate for events is so high is because people get invited to so much stuff. They just sign up to all of them and then they're like, I'm kind of tired today, so I'm not going to go. Yeah, exactly. So that's one of the things where I'm like, yeah, you know what? We should have this in Seattle. Yeah, it is surprising because Seattle, it's not like Seattle, like it from like a proximity to like a VC hub, like Seattle is
Starting point is 00:08:36 not that far away, you can do a day trip to San Francisco if you needed to hit up, you know, you know, a bunch of VCs or something like that to run through pitches. I wonder, in other cities in the US, like New York, Boston, cities that do have a pretty decent startup culture and a number of startups there, what is it that they did in order to build that?
Starting point is 00:09:01 Is it that they have some sort of incubator or something like that, or some prominent person? I also think is that they have like, you know, some sort of incubator or something like that, or some prominent person. I also think sometimes that you have like one company that emerges as like a really successful startup out of a place and then suddenly everybody's like, oh, okay, like I can do a startup in, you know, wherever Omaha, Nebraska. You know, I think about that too. And so like, SF has a bunch of different startups that are very successful, obviously. Seattle has had two huge successes in the last like, well, now it's like 50 years. But you know, like, yeah, why doesn't that fling off a bunch of other sort of people
Starting point is 00:09:39 leaving there and starting something else and just like creating a whole ecosystem like like we've seen in other places? Oh, okay. So I actually have looked into this a bit. And here's what I think that is. and starting something else and just creating a whole ecosystem like we've seen in other places. Oh, okay, so I actually have looked into this a bit and here's why I think that is, okay? So Amazon and Microsoft are so big and they pay you so much money that it is very hard to leave. On top of that, a lot of their employees,
Starting point is 00:09:57 a lot of their software developer employees are actually people who are on H1B visas. Interesting, okay. And you think that's more, I mean, cause like that's, some of that's gotta be true about the pay in the Bay area as well. But then do you think there's less of the,
Starting point is 00:10:13 I guess H1B just proportionally in SF compared to Seattle? It could be. So Seattle has like the two big, like Amazon, Microsoft companies that do that. I think Google and Meta, Google, Meta, Apple, they all they all do that as well. I think a couple of other things that are at play here as well is like the Bay has Stanford, Berkeley, right? These are very big school.
Starting point is 00:10:37 Well, I get a Berkeley is a very big school. Stanford is not that big, but it turns out a lot of people. You we have you dub up here, which is one of the research institutes. We also UW up here, which is one of the AI research institutes. We also have AI2, which is a really good AI research hub as well, research institute as well. The thing with this is AI2 is tiny, and UW is roughly the size of Berkeley.
Starting point is 00:10:57 And so also, there's a lot of people who go to UW and then go to AI2. So it's not kind of like the same as you have down there with Berkeley and Stanford. And then also Silicon Valley, it started because it got its name from the people who were doing the Silicon chips back in the 60s, right? Like Hewlett Packard and whatever.
Starting point is 00:11:22 And they were successful and they have a ton of money and they're hanging out in Palo Alto. So you've got like, you know, you walk right the whole like, the whole like startup, like shtick meme, whatever, is like you're walking down. What is that? Stan, Stan Hill Road? Yeah, yeah, yeah, that road and you know, there's like billions of dollars like right there. And in Seattle, you don't really have that. We have Madrona. Madrona has like $3 billion under management. Here's another plug, John Turrow from Madrona,
Starting point is 00:11:51 he's a partner who's going to be coming to speak at Seattle Startup Summit. But there's just not as much money, like MVC money here. There's a lot of money in tech money. People who, there's a lot of money in like tech money, right? Like people who like there's a Microsoft executives. I don't know if you have you guys have either of you guys been up here? Yeah, I was there last month. Yeah. Oh, yeah.
Starting point is 00:12:13 Oh, okay. Cool. So in Bellevue, there's this place called Medina, Medina, something like that. And this neighborhood is literally like just mansions. And that's all the Microsoft execs, all the Amazon execs. So there's a lot of money up here, but I just don't think it's being activated in the same way. Boston, you know, Boston also has MIT, Harvard, a lot of like medical stuff out there as well. Yeah, Boston, I think, has a good base of bioinformatics, biomedical-focused startups.
Starting point is 00:12:49 We're only tech-seeing around that, which I think has emerged from those universities. Yeah. Yeah. And of course, New York has Wall Street. Yeah. Yeah, exactly. So what's the agenda for the summit?
Starting point is 00:13:02 What's the, if I'm attending, what should I expect? Yeah, so I think the agenda for the summit? Like, what's the, you know, if I'm attending, what should I expect? Yeah, so I think like the idea with this summit is like really like, hey, I'm gonna try to gather all of the people who are working on startups in Seattle. Now I haven't gathered everyone, but I've gathered a lot of people. So at the beginning, we're gonna start,
Starting point is 00:13:20 we're gonna talk with John Turow. I haven't like, I don't know who's gonna moderate this yet, but we're gonna have a talk with John Turow. I haven't like, I don't know who's gonna moderate this yet. But we're gonna have a talk with John, he's he's the partner of Amazon, and he's gonna talk to us about AI, AI agents, AI infrastructure, right. So this guy was building like a voice agent, like, company back in like 2011, 2012. And, you know, he got he got out of it, he went to Amazon, he built for like 10 years. And now he's kind of like seen the investor side,
Starting point is 00:13:45 the startup side, et cetera. So he's gonna be giving the first fireside chat. The second fireside chat that we're gonna do is with Pasquale De Mayo. So he's currently the head of Amazon Connect, right? So that's their customer service platform. The story that we wanna cover there is like, he was part of that journey from the beginning
Starting point is 00:14:12 where they took Amazon Connect from zero to 150 million in like less than four years. And that's something that I think a lot of founders would find very interesting, even though they were internal to Amazon, it's still like, that's a lot of revenue very quickly. And then the third one is we're doing a fireside chat with Vijay Raji.
Starting point is 00:14:33 So he's the CEO of StatSig. Have you guys heard of StatSig? Yeah. Yeah. They've been doing a lot of stuff in the Bay. They're like experimentation. And one of the things I talked with his team about his statistics in person
Starting point is 00:14:48 five days a week, by the way, and they're based out here in Seattle, and they do stuff in the Bay. So I was like, I was like, dang, like you guys all travel like that far to the Bay like, you know, pretty often do this. And they're like, yeah, we're really interested in having more people come to Seattle. I was like, me too. So, you know, that's part of it. And then we'll have some, we'll have a couple of panels. So let's see who's on our first panel.
Starting point is 00:15:11 So we got Alex Ye from GMI Cloud. So GMI Cloud is one of the Neo Clouds. I think we'll touch on this later as well. But they're one of the GPU Clouds. They just raised like 80, 8282 million, something like that. Then we'll have, who else is on that panel? Tony, Tony Holdstock Brown from Ingest. So he used to be in the healthcare space actually.
Starting point is 00:15:35 And he's got an interesting story. So Ingest was like an orchestration platform. So you guys know like AWS like step functions, right? Yeah. Yeah, so they do something kind of like step functions. And then like when the whole AI agent like thing picked up, they found that people were using what they were building to build AI agents.
Starting point is 00:15:53 And so they kind of like pivoted into this space. And then there's also Rob Williams from ReadAI. So for context, Riverside is the third streaming platform that I've been on. ReadAI is like one of like seven, like, you know, meeting assistants that I've seen. And they're doing quite well. They just raised the $50 million Series B. I think they're expanding into the enterprise. So that panel is going to be talking about the AI stack.
Starting point is 00:16:23 You'll notice that all three of these people are from different pieces of the stack. So Alex is from the super infrastructure core layer, like GPU cloud. And then Tony is the orchestration framework layer. Like you want to build an agent, this is what you use. And then Rob is from this application layer, where it's like,
Starting point is 00:16:48 here's what we're doing, here's how we're using AI, right? And then our other panel, I've titled it Enterprise Insights, which honestly is a very vague title, but these people work with enterprises, right? So we have Sharon Zhou from Lamini. So by the way, I don't know if you guys know this company, right? Lamini. Have you heard of this company? No, no, I don't know that one. Okay. Well then you, you won't have, you won't, you won't, you won't have
Starting point is 00:17:13 the context for what I'm about to say, but it's spelled L A M I N I. And I thought to myself, I was like, Oh, this has got to be pronounced Lamini. And then I talked to one of my friends and she was pronouncing it Lamini. And I was like, what is this has got to be pronounced Lamini. And then I talked to one of my friends and she was pronouncing it Lamini. And I was like, what is this? How do you say this? And then I got on a call with their team and they're like, oh, it's pronounced Lamini. And I was like, how am I supposed to guess this?
Starting point is 00:17:34 And then they're like, no, it's a real word. It actually means multiple llamas. It's a type of llama or llama types or something like that. I was like, wow, okay. Wait, like a herd of llamas is a Lamini? Yeah, yeah, yeah. Yeah, something like that. And I was like, wow, okay. Wait, like a herd of llamas is a llama night? Yeah, yeah, yeah. Yeah, something like that. You didn't know that, Alex?
Starting point is 00:17:50 I did not know. I'm on their website right now and in the M of their thing, it's got a little llama embedded in there. So I could see it's llama related, but it's weird because it doesn't look like llama generally. It's only got one L rather than the two. Yeah, so. Does this have anything to do with like, the llama model from
Starting point is 00:18:10 meta? Or is this completely unrelated? No, it's totally unrelated. That's a lot of llama on like, I would be a little bit worried about sort of trying to use another mama related word for something that's so well-known and branded from meta. Yeah, I don't know, man. I think people, because it's LLM, people are like, oh, llamas, llamas.
Starting point is 00:18:37 But yeah, so they're like an AMD cloud, right? So they basically provide you AMD chips. And Sharon is interesting in the sense that she is a student of Andrew Ong, who is super, super popular in this space. And then, who else on that panel? Denny Lee from Databricks. You guys know Denny?
Starting point is 00:18:57 Yeah. Yeah, yeah. He gets around. He's everywhere. You can't not know Denny if you're in this space. And then Yina, I don't know if I'm saying this correctly, I think it's Yina Arenas, but she's the VP of Azure AI, Azure Core AI,
Starting point is 00:19:13 so they're like new AI platform. She's also based in Seattle. So at first I thought she was based in the Bay, I was like, oh, yes, we're gonna get her to come out from the Bay for this. And I was like, oh, she's in Seattle. I was like, oh, even better. She has tech leadership from AI in Seattle. So yeah, that's those are the panels. And then afterwards, we'll have a lunch break and you know, stuff
Starting point is 00:19:31 like that. And during the day, there's like a bunch of exhibit tables, you know, like when you go to a conference, there's always exhibit tables, you walk up to the zip tables, you talk to the people see what's going on. In the afternoon, this is my like, I would say like my like defining feature that I do in my events is like the demos, this is my defining feature that I do in my events, is the demos. It's always really fun to see lightning demos. If you ever come to my events in the Bay,
Starting point is 00:19:51 you'll see that I have people do these two-minute demos. And after the two minutes is up, I will shush them off the stage. And I'll be like, get off the stage, get off the stage. And so at the end, at the afternoon, we'll have four-minute demos. I'm giving people a little bit longer. It's a little bit big. Very generous. I know, right?
Starting point is 00:20:09 But we're having these four minute demos at the end. And currently we're at 26. I have spots for, oh no, 28. I have spots for 52 of them. So we'll see if we get up to 52. If it's 39, I'm happy to just cut off the last 13. I'm doing like 13 an hour. So. And so these will be like AI companies showing off what they're doing.
Starting point is 00:20:32 Partly with the idea to like, I mean, are they attracting dev talent? Are they attracting investment? Are they attracting customers? Like what's the goal there? Yeah, yeah, yeah. So there's the idea behind this is like, the original idea behind this was that,
Starting point is 00:20:46 and it's kind of evolved a little bit, but the original idea behind this was like, some early stage companies, and I've given a lot of free demo slots to early stage companies, is early stage companies really need product feedback, right? Like devs, here's what devs are good at, building things. Here's what devs are not good at, talking to people.
Starting point is 00:21:05 I was like, well, what I can help you do is you can come and show us what you're building, and then people will come to you. You don't even have to go out and find them. They will come to you and they'll tell you, hey, this is really cool or hey, have you thought about this or blah, blah, blah, blah. That was the original idea behind this.
Starting point is 00:21:22 Then for this event, this has evolved into like, OK, like we, our theme is AI dev tools. Let's try to get a whole view, like, you know, just adding in that idea of like getting like the whole view of the landscape. Let's see as many AI dev tools as we can and like get to kind of like compare and contrast. That's cool. Because I find like, you know, I talked to a lot of these companies that are interested
Starting point is 00:21:43 in doing, like every company in the world basically has some sort of AI initiative. But I'm saying like 99% of them have no idea what that actually looks like. So a lot of it comes down to, I think them seeing like the art and the possible. So if they see like what other companies are doing, and then they're like, oh, okay, like, we could do something similar, or we could, you know, maybe we can work with these people or whatever. But I think there's such a huge
Starting point is 00:22:09 interest in this space, but very few people who kind of have like a definitive plan for what that means. Yeah, 100%. I think that's that's like, that's definitely part of it. And I think also, you touched on something else there, which was like, you got to see other people do things. So you kind of like have an idea of it. That's also part of it for me is like, about I think about like 70% of people
Starting point is 00:22:31 who have signed up for this conference are like some sort of builder, like product or engineering, some sort of builder like that. And I think like one of the other things is like, if you see other people doing startups, you see other people who are devs that move into this founder kind of role,
Starting point is 00:22:44 you're more likely to go and take that job. Take that like, hey, I know this guy. I know what he can do. Maybe I can do that too. I know what she can do. Maybe I can do that too. That's also part of this as well. Awesome. Well, it sounds like a great event. It looks like you got a good set of speakers there. Should be really interesting for folks.
Starting point is 00:23:07 Hopefully, if you're in the Seattle area or you want to travel there and you're listening to this, you'll sign up. One last question. If you happen to be in the Omaha area, is there a livestream or anything like that I can watch? I especially want to see the demos and see all these things.
Starting point is 00:23:20 I think those are so fun. That's such a great idea. I don't know if we'll livestream, but I just, oh my God. Okay, so this conference is like $100,000 to do. And this was like before getting like the recording AV. And so I just was like, you know what? We should record this. How much more is that gonna be?
Starting point is 00:23:36 And they're like, oh, it's gonna be another $10,000. And I was like, okay, well, we'll figure this out then. And so I was like, all right, we're almost at the breakeven point. And you know what? I just went ahead and got the recording package. We'll record it. I don't know if we'll stream it, but we will record it at least.
Starting point is 00:23:49 Okay. That's perfect. I just want to see all these demos and watch that. Yeah, that'd be great. Yes. I will let you know, Alex. It's like the YC demo day. Exactly. It's like demo day, but in Seattle. Yeah, that's awesome. Cool. Let's get into some other AI-related topics here.
Starting point is 00:24:09 I think there's been a lot, certainly since Alex, you and I last talked, but even the last month or so, just like models, models, and more models everywhere. DeepSeq basically came out, broke the internet. Now almost like DeepSeq seems like old news at this point with everything that has happened since that. It's like, you know, quad three sevens out, OpenAI is launching 4.5 soon. As response to DeepSeek, like Google has Gemini Flash 2, OpenAI has O3 mini. Like, all so much stuff is happening.
Starting point is 00:24:43 You know, maybe we'll go to you, Jensen, you've been working in this space for a long time. Like, what's your thoughts on this? Do you have any specific takes? I mean, I love it. I think it's great. There's a lot of AI models coming out. I think that's always fun.
Starting point is 00:24:56 Competition is what makes people make more like interesting things. There's a lot of talk around GPT 4.5. So I just did an event in SF last week where it was a full day AI production mini conference. Basically, people were just talking about how GPT 4.5 is really expensive. That's the number one takeaway that I think people are getting right now. Someone said that O3 was really nice. I didn't get to see a demo of it, but someone did say that O3 was really nice.
Starting point is 00:25:26 I'm actually, for my workflows, I'm using GPT-4. I'm like, I don't even, I'm not even doing anything that complex. So I'm still just playing around with GPT-4. I think, you know, like, I don't know if I mentioned this to you last time, Sean, but I think one of the things that I find particularly interesting about this is everybody that,
Starting point is 00:25:46 at least that we can see, is still on this kind of like transformer model architecture, right? And then last year, literally December 31st, 2024, the last day of last year, and I don't even know how they got people to do this, release this at the end of December, but Google released this paper that is on like a slightly different architecture.
Starting point is 00:26:05 And it's called Titans. And they use this thing called a long-term memory module. And I think I would like to see... I don't know if Flash 2 is built on this. I don't know anything about Flash 2, actually. I don't know a ton, but I believe it's a derivative model from their existing Gemini models because it came out so quickly. I think a lot of them are using a lot of these models, even though the incremental 0.2, 0.3 are usually using the same base model, and then they're doing some additional fine-tuning
Starting point is 00:26:39 on top of that to create a new version of it. It's probably still based on the Transformer model, but this new paper they released called Titans has this interesting thing called a long-term memory module, which makes it kind of work. So it's a lot smaller than Transformer models, which means it's more efficient, which means it's less costly, it's going to less like energy cost,
Starting point is 00:27:00 it's better for the environment, et cetera, et cetera, et cetera. But the thing I find interesting about it is it learns on... It learns kind of like people do, right? So what the long-term memory module does is it says, if you see something that's very statistically unlikely to happen based on your current data, you should remember that. So, for example, if you're walking down the street and all of a sudden one of the trees turns red, you're going to remember that, you're like walking down the street and all of a sudden, like, one of the
Starting point is 00:27:26 trees turns red. You're going to remember that because you're like, what's going on here? This is not, this is not unusual. So I think that's like an interesting architecture thing that's happening. I think like DeepSeq, for example, the whole like distillation model thing. I don't really think it's quite the open AI killer that people have been kind of like making it out to be. It is very popular, of course. But the whole like distillation model thing, like, yes, you can train it for a few
Starting point is 00:27:57 million dollars, but that's the last training run. They still took data from open AI, you know, they still need that that large language model data. And then the primary cost is inference. So I don't think it's quite as much of a breakthrough as people think it is. But yeah, that's kind of what my opinions are. What about you, Alex?
Starting point is 00:28:14 What about you, Sean? I mean, the interesting thing about DeepSeek is just the state of the art, it seems like, is still going to be super expensive to train all the time. But then anyone is going to be able to replicate that pretty quickly and easily. So it's just like, it seems like is still gonna be super expensive to train all the time, but then anyone is gonna be able to replicate that like pretty quickly and easily. So it's just like, I guess it's gonna be a super decentralized technology, right? You know, like it's gonna be hard to stop in terms of just like, if you get a new model out there, people are just gonna be able to copy that and really say open source or whatever
Starting point is 00:28:38 they want to do there like pretty easily. And like we've seen like cost of tokens is just like continuing to plummet all the time. You know, you're saying like GPT 4.5 is expensive, but that's gonna be,, like, the cost of tokens is just, like, continuing to plummet all the time. You're saying like GPT 4.5 is expensive, but that's going to be a tenth of that in probably six months or a year, something like that. So it's just like, yeah, it's pretty wild, just the improvement of diversity, I think. Yeah. I mean, I think that, like, overall, my impression when I saw all this come out and people were talking about, like, oh, like, the reduction in tokens and, like, the impact this is going to have on GPUs and NVIDIA and stuff like that, like lowering the cost of tokens doesn't mean there's going to be less use of GPUs. It actually means the opposite. It's just like when we lowered the cost of transistors, it's not like there was
Starting point is 00:29:19 less computing and that there was a whole lot more computing. But so this is like net good, I think for everyone. And also going back to your point at the beginning, Eugene, about competition is good. It's forcing companies to essentially compete. And that's why you have new launches on December 31. Or when people are launching models on weekends, they're not waiting for the right time.
Starting point is 00:29:44 It's such an arms race essentially between all these different companies. And I think that's good for the can ultimately good for the consumer or good for the developers are building on some of these technologies like there's a there's an open source project out of Stanford, I forget the PhD students name, but it's called s one. And you can you can download it, you can try it out. And like they did basically applied the same approach
Starting point is 00:30:08 of this distillation using open AI originally using their reasoning tracing and then open AI shut that down. But then they use DeepSeq as well to do the training and they get a similar performance and the final training run costs $20 to run on GPUs. So you can recreate this whole thing yourself if you want, if you download their source code and run through it.
Starting point is 00:30:32 So that's great for, I think, creating this distributed environment. And I also think that a lot of the model companies are, because there's so much competition and the switching costs from model to model is not that expensive. You're literally mostly just changing like a string value typically, then they're I think they're trying to find other ways of having sort of stickiness in their platforms
Starting point is 00:30:58 and adding value that goes above just like, hey, here's a model that you can use. You know, they're going sort of up the stack, they're adding more like agentic capabilities, reasoning capabilities, data integrations, all this kind of stuff. So, you know, it's definitely an interesting time. Or even just like focusing on certain areas like Cloud 3.7, they were saying like,
Starting point is 00:31:19 hey, we realize everyone is using our thing for code. We're gonna train them, like make it much better at just coding and like worry less about just like being a general purpose thing, which like it's good for general purpose stuff, but it's like amazing for coding, you know? So yeah, I think they're trying to find their niches and stuff like that. I think quad, you know, anthropic really prides itself, I think on, you know, safety is one, you know, key thing. And then also being really good at coding problems. And I think like real world reasoning, like they, I listened to their CEO talk recently in a podcast, and
Starting point is 00:31:52 they were, he was saying how, you know, a lot of times when models come out, they're sort of optimizing for like the the wow benchmarks of like, oh, this, you know, model can pass the LSAT or something. But like, that is very different than what a lawyer actually does in the real world. Just coding problems in Olympiad are very different than the engineering type, like the real world engineering that you're doing. So Anthropic really tries to focus on, how do you do this in the real world? Yeah. What do each of you use day to day or which models or products or things are are using in your in your day to day?
Starting point is 00:32:28 So I mostly like when I when I do like LM like work, I mostly just use GPT-4 it's pretty cheap and I have my open AI key saved and I Just I don't even want to bother switching actually, you you know what? I do play around with GitHub models. So GitHub has this thing called GitHub models, which I actually only found out about because I'm a LinkedIn instructor. And they're like, hey, so we're releasing our own like set of ways to access LLMs.
Starting point is 00:32:58 And we want our instructors to like test them out. And so I'm playing around with that. They have, they have. They have OpenAI, they got DeepSeq, and they have another one. I forget what the other one is. But interesting. So it's routing through a unified GitHub interface, but it's going to the different provider models that you specify?
Starting point is 00:33:21 I think it's Azure underneath. I think it's GitHub, and then they just put all these models on Azure. Yeah, OK. Like, self-subsidiary, right? Yeah, yeah, for sure. Okay. And so that's like, you're talking about that's when you're using those when you're like building something into a product or like you're using those as just like chat type stuff or like what sort of use is that for? So I don't really chat with elements. I don't really use them for that, no. Oh man, I use that all the time.
Starting point is 00:33:50 I use it so much, yeah. I'll use it for things like, hey, like here's this copy that I'm writing, can you help me edit it? I'll use it for things like, hey, I need to organize this information, can you help me organize it? Or I'll use it for like, oh, I need to organize this information. Can you help me organize it? Or I'll use it for like, oh, so for example,
Starting point is 00:34:08 I took like the attendee list for Seattle Startup Summit and I fed it into ChatGBT and I was like, hey, can you give me a breakdown of like the demographics of this? Like who's, you know, what, are they engineering? Are they products? Are they marketing? Are they tech leaders?
Starting point is 00:34:24 Are they whatever? So it's, are they marketing, are they tech leaders, are they whatever. So it's really useful for things like that, but I don't like chat with it. And then the other things I play around with a lot is I build a lot of tutorials. I build a lot of tutorials because I teach courses and stuff, and so I just mess around and put different ones on there. Yep, yep. And are you using like cursor or anything like that? Or are you just like, are you using anything to help you generate the tutorials or you write it all by hand? No, I do, I write them all by hand.
Starting point is 00:34:52 Oh my goodness, man. What are you doing? Get you some cursor when you're like, you're doing all by hand. I don't know if cursor can do this. It might be able to do this now, but like when I tried it, and now it's been a year actually, wow.
Starting point is 00:35:03 When I tried it last year, it wasn't really able to generate like Lama index or like main chain or like, you know, codes from like more recent frameworks, which you might be able to do now. I'm not really sure I haven't tried it recently and I probably should. I would give it a try.
Starting point is 00:35:18 I would say get cursor and get clot three seven and like let it run. I bet like, I'm pretty surprised at like what it at what it can do. Last week I was out of town. I went to Disney with my family and I heard Cloud37 came out and I was just like, oh man, I want to try this, but I told my wife I would not work on this trip and all that stuff. I just got back this week and I'm like, wow, it's pretty wild. Yeah, it's good. Yeah, it's good. I would recommend. Yeah, I use primarily for my consumer-facing application side, I use primarily ChatGPT.
Starting point is 00:35:55 I'll use Gemini a little bit as well. And I started playing around with Notebook LM2 from Google for doing research. But for programming, I primarily use GPT-4 and Quad, depending on what the project is. And sometimes I'll use the earlier models of GPT-2 as well, like 3.5 and stuff like that. If the thing that I'm doing doesn't really, I just need to do the quality doesn't matter that much.
Starting point is 00:36:26 Then I'll do that just to reduce the token costs and stuff like that. But primarily, Claude and GPT-4 are my go-to models. Gotcha. Do you use any like Cursor or even GitHub Copilot or anything like that? No, I use, I do want to experiment with Cursor, but primarily, I do use OMS a lot for helping with coding, but I'm
Starting point is 00:36:48 mostly using like chat GPT for a lot of stuff right now. Yeah, yeah. What about have you any of you have either of you used v zero from Versailles at all? I haven't like, but I think, you know, there. I mean, I do a lot of next JS for like the front ends of things I build, but most of the stuff I'm building is from like a UI perspective, is really just like to serve a purpose of to go do something more interesting on the back end. So like, I, you know, I wanted to look half decent, but I don't need to like, I'm not building like full on, you know, end to end front end applications and stuff like that
Starting point is 00:37:24 for the most part. Yeah, that's true. That's true. I'm like, I'm like mostly a backend dev, but I have like a project right now where I'm doing like full stack, including some front end stuff. And it's like, I'm just, I can, I can write the front end stuff, but I cannot design it. So even just to like, I just like,
Starting point is 00:37:37 I'd ideate whatever with, with V zero and just like get some, some stuff. And I know it's just like sort of, you know, it's not super unique, but it's like, it looks good by the standards of what I could do anyway. So I like to get that just like a first cut. Yeah, even I fell in even like, techie Petit, if you're like, hey, I'm writing in next.js, I need a form to do this. And then you're like, make this look more professional, like this
Starting point is 00:38:02 website or something like that. It you know, it does a pretty good job of kind of giving you out of the box styles and stuff like that for these one-off projects. I think if you're doing something more professional, then you probably want a more integrated experience that can do a deeper analysis of what's actually within your code base and stuff. I think one of the interesting areas that's emerging now. Obviously, all these coding assistant, I think when it comes to AI, the initial frontier was, how do you do supercharged code completion?
Starting point is 00:38:33 And then it went from there to, how do you go from a prompt to actually being able to generate a more comprehensive part of an application or some fully formed application? And now there's a lot of going on as well with just like automating PR reviews and stuff like that, which I think is interesting.
Starting point is 00:38:53 There's a bunch of companies that I think are finding a lot of success. And that seems to like the next frontier of optimization. I know CodeRabbit is doing that. Who else is doing that? Let me get back to you on that. Okay. I've only seen CodeRabbit because... Oh, SourceForge, sorry. SourceForge is.
Starting point is 00:39:12 SourceForge. Oh, but they've been around for a while, right? Yeah, so they just released an agent for doing this. I talked to them recently, but it looks really interesting what they're doing. Okay, cool. I think one of the things about, like, with, I think one of the reasons, you know, there's been a lot of success when it comes to engineering is because, you know, one of the challenges with any of these, you know, LLMs is, you know, around reliability, but, and how do you build like a good eval set for like testing the application? But there's so much example, so many examples exist as well as when you're coding,
Starting point is 00:39:51 you have additional checks and balances like I can run it, I can compile it, or I can run it through unit tests. If it's a PR, there's tons and tons of example PRs that are out there that can become sort of my grounded source of truth that I can test against. And that really really really helps Whereas if you're doing sort of just like super general purpose stuff Like what is how do you build an eval test set for that? Like how do you test it and measure the reliability of is really really hard Have you talked to the folks that arise or or common about this? I have it. Okay. Okay. Let me let me let me get you
Starting point is 00:40:23 Taxed it. I'm sure they have some interesting things to say about this, because that's their whole business. Wait, sorry. Their whole business is evals? Or it's? OK, interesting. OK. I have talked to BrainTrust and some other folks in the space. Yeah.
Starting point is 00:40:39 So I mean, I think this is also another area where a lot of companies, but this is the emerging area too, is how do you test these things? What's the framework look like? I think a lot of companies end up building sort of hand curating their own eval test sets to start and stuff. But now there's companies that are focused on this.
Starting point is 00:40:56 But for general purpose stuff, it's a challenge just because the range is so wide. I mean, that's like the advantage of the sort of predictive ML or purpose-built models is the, you kind of knew what the inputs and outputs should be. So you can build a pretty good test set to measure, you know, precision and recall. It's hard to do that with like a general purpose research
Starting point is 00:41:15 tool or something like that. Yeah. Yeah. Yeah, for sure. But I agree with you, like the rapid feedback loop in programming makes it so much easier to like use this stuff where I was talking I was talking to a lawyer the other day I'm like how much do you use this he's like well we tried it but you just have to like you have to review it so closely anyway it's like I might as well just do it myself and in a lot of cases because there's not that like you know did it
Starting point is 00:41:36 compile or or did it pass tests or does it look right on the page or anything like that you have to like really look through every single word a lot to a lot more yeah especially something where the consequences of being wrong. I mean, the consequences of being wrong in coding is high potentially as well, but you have those additional checks and balances. You have automated tests, well, hopefully, and you can run a compilation. You can make sure that it compiles at least. Presumably, there's a review process and stuff like that. So there's like multiple stages that it's gonna go through.
Starting point is 00:42:08 Not every profession has that. They're basically banking on this like one person who's responsible for it to not only craft it, but make sure that is correct. Yeah. Yeah. I think it's really interesting that like, you know, you give a bunch of software engineers, this AI that can do a bunch of things.
Starting point is 00:42:24 And the first thing they do is, oh, how do I make it so that I have to do less work? I'm gonna jump in there now. Yeah. One of the things I've been thinking about is, over time with any industry, and this has certainly has happened with engineering is, because there's so much to know at this point,
Starting point is 00:42:44 there's a lot of like specialization. You know, as you go more and more in the industry, you might be like, hey, like I am not only a front end engineer, but I'm a front end engineer that has deep expertise in this particular framework. Or, you know, I work on this part of the back end and I really know, you know, these technologies deeply and stuff like that. But I kind of wonder is, if you could have a tool, like an AI tool that can essentially behave like the expert, and I really just need to know more of a high level sort of architectural standpoint and work at a bit of a more abstract level,
Starting point is 00:43:19 does that change the way this kind of specialization does? Suddenly having these tools at your disposal, it pays off to be sort of more a million miles wide and an inch deep to be able to give enough sort of instructions to these tools to get what you need. But you don't necessarily need the deep expertise of like, OK, well, I know that this is happening at the thread level when I'm calling this particular function and stuff like that.
Starting point is 00:43:44 I think that's a really good application, actually. this particular function and stuff like that. I think that's a really good application actually. I hadn't even thought about that. This is what keeps me up at night. I definitely think so. Like I've seen that myself over the last year. Like I have taken on like a few full-step projects and I'm like not a great front end developer but it's like I can make it work.
Starting point is 00:43:59 I did a mobile app. I'm working on a mobile app right now. I've never done mobile work before, you know, last year. And it's just like, yeah, I think if you have a good base of skills, I think you can apply it just to like way more areas than you could before. And you're not gonna be as good as like the people that are true experts and craftsmen in that particular area,
Starting point is 00:44:18 but like a lot of apps don't need that, you know? So yeah, I think you can like, absolutely. I think you can branch out, absolutely, I think you can branch out into a lot more things. It's also like good for like the, like, I guess, like the software engineer market in general, like, because I think one of the things that I at least I remember seeing this, like five years ago, is people are like, oh, like, you know, what if you get really into a framework, and then it becomes like obsolete, you know, becomes obsolete, then you're just screwed. But now this idea of being able to be like,
Starting point is 00:44:49 hey, I understand the ideas, the applications, how to apply these things in software. I'm just not an expert in this, but I can just ask the LLM. That's actually really helpful and I think that's better. I think that's better as a software engineering experience. Well, this is partially biased in my own experiences.
Starting point is 00:45:08 But in software engineering, people who are software engineers, who become good software engineers, are people who really enjoy learning about the high-level architectural stuff and how do you put all these pieces together, and they have a lot of curiosity. And so like you were saying, Sean, it's nearly impossible to learn everything.
Starting point is 00:45:28 And so I think this is nicer for the people, that persona of people who enjoy the software engineering archetype, but don't necessarily wanna get super deep into things just because it's literally impossible time-wise. So I think this is a really cool idea. I haven't seen anything that particularly does this, but I guess you can just go to Chatuchipati,
Starting point is 00:45:48 you can be like, tell me about the high level architecture of the COBOL, and then you can go and do that if you want. Well, yeah, and also, all the TypeScript professionals can now maintain those COBOL mainframes because they can be like write it in TypeScript and translate it to Cobalt for me. Yeah. That sounds terrifying. Yeah. I want my bank to just stay on Cobalt for now. Yeah.
Starting point is 00:46:14 Yeah. I don't think we're quite there yet, but I just thinking sort of bigger picture, like what is the future impact of some of these technologies on various industries. And I think the one industry that's leading the charge is to the tip of the spear is I think engineering in terms of what impact this is having on people's, the way they do their job is gonna impact the way that people probably think about hiring as well as like what skill set do you now test for? If I can give someone a problem
Starting point is 00:46:44 and they can literally just type it into a prompt during an interview and get like a decent, you know, decent output, like, you know, you can't obviously can't test that way during an interview. I don't think those are great tests anyway, but that has historically been a way a lot of companies have hired for a very long time. Have you seen like the recent thing
Starting point is 00:47:03 about like the Columbia student who like, he he like he took like his live stream of this video he just like got this like I don't know like this like transparent looking thing that like sits over his like coding interview and he would just use that and he used it in a coding interview and he passes like Amazon, Google, XYZ, ABC like interviews and he just like was like I'm not going to take in these offers and he has like a 30 he's like he's like 30,000 MR and like literally like, I don't know, like a month
Starting point is 00:47:28 or something like that. I was like, what? What the heck? That's that's amazing. I didn't I didn't see that one. That's that's funny. Yeah. I mean, I think it was similar to if you're in your teacher, and you're used to the the thing that you're giving a student is a book report well a Book reports not a great test of someone's You know comprehension of a book at this point because you can literally generate a book report in a single prompt and have it So you need to figure out new ways of essentially, you know testing skills
Starting point is 00:48:02 And I think it also evolves like evolves what are the core skills that you need to be successful in a job. It's changed a lot. There was a time where writing cursive was maybe a critical skill, or knowing the Dewey Decimal system, or these types of things as well. And those skills haven't completely gone away, but you need to essentially develop
Starting point is 00:48:26 new types of skills that are relevant to the workspace. I've never written a book report in my life. Wow. I've never read a book for a book report in my life. I grew up in the age where I had the internet, and so I was just, you see internet. I really want my kid's school to have like a, an LLM class where it's like,
Starting point is 00:48:50 I don't know how to explain it, but basically I want the teacher to give them like kind of wacky assignments, like things that they don't have any context for and just be like, you know, and buy, get them a chat GBT subscription or something like that, and just like go do something and tell them,
Starting point is 00:49:04 like spend two hours on this and give me something at the end and also send me the chat that you did for it. And like sort of like get them used to like using these things to like make like, basically I wanna say like, here's an Excel spreadsheet and I want you to find out these like 10 answers from me. And like none of the kids have probably used Excel or something like that,
Starting point is 00:49:22 but like they could go to the chat GBT and be like, hey, I need to figure out. I have this spreadsheet. Like, what do you need from me to figure this stuff out? I think it'd be cool. Yeah. Yeah. And I think that's that's that's really interesting. Like that test, you know, someone's ability to like learn and use these tools as a learning tool, which is is probably a more important skill set now. Because I also think like even as amazing as some of these coding
Starting point is 00:49:47 tools are or other types of tools that are out there, you still need enough sort of technical context to direct them in a way that's going to give you a good result. I don't think they're at a place where you could give this to the hands of my mom and my mom would be able to build a mobile app like you're doing right now. Exactly, yeah. That's totally true. But in terms of just so many things that you should be able to just,
Starting point is 00:50:14 you might not think you could do it, but I think my 12-year-old daughter should be able to get answers from an Excel spreadsheet, even though she's never used Excel before, with an hour of chat to be honest. I think that's totally doable. And I think the big thing is I want them to learn how to use these LLMs and also just train their agency a little bit of just
Starting point is 00:50:32 if you don't know how to do something, you have something that can really help you figure it out, at least to a medium level. It's not going to make you expert on any of this stuff. Yeah, absolutely. Some decent results from that sort of stuff. Cool. All right, well, so the last thing I wanted to chat about
Starting point is 00:50:47 was there's also been a ton of recent acquisitions that have happened. Voyage AI was acquired by MongoDB, data stacks by IBM, weights and biases by CoreWeave. I'm sure there's a ton of other ones that are missing here. And given, Eugene, your background working for Zillow, I'm kind of curious, what are your thoughts on the acquisition by Voyage AI? It seems like a kind of trend where a lot of databases are going to continue to increase value for their platforms. It's like, go beyond just a database to building sort of like a data platform.
Starting point is 00:51:20 So we're going to offer from an AI perspective, we're going to have embedding models built directly into the platform, we're going to have, you know, chunking all these other things that, you know, if you weren't using them, you'd have to go to some other third party service or write yourself book to get that stuff ready before you sort of ingested in the database. Yeah. Okay, so this is what I was gonna say earlier. So, Frank, Frank Liu, Frank Liu and I worked together at Azales
Starting point is 00:51:46 and he went to Voyage.ai last year in September. And I started hearing rumors, not from Frank, from other people. I started hearing rumors that Voyage.ai was going to get acquired in like December, January, something like that. And I was talking to one of my friends about this and I was like, so did you hear about this VoidJI thing?
Starting point is 00:52:06 And they were like, yeah, I heard VoidJI was getting acquired. But at this point, I'm thinking like, if you're one of these companies that raised in 22, that raised in maybe not 22, maybe 23, but like during that like kind of, no, 22. Yeah, at the peak, like there were still many companies that raised $100 million at 100 million dollars at two billion dollar
Starting point is 00:52:26 valuation with zero revenue. Yeah, yeah, yeah, exactly. And my friend was saying he was like, hey, like, you know, like, I'm thinking if you're one of these companies, you're either an acquisition target or you're dead. And I was like, yeah, I mean, that's that's fair enough. And then I saw Frank in. I don't know if it's early February, January,
Starting point is 00:52:46 something like that. We went out, we went out for drinks. And he was like, he was like, yeah, I've heard those rumors too. I was like, oh, okay. But like, like, like, Zillis, for example, like, you know, the CEO of Zillis is he's, he doesn't want to do an acquisition like he's like, he's like IPO or bust. But Tang you the guy who does voice. Yeah, I obviously is open to an acquisition and they got acquired. And I remember hearing about this like the week before it happened. He was like, so you know how like we're talking about like coming on board to like sponsor a Seattle Star Wars Summit and your AI production meeting conference. And I was like, yeah. And he was like, well, we wanted to do it, but we're actually doing something else right now.
Starting point is 00:53:28 And you'll see an announcement on Monday. And I was like, OK, OK. And then he actually told me literally the Sunday or the Saturday before. So I was like, I knew a couple of days before. And he was like, just don't tell anyone. I was like, all right, I'm going to keep my mouth shut. But now that it's happened, I'm like,
Starting point is 00:53:40 I'm sure I can talk about it now. But yeah, he was like, yeah, MongoDB, good acquisition. Good for them in terms of I heard they got very good valuation out of it. I, I personal opinion, in vector database space, MongoDB has a lot of work to do. has a lot of work to do. Yeah. Yeah, I mean, I think that, similarly to the model companies that we were talking about earlier, where they're looking for ways to increase their value beyond just being a foundation model, I think databases
Starting point is 00:54:16 are doing some more moves. How do we move up the value chain and own more of the pipeline natively? You see the same Databricks, Snowflake have continued to essentially layer on more and more functionality. They're way beyond just like, hey, there's a data store too, like this entire platform where you can, you know, do all kinds of different things there. I think replacing a lot of the ETL layer for traditional, you know, data infrastructure, as well as trying to own a lot of the AI space of,
Starting point is 00:54:48 you come to us and do everything essentially. Yeah. Also, DataStacks literally just recently signed on to come on board as a sponsor for Seattle Star of Summit. I saw they got bought by IBM and I was like, that deals dead in the water then, that's gone. But then I got an email and they were like, yeah, we're in. And I was like, oh, okay, great, let's do it. I think the HashiCorp IBM deal is official now too as well.
Starting point is 00:55:12 Oh yeah, yeah, that was huge, dude. That was like 7 billion. Did they announce how big the data stacks one was? I don't think I saw it, but maybe. Data stacks has been around for a long time. I think they're over a decade old at this point. Yeah, Astro. Yeah, Astro DB. Yeah, Astro. Data stacks also
Starting point is 00:55:33 bought like some company like in like the last like two or three years. Yeah, forgot which one it was. But guy Ben, Ben chambers was one like the CTO of that company. I can't remember what it was. So yeah, the CTO of that company. I can't remember what the company was though. Yeah. Data stacks also, vector database functionality. Yeah. Yeah. So IBM is- They're making a lot of moves.
Starting point is 00:55:53 It's very interesting whenever I think about IBM, because it's just like big giant old company. It's like, what is this company doing? I think our generation doesn't think about them. We'd never think they'd buy something from IBM, but yet they're buying all these other companies. Well, I mean, they're such a huge conglomerate. They own a red hat.
Starting point is 00:56:17 I mean, there's so much stuff that's under the umbrella of IBM at this point. They're called the international Machine, you know? That's what they are. That's like information superhighway, but from like 1920. What a name. Yeah, that's amazing. And then what are your thoughts on weights and biases, CoreWeave?
Starting point is 00:56:42 CoreWeave, I'm not super familiar with, but you know, they're an AI infrastructure company, you know, focused on sort of the GPU layer in the cloud. Yeah, that one really surprised me. I thought, like, okay, so I'm actually not super familiar with either of these companies, but I thought Weights and Biases has been around for longer. They have. They've been around for over a decade, I think, because they were originally traditional sort of ML ops. I interviewed one of the founders of Weights and Biases at one point.
Starting point is 00:57:10 So they, I mean, they pre-date everything going on in general AI and then of course they added a lot of functionality on top of that, but their original core customer was to serve sort of data scientists that were building models. How do you do versioning? How do you set up your training runs, all that kind of stuff? And I always, from an outsider's perspective,
Starting point is 00:57:33 felt like they had a very good business, but they, obviously they see some sort of opportunity with CoreWeave. Yeah, that one really surprised me because CoreWeave's a little bit newer. It's not that big. I mean, it is pretty big. They've they've raised a lot of money. Well, I think to do any cloud GPU infrastructure company, you have to raise a lot of money.
Starting point is 00:57:58 Yeah, you got to raise a ton of money. But yeah, I was I was really surprised. I just I don't know. But but it says they spent 1.7 billion on whites and biases. Like how much did they raise they have 1.7 billion? DC money. Yeah Do they have this investors they have they might have similar investors It might have been like a deal that some of the investors like helped like put together. Yeah, what are your thoughts on? these kind of like specialized cloud infrastructure for AI workloads versus the public cloud? Like, you know, I can do you think that because there's more and more of these companies that are emerging.
Starting point is 00:58:36 Do you think that we're really going to have a dozen different places where you're going to go run your AI workloads independent of where you might be running your application stack today? Yeah, so I've thought a lot about this because I talked a lot with the GMI Cloud people. And I also talked a lot with this company called Amara, which says data centers. So I think the thing between these new Neo Clouds or AI-native GPU, whatever, is that unlike the CPU marketplaces, like AWS,
Starting point is 00:59:08 for example, they're basically just running their own data centers. AWS has these CPUs. They basically just have a data center where you walk in and you have these racks, and these racks just full of GPUs. And there's different cooling that you need for that. There's different like kilowatt edge that you need for that. There's like different, like, um, there's different requirements that you need to run one of those data centers. So I think it's like, it's like the, the physical requirements are different enough from like your classic cloud compute stuff that they will exist. I don't think we're gonna have, I think there's currently 45 of these.
Starting point is 00:59:45 I don't think we're gonna have 45. We're gonna have some. I think the two kind of like, the two plays that I see that are gonna be most likely to be sticking around for a while are the ones that own their own data warehouses and the ones that are basically like a marketplace for GPUs. I think those are the two kinds of companies that these are,
Starting point is 01:00:08 I kind of like, I see these as the two kind of extremes. And I think those are the two that are going to stay around. I think we're going to have maybe like three or four of each. Interesting. Are most of those specialized data centers, this is my ignorance on it, are they mostly used for training or are people doing inference from these specialized places as well?
Starting point is 01:00:24 I think, I think they're doing both right now. My intuition on this is that the training is actually what's making them a ton of reliable revenue. Then the inference is actually what's mostly like it's probably used more, but it's not as reliable. I don't know how that's going to play out but it's not as reliable. And I don't know how that's gonna play out in terms of the business model. It's probably just some sort of uncertainty
Starting point is 01:00:50 they're gonna account for. But that's one of those things that I think is not very clear. And there might even be companies that are particularly... So for example, GMI Cloud, they provide training. That's their current model. But they're building an inference platform. And they want you to go and use the inference. So it's just not very clear yet.
Starting point is 01:01:14 You think the TAM on the total addressable market for inference is much, much bigger because most companies are going to be doing it. But then you have players like Base 10, Together AI, and others that are focused on inference and optimization. We've had Base 10 on, Together AI is coming on soon. Those companies are interested, but they're focused,
Starting point is 01:01:36 specializing essentially in multi-tenant inference and just optimizing the heck out of it. Within public Cloud? Those are more within public cloud, for those ones, those are more within public cloud, right? I don't know. Is they? Like rather than build their own data center, I imagine? I'm not 100% sure.
Starting point is 01:01:54 I think they probably work with other providers, because I know that Base 10 and GMI Cloud are doing a dinner on Wednesday for GCC in like two weeks. So, okay. Yeah. It's definitely interesting times. Anything else that other you wanted to chat through? I think that's all I have.
Starting point is 01:02:14 Yeah. Yeah, that's all I got. All right. Awesome. Well, Yujin, thanks for coming back. Excited about the, you know, what's going about everything you got going on in Seattle. Hopefully, you're able to accomplish your goal of turning Seattle into the startup hub. For those that are listening,
Starting point is 01:02:33 check out the Seattle Startup Summit at the end of March. Now, it's good to see you as always. Yeah, for sure. I'm going to check out the recording, so make sure you put those on for me later on. Even if it's not live streamed, I'll catch those recordings later. Yeah, 100 percent. I'm gonna check out the recording. So make sure you put those on for me later on. Even you know, even if it's not live streamed, I'll catch up. I'll catch those recordings later. Yeah, 100 percent. I'll get them. Yeah. Thanks, Sean. Thanks, Alex.
Starting point is 01:02:51 Yeah, thanks for coming on. All right. Cheers.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.