LINUX Unplugged - 551: AI Under Your Control

Episode Date: February 26, 2024

Corporate AI is a hot mess, but open-source alternatives can be open-ended chaos. We’ll test some of the best ways to get local AI tools under your control. ...

Transcript
Discussion (0)
Starting point is 00:00:00 Well, boys, I've been chewing on something long enough now that I think I'm ready to talk about it on the show. I've been trying to think about all this AI stuff that's been just thing after thing. And all of the discussion around safety and then the news that we're going to talk about in a little bit with Gemini. And I feel like perhaps I've stumbled onto a hard truth. I want everybody's feedback, but we're always talking about user consent and how much it matters. But I don't think we're thinking about it in terms that regular people care about. I think user consent matters and hijacking your intentions as a user is wrong. Telling me something like this isn't appropriate is hijacking my intentions.
Starting point is 00:00:40 Enforcing a particular worldview, I think that's unethical. I mean, I agree on some of the things that they're trying to protect, like racism. Those are bad. But so is cramming a for-profit company like Google, cramming their ideals and morality down my throat just feels like the worst case sci-fi dystopian future that we worried about. And we have these corporations that are setting themselves up as like arbiters of reality, of what's right and wrong, when morality should be set by the people, not corporations or governments. And the more I think about this, I think this is just going to be a hard reality because all these companies are so risk adverse.
Starting point is 00:01:19 Yeah, they're not really set up to be speaking truths. They're there to set up to be to incentiv you know, incentivize their profits and say things that won't get them in trouble. They're not necessarily doing a good job of that, but that's what they're trying to do. But like if I run it on my system and I have control over it, it's on my Linux box, those aren't my priorities.
Starting point is 00:01:37 And I think if this is a hard truth that we kind of accept, then really the future for interesting AI that's going to be pushing the limits and for good and bad, and we'll see stories around both, I think that's going to be pushing the limits and for good and bad and we'll see stories around both I think it's going to be open source I think it's going to be the open source stuff and that was really the motivation for today's episode Hello, friends, and welcome back to your weekly Linux talk show. My name is Chris. My name is Wes.
Starting point is 00:02:11 And my name is Brent. Well, hello, boys. Coming up on the show, you know what they say, if you want it done right, you got to do it yourself. So this week, we're going to see if we can tame the moving target that is the current AI stacks out there that you might want to try. We're going to see if we can tame the moving target that is the current AI stacks out there that you might want to try. We think maybe we have found one of the best ways possible to deploy these AI tools under your own control on your box. So stay tuned and find out for that. And then we're going to round the show out with some great boosts and picks and a lot more.
Starting point is 00:02:39 So before we go further, let's say good morning to our virtual lug. Time-appropriate greetings, Mumble Room. Howdy. Howdy. Room. Hello. Howdy. Say hi, Sam. Hello, Brent. Hello. Hello. Hello, and a big good morning to our friends over at Tailscale, tailscale.com slash linuxunplugged.
Starting point is 00:02:53 We're playing around with these tools. We're deploying it on our tailnet. We can all access it by name. We don't have to worry about security vulnerabilities in these brand-new projects because we never put it on the public internet. I don't really deploy anything on the public internet anymore. It is the easiest way to connect your devices and services to each other wherever they are, regardless of the platform,
Starting point is 00:03:10 and it's fast. Go try it on 100 devices and support the show at tailscale.com slash Linux Unplugged. All right, so this week, I think if we could have one overarching goal for the show, So this week, I think if we could have one overarching goal for the show, I would love to at least get the listeners thinking more seriously about open source AI tools and less about things like chat GPT and co-pilots and things like that. We're going to talk about the news that's come up this week in just a little bit. But also the tooling has come a long way for the open source stuff, just on the web and communities, the stuff you can install. And also some of it's gotten complicated and some of it's broken. And I think the other reality we're sitting with is these commercial solutions, they're so risk adverse that it's almost reducing the usefulness of the product, right? It's embarrassing, really. It worries me too, because that's just more of the, you know, it's already an imbalance because only these giant companies seem to have the
Starting point is 00:04:06 resources to sort of train the foundational models. And of course they're the ones applying all of these filters that end up harming the actual productive use. And so if you can't say anything, I mean, let's not even say offensive, but just sort of creative, weird,
Starting point is 00:04:19 out of the ordinary, then they're left with the only sort of access to these tools that, you know, they can tune down those filters internally, presumably, and then they can make the interesting stuff. That's a good point. Yeah, they're the only ones, really. They become the gatekeepers, essentially, of this. And this is why I sort of feel like we have a limited time to win the race here in just a moment that there are several factors that are working not necessarily together, but the result is they work together to limit the usefulness and utility of the open source stuff. And I want to cover all of that in a moment. But first, we have a couple of events that are just around the corner.
Starting point is 00:05:00 We are three Sundays away before we hit the road, boys. Okay. Three Sundays. I'm sure you're totally ready, Brent, right? Because, I mean. Oh away before we hit the road, boys. Okay. Three Sundays. I'm sure you're totally ready, Brent, right? Because, I mean... Oh, yeah, yeah. Always. You know me. I'm early on the spot. I'm not worried about that at all. He committed on air. Yep. Multiple times for weeks in a row. And I always end up there.
Starting point is 00:05:18 Yeah, that's true. It's true. It's just how much that I stress about it is really it. We are getting so excited about the very first NixCon North America. We can't wait to bring you coverage of that. That is co-located with Scale in Pasadena, California. It is the Linux Expo number 21 for Southern
Starting point is 00:05:34 California. It all kicks off March 14th. We're hitting the road just like a day or two before that. If you want to sign up for Scale or go to NixCon or go to them both, get a Scale pass and use our promo code JBFTW JB for the win to save 50% off and uh and then join us at lunch Saturday at 1 30 p.m at the art house we want to see you there and then of course go grab fountain.fm because they are creating a
Starting point is 00:05:57 totally upgraded live experience for us in app and on the web oh they're working on a cool web app that they've prototyped for some of the live value for value music streaming events. Okay. That can toggle between audio and video. You can use it in the app or you can use it on the web. Nice. Well,
Starting point is 00:06:14 I want to see this. Oh, it's going to be great. So we're going to have a whole batch of live streams, one on the 12th, the 14th, the 15th and the 17th. And you don't have to really worry about any of that right now,
Starting point is 00:06:24 but we're looking forward to it. And I think depending depending on how all this goes we're kind of going to be defining a new live experience for the show and for future events so i think it's going to be kind of a make it or break it trendsetter for us we're kicking off the first event bigly bigly as wes always says that's right okay well that's kind of all i had announcement wise uh you feeling good i am feeling good Did you decide you're flying back? I am. Yeah, for the work, right? For your hashtag J-Job?
Starting point is 00:06:49 Yeah, get back to the regular things. Yeah, that's right. We're going to take Brent maybe somewhere special then just to make you jealous. Dang it, I knew this would happen. I'm going to have to fly back to then visit for that part of it. What are the options? Brent, you know what I'm going to be advocating for. I don't know how we're going to make it work.
Starting point is 00:07:06 That's what you said last time. I'm going to be advocating for the coast. Oh, this is exactly what you said last time. That never happened. I know. I'm aware. But we haven't booked any of our return Airbnb-zles. Bring it on. You've got flexibility. We do. Maybe we just do like hotel or something. I don't know.
Starting point is 00:07:22 But I would love to take the coast. I think one of the prettiest places that West Coast is one of the prettiest places in the world. It's or something. I don't know. But I would love to take the coast. I think one of the prettiest places, that West Coast is one of the prettiest places in the world. It's really something. We'll take pictures though, Wes. We'll take pictures. All right. So you probably heard the news this week about Gemini just being embarrassingly bad. As the Register put it, Google halts Gemini image generation to fix a white balance.
Starting point is 00:07:46 it, Google halts Gemini image generation to fix a white balance. They go on to say Google has suspended the availability of text-to-image capabilities in its recently released Gemini multi-model foundational AI model after it failed to accurately represent white Europeans and Americans in specific historical contexts. And their sub-headline, which I thought was really on point, quote, big tech keeps poisoning the well without facing any consequences for its folly. Well, I say today here on this show, they face the consequences, gentlemen. Today, we say this far, no farther. The line shall not be crossed. We are going sovereign with our technology. And I was excited to see Stable Diffusion 3 came out. Big fan of Stable Diffusion image generation stuff. Big fan of the Stable Diffusion image generation stuff. I'm not so sure, though, the hosted version is really going to hold up.
Starting point is 00:08:30 In their announcement, they write, We believe in safe, responsible AI practices. This means we've taken and continue to take responsible steps to prevent the misuse of Stable Diffusion 3 by bad actors. Who bad actors are, though? I'm not sure. It doesn't say. I guess they define who the bad actor is. It's Brad. Oh, yeah, of course. Well, that's definitely one of them.
Starting point is 00:08:50 They continue. Safety starts when we begin training our model and continues throughout the testing, evaluation, and deployment. In preparation for this early preview, we've introduced numerous safeguards by continually collaborating with researchers, experts, and our community. We expect to innovate further with integrity as we approach this model's public release. So all of the talk right now really is leading with responsible AI practices and safety, perhaps as they should be, but it's definitely the focus. And I thought commenters on Hacker News
Starting point is 00:09:23 nailed this when they were talking about how it's basically making the products less useful. Their sensitive nature is causing the rest of us to just want to look for something that we can self-host and run because I get offended by them projecting some intentions that they manufacture onto me. Yeah. Well, and I mean you just – you're kind of left out. you're on to me yeah well and i mean you just you you're kind of left out you know it's not hasn't been a perfect system but uh by and large you kind of get to decide you know if you're going to repost something if you're going to share it if you're going to use your speech to have it and you're you're up to that and you're it's sort of implied like well you could do anything with this even though you might just want to make a silly limerick you know right there in the app and then
Starting point is 00:10:04 you'll never use it again it kind of feels to me like this is potentially going to create another divide that we've seen in, you know, computer privacy and data privacy. Like those of us who are technically sound enough to host our own stuff will have far more privacy than those who just don't have those skills. And this feels like it's headed in that exact direction. It's also alarming. The point that the reg makes is actually well taken. It's like they keep screwing it up, but because we have no other choice,
Starting point is 00:10:34 or no other viable choice, they really feel, they face no consequences for their folly. I mean, you think about something like this that had to ship, how they miss something like this in testing. And we were sharing screenshots this morning, I mean, you think about something like this that had to ship, how they miss something like this in testing. And we were sharing screenshots this morning, I was, because throughout the weekend, people are just trying various different things that you think would be easy for it to answer, and it just can't answer them. Sure seems like it is kind of like, you know, end of the pipeline. Like, you know, they're not necessarily being able to fix or address the systemic problems in their data sets.
Starting point is 00:11:05 And then at the end, they're kind of just like, oh, well, we will apply some little fix to try to tweak it to make sure it's like sufficiently diverse or whatever the, you know, like correct for other things that they want to make it generally more accurate. But then you're doing it so late in the process that you get crap results. Wouldn't it have to understand to not answer? I think it depends on what layer this is being applied. But your point taken is it must realize the truth in order to generate something that is either the opposite of the truth or to say it can't answer it because of moral reasons. Like it must have some context awareness then of the actual question underneath. But then you're right. The presentation layer, they're restricting.
Starting point is 00:11:41 than of the actual question underneath. But then you're right, the presentation layer, they're restricting. Well, I just mean like what is implied or learned from the data sets that are being selected. And then they're just trying to apply things at the end that are going to be very coarse grained and are not at the level that a giant data set has. Mm-hmm, okay.
Starting point is 00:12:00 And I think at the same time, as we have this sort of ridiculous, over-safe, we had an example on Coder Radio. Somebody wanted to know the absolute fastest way to do something in C-sharp, and the LLM would not respond because it was too risky and dangerous to do something just for speed and not for, you know, good code quality. Couldn't tell you the answer. And that's where it feels like it's another level. Like, I think probably a lot of us, especially for hosted for hosted services like i expect there to be some level of this you know like if i'm just like generating obvious hate speech i could see and it probably being reasonable for
Starting point is 00:12:33 the general public of something that's like um we're not cool with this from our product for you like and if it's an edge case it's fine it's like a reasonable limitation but right now it feels like every other thing that you ask these these machines do, you get back a little slap in the face. I mean, sometimes it's even like a parody of a brand name or something. It's like, oh, you know, we can't really let you speak. Hello, the Burger King. Yeah. Or like, you know, public figures, which an artist could whip up and it would be totally free speech. It's an important part of our public discourse.
Starting point is 00:13:05 Yeah, right. And then at the same time, you have a group out there that Politico has labeled as AI doomsayers that are funded by billionaires that are ramping up lobbying right now as we record. So two different nonprofits, the Center for AI Policy and the Center for AI Safety, funded by different tech billionaires, have begun directly lobbying Washington to address what they perceive as existential threats posed by AI. So each nonprofit has spent around $100,000, north of $100,000 on lobbying efforts, both of them. Specifically, they've spent that money in the last three months. And they're drawing their funding from different organizations throughout the industry that are tied to AI.
Starting point is 00:13:48 And the part that I love is, like Brent's good buddy, Sam Bankman-Fraud, both nonprofits have ties to the effect of altruism cause. Which has been an absolute, I mean, I just think it's been absolutely disgraced with folks like Sam Altman and others. I mean, I just think it's been absolutely disgraced with folks like Sam Altman and others. And these folks all have what, in my opinion, is a God complex where they believe they have to save humanity. So they create problems and then they panic about them and pretend like they're the only ones that can save us from them. And that's how they get off as rich billionaires is saving the world from these existential crises that they're very funding created in the first place. Previously, efforts to influence Washington on AI were conducted through going directly at staffers and think tanks. That was really the approach they took before. And we got some executive orders here and we got some action from a couple of senators, but not much.
Starting point is 00:14:40 Nothing really formed any kind of law. But so now they're going with the full on lobbying effort. Nothing really formed any kind of law. But so now they're going with the full-on lobbying effort. And essentially what they're pushing for is in order to work on AI projects, you need to be licensed. And you need to be essentially re-licensed every couple of years to make sure you're not creating dangerous AI. And then the projects you're working on need to be checked and audited by a group, an industry group. And they even want to have like restrictions around like the power and size of the models and whatnot is what they're advocating for.
Starting point is 00:15:09 And they want this implemented at a law level from Washington, D.C. here in the States. You know, I'm not against like any kind of regulation for this kind of stuff. I'm sure at some level, especially as the scale picks up, we'll need some. But boy, do I like it does not seem like a we're long enough into it to really know what the effective regulation should be and then be just the i don't know that our current system is really set up to like it's all going to change a lot so like whatever we need to do the qb get something reasonable and then something nimble that you would actually update as these things develop and that i have not a lot of faith in and something that didn't require
Starting point is 00:15:42 millions of dollars of lawyers to nursemaid through the system, which it would seem these systems do require that, which almost means you're just. Can we not lock in the incumbents? Right. You're just locking in the incumbents. Right. And you're making it particularly difficult to use open source AI tools in business and in business development and software development. You know, if you want to run a little model on your little GPU, that's fine. But don't go building something that's going to be running as part of a software as a service. It's the sort of same thinking that reminds me, I don't know, the folks who get prosecuted for hacking when they download a file that was in a public S3 share.
Starting point is 00:16:19 Yeah. Yeah, it's information. And I think that's why I want to advocate to our listener base that even if you're not currently wrapping your heads around this, just stay informed a little bit. We'll give you some resources. I'm actively soliciting more from you out there in the audience because I feel like we have kind of a race to win here to a degree to get to a certain user-based size where they can't just completely ignore it at a legislation level. user-based size where they can't just completely ignore it at a at a legislation level if we got to a certain size adoption before it goes too far then it's like the genie's out of the bottle and so i feel like one of the things this show could do to help contribute to that is cover the tooling that is accessible to all of us and help you deploy in ways that are consistent
Starting point is 00:16:59 um that maybe are nimble like west, because the thing is changing really fast. So I think, you know, that's what we focus on next. Warp.dev slash Linux dash terminal. Yeah, that's right. Warp, the modern Rust-based terminal with AI built in, is now available for Linux. It used to only be out there for Mac OS users, but it's available for Linux on multiple distributions,
Starting point is 00:17:29 including Ubuntu, Fedora, Debian, and Arch. You can give it a try today at warp.dev slash linux-terminal. It is something that feels like they rethought the terminal for the modern era. If you work in the corporate world where there's Mac and Windows and Linux systems, you're going to feel like you have a tool that just keeps up with the rest. This is really slick because, A, it's built on Rust, so you know it's quick. But it has a really nice modern text editor built in. So when you're editing your YAML, your Docker Compose, your JSON, whatever it is,
Starting point is 00:18:02 you know, it's got a modern text editor built in. And it lets you navigate through blocks of output really quickly. If you forget a command, there's that warp AI that you can invoke that will suggest the right command for you. This is what I've been playing with and it is really handy. There's also just the ability to customize the UI. Yeah, I know you like to do that. You can create your own prompts too and have all the nice things set ready to go so you can recall those when you need them. It's a great user experience.
Starting point is 00:18:28 They have a collaboration feature in there. All of it is fantastic for developers or engineers who are working or living in that terminal. I've sometimes looked over at some of the other terminals on the commercial platforms and I've been a little jealous. But now, no, no, not anymore. Go check it out. Warp.dev slash Linux dash terminal. Now available for Linux, including Ubuntu, Fedora, Debian, Arch, Linux. And I bet a lot more coming soon.
Starting point is 00:18:56 You know how those things work out. Go check it out. Give it a shot. Support the show by going to warp.dev slash Linux dash terminal and give it a try today. warp.dev slash linux dash terminal. It's not all bad news, though. I mean, at the same time that these big corporations are kind of, you know, cutting their foot in public, they are also releasing free and open-weight models that you can go play with, like Google's new Gemma.
Starting point is 00:19:26 I mean, this is pretty nice to see it this soon after Gemini comes out and it's a two billion parameter model. Wow. Two sizes, the two billion and seven billion parameters available. Now, they do apply like all their safety stuff to it. So, you know, it's that's these different models you can grab. We've talked about this before. They all whatever institution makes them. It comes with with that institutional bias there is that aspect to it but wes why are they doing this why are they taking like these these models that are kind of behind all of this work they've done and releasing to the public is it is it like the like the github
Starting point is 00:19:58 idea where you make it free so developers kind of get hooked is it i mean i guess i don't really understand the motivation for all of these models coming out, especially when Google's trying to get people to pay for an outrageous amount for Gemini. Yeah, I mean, I suppose on one end it is to help mindshare adoption. I think because machine learning has been the area of academic interest for a long time and then has in the last few years sort of blossomed now to being really an applied science and industry um there is sort of a reputational aspect of publishing what you're doing like whether that be a paper and then a lot of times now it's a paper and the model behind it and maybe they just feel comfortable that there's enough you know special sauce that they can sort
Starting point is 00:20:38 of keep on their own plus uh it still takes resources and knowledge to sort of get this you know you and i can get this to work and tie it into stuff. But it's a whole other level to build into a nice service. And I hadn't thought about the historical kind of momentum around releasing this stuff too as sort of educational background. And speaking of that, Olamah, which we've talked about before on the show, has already incorporated this. So Gemini's Gemma comes out and then within a day or so, the Olamo folks are already incorporating it. I mean, the free stuff is moving really fast. Yeah, Olamo seems to be doing pretty nicely.
Starting point is 00:21:11 And you see they have a REST API. They've got, I think, chat GBT style API compatibility. There's also some bindings like Python bindings, probably JavaScript bindings as well. So it's getting easier and easier to adopt and build into your apps if you so want to. So I thought we should focus on image generation this episode. There's a lot of different tooling out there, but all of the hoopla around Gemini has been with image generation. And we've talked about easy stable diffusion before. It's a Docker container that you can get up and running pretty quickly. This week, though, I want to talk about Invoke AI. Invoke AI
Starting point is 00:21:45 is an implementation of stable diffusion that has, along with it, a great set of tooling and a streamlined process to generate images, pull in updated models or modifiers, and work to generate images, whatever you want, really, depending on the model you're using. And we've been experimenting with it on and off for the last few days. We be talking about that but the reality is there's a lot of ways to install something like this and i've noticed a lot of bad habits invoke ai in particular is guilty of this a lot of like just run this script and it just dumps stuff all over your system and running stuff and i mean i think also some of the, you know, just the way these, the history, maybe some of the coming out of academia,
Starting point is 00:22:31 there's a lot of specific stuff. There's different package managers. There's a lot of specific dependencies that you need for your particular hardware. Sometimes like sensitivity to which version is going to work and, you know, new ones won't or way older ones won't. You need this set over here matched with this one over here. Plus you're saying there's kind of a lot of stateful stuff where you're you're running scripts to get everything just so set up and then it's kind of based on
Starting point is 00:22:52 whoever happened to tie all those things together and then some of it's kind of brittle because it's dependent on certain like video acceleration libraries working and being there installed correctly and you end up with a bunch of stuff scattered plus then there's the web uis that go along with these things so you got a separate sort of javascript app probably that you gotta you gotta handle and build and then have a web server for see and then because you know you can't just trust them then there's not like a unified system for it everyone's gonna run their own thing and you know install apache or add to your nginx or oh god you're making my blood pressure go up just talking about it. It is really a mess. And then you also have like Docker containers. Of course, those can be loaded full of some stuff, but not have other things that you need for your particular hardware. Or
Starting point is 00:23:32 like if you deploy it on the Mac, you won't get access to hardware acceleration in some cases. So there's a lot of edge cases around the containers. And I haven't really seen any of this stuff shipped as like a flat pack yet.'ve been waiting for something like invoke ai and stable diffusion to just be a flat pack and you install it it starts a little web server gives you the url and like a little pop-up and you click that and it brings it up but so far that's not happening and so you have containers with their limitations sometimes they'll work great be exactly what you need and sometimes you're going to be execing into that thing and fixing stuff and adding stuff so it works on your machine. You got the blast and spray and pray method, which is what the Mac users seem to be just going for, which I don't understand. You
Starting point is 00:24:15 know, there's other ways you could do it. And then, of course, you could do it from scratch and you could actually build it, pull down everything. Man, is that a job? Yeah, right. You can kind of follow the developer path and be like, okay, if I was going to be working on this project, what all would I have to do? So as you can guess, we didn't want to do it this way. As you could probably guess from our tone, this is not how we did it. And we think maybe you shouldn't do it those ways either. We think maybe we have a much better way to do this. Determinate dot systems slash unplugged. Bring Nix to work the way you've always wanted with flakehub.com. Go register for the private beta and get secure, compliance-friendly
Starting point is 00:24:54 Nix and all the support you need at determinate.systems slash unplugged. So let's talk about flakehub.com. It's the all-in-one platform for secure, compliant, transformative Nix development, and it is made by our friends at Determinate Systems, the builders of the best Nix installer, including the massively popular MDM and Skiff mode support, which gets 20,000 installs per day, massively successful. And they have an advanced binary cache with sophisticated access controls that if you're working in a team is going to be mind-blowing for you.
Starting point is 00:25:26 They're also the ones behind that cloud-native Nix project we've talked about and just a plethora of open-source tools and documentation for the Nix community. And FlakeHub. FlakeHub is solving big problems if you're building with Nix. Nix caching traditionally involves a globally flat namespace. So imagine building software. It's like having a massive toolbox with tons of parts and tools in there. But finding the right one can take forever when it's a mess, and auditing everything that's in there that got pulled in is just,
Starting point is 00:25:54 well, it's a nightmare if maybe not even impossible. Traditional Nix caching is like having multiple toolboxes scattered everywhere, and each team member has their own set, and it's hard to share the tools or know who has which one and if they have the right one. It's not really elegant. Well, FlakeHub is. It's a single organized toolbox for your entire company, if you will. Imagine one identity-aware cache where everyone gets the right tools based on their permissions.
Starting point is 00:26:22 No more searching through that mess with fine-grained access control so you make sure things stay sensitive and they're in the right tools based on their permissions. No more searching through that mess with fine-grained access control so you make sure things stay sensitive and they're in the right spot. And FlakeHub has that globally distributed build cache. Man, has that been useful here when we've been working with AI tools. If you can use GitHub, FlakeHub will fit into your existing DevOps workflow.
Starting point is 00:26:41 And it's the only SOC 2 certified platform for Nix, and it delivers a low-friction Nix development experience that you've always wanted. Go bring Nix to work the way you've always wanted with flakehub.com. You guys, you've got to check this out. Go support the show and register for the private beta to get secure, compliant, friendly Nix
Starting point is 00:26:59 and all the support you need. You go to determinant.systems.unplugged. We'll have a link for that in the show notes too. That's determinate.systems slash unplugged. And a big thank you to Determinate Systems for sponsoring the Unplugged program. Determinate.systems slash unplugged. Now, if you've been following the show lately, you might realize that we have a favorite way of solving this problem. That's right. We wanted to find out if we could Nixify our AI tools. And we didn't have to go very far.
Starting point is 00:27:39 You know, sometimes it's easy. Nixify.ai is a place you can start. They are trying to just make things simply available, a large repository of AI executable code that might be impractical or hard to build or figure out yourself. They've got it. They've got it over there. You know, I think we saw it kind of float by. Some of our Nix friends were sharing it around sometime last year, but we actually hadn't had a chance to give it a try. And, I mean, yeah, it's got exactly what we want, at least if it works.
Starting point is 00:28:06 Here's how they pitch it. Each project is self-contained but without containers. By the way, when we say Nix, we don't mean NixOS. It would work on NixOS, but this will even work in WSL. You just need to get Nix, the package manager, installed. So you got self-contained applications. It's easy to run. It's like a one command to actually install and run it once you've got Nix set up. Their projects have support for NVIDIA and AMD GPUs. So it'll work with other. And like I mentioned, you WSL users are not left out. And I recognize there's more and more of you out there. Go get NixOS WSL and then you can run GPU accelerated AI workloads on WSL using these commands.
Starting point is 00:28:46 So you can get it right there on your Windows box if you want. Now, Wes, am I correct in saying that they're using flakes under the hood? Yeah, I think that is the primary method that they advertise. You know, like if you go on the site and you want to go get started, let's say you've got an AMD system, you do Nix run, and then basically a GitHub Flake reference that points you over at their Nixified AI slash Flake repo. And then you tell it you want either the AMD or the NVIDIA version, whichever project you're doing, like we were doing Invoke AI.
Starting point is 00:29:19 I want to ask you how you got it working in a moment. I did the spray and pray approach on a demo system. I'm not proud of it. That's why I don't recommend it. in a moment. I did the spray and pray approach on a demo system. I'm not proud of it. That's why I don't recommend it. But a flake is perfect for this kind of thing. I think you could think of a flake as, it's actually pretty simple. Think of it as like a building block for a Nix project.
Starting point is 00:29:37 Imagine a flake is a self-contained box and it has code instructions for a thing. And inside that there's a file named flake.nix and it tells Nix what to do. It'll say, here are your inputs. These are the other bricks you need to build the full wall. And here are the outputs. This is what it should look like. This is how things should be set up.
Starting point is 00:29:53 These are the parameters. And then you can combine these to create bigger and bigger systems. They're just inputs you put in your config file. And each flake has a lock file to ensure it's always using the same exact version of everything that makes it reproducible and reliable so that's really big from an audibility standpoint or maybe you're deploying an appliance or maybe you're just trying to get something that worked that worked on this machine to work just like it did on another machine and that's i think particularly where it's useful for these types of tools they've got a lot
Starting point is 00:30:23 of brittle parts and are moving quick. So you could think of like regular NICs is you're building everything piece by piece with bricks, and you're assembling those bricks and you're building the wall. With flakes, they're pre-made chunks of the wall with clear inputs and outputs. They're like sets of Legos that you can grab and set down in a whole set of Legos all at once. Yeah, you know, if you've ever done anything with, you know, like PIP or NPM or any of these sort of programming language specific installers and package managers, you know, they've got these lock files and they kind of let you manage everything and make sure you do things in chunks that you can comprehend and you've tested and, okay, I do want to
Starting point is 00:30:59 update them all for me or, you know, leave them pinned for now. Flakes and Nix kind of let you take that to all of your packages. You know, it extends that approach to the whole system and it lets you easily integrate with just GitHub repos, whatever Git URLs you want to use. And like you're saying, this stuff's all moving so fast. You might have custom versions of stuff, forked packages, different versions of the web UI.
Starting point is 00:31:23 You're not necessarily going to want to wait for that to get into a centralized Nix packages repo. Or any distro. Right. Yeah. With the inputs, the Nix Flake stuff goes and handles going and checking what is the most recent commit, grabbing that, writing it in your lock file, pinning it, and then getting you all that stuff onto your system and then feeding it into the rest of your Nix stuff. So all that is just taken care of
Starting point is 00:31:45 for you. Now, late in the night, we all tried our various methods, but Wes, of course, did the Nixified approach and got it working. Did you use a flake? Did you use a different installation approach? How did you get things working on Nix? No, I just went with the Nixified AI
Starting point is 00:32:01 stuff. I did see that Invoke AI has a flake.nix in their repo. Yeah. But it's mostly just set up for getting the raw dependencies. So it will get, like, handle all the CUDA stuff. I think it built OpenCV, so it took a little while. But, you know, thankfully that rig has plenty of CPU. Shout out to our audience that donated very nice servers.
Starting point is 00:32:20 I think that would be a nice setup if you were explicitly trying to develop on it. Because you kind of get all the different Python libraries and the stuff you needed. And then you could just do the build yourself right in that environment. But if you want something that's a little more end user focused and packaged, that's where the Nixified AI version came in. I think they had some caching in place, which was nice. So you didn't have to build absolutely everything. Right. You can do the Nix run, but you might want to do Nix shell because there's more than one binary
Starting point is 00:32:48 for some of these. So in particular for invoke AI, you've got just the invoke AI command, which does like a command line version, targeted more advanced users, maybe for like scripting or automation or that kind of thing. But you probably want the invoke AI dash web command because you're trying to run the web UI.
Starting point is 00:33:04 Yes, right. And I think by by default maybe there's one it comes with but there's a whole bunch of different models this thing can use so you're going to need to do some configuration and most importantly you're going to need to actually pull down all that stuff so you're going to want to run some stuff before you get it all launched uh where they have a configuration command you can run uh it's even got a little sort of mouse enabled uh command line interface that you can run. It's even got a little sort of mouse-enabled command line interface that you can kind of click through either with tab and spacebar or double-click right there. Yep. Let's you select what models, some of the options, some of the enhancements further
Starting point is 00:33:32 on in the pipeline that you need. Yeah, it makes it, it sounds like it's a lot of work, but that part of it makes it really simple for anybody that's even new to this. Yeah, I didn't super know what I was doing, so I just kind of clicked some of the ones. I was like, yeah, that sounds good. And it, for the most part, just worked. It does take a little while because there's gigs and gigs of data to download, depending on how many models you choose and which ones. But it downloads it. You don't even need to restart anything. The web UI will pick up the new models. There's even a little button
Starting point is 00:33:59 on there to refresh. And now you've got InvoKI. And you've got a web UI that lets you generate images, and then you can play around with the different plumbing and the different models behind it and the different accelerators that they have to get everything tweaked just right, and you can start creating your own pictures. And, you know, what stands out to me so far about InvoKI
Starting point is 00:34:18 is you can tell it's kind of like there's workflows that they have in mind. There's all kinds of, like, you know, setups here so you can see the queue that's going on yes right um you can have like we could have no sorry to interrupt but just to something was really neat is we could share a central invoke ai server and we could all have different images being in the queue and i could see yours and you could see mine and they would just be generating in the order they were submitted yeah and you can you know you can set up different workflows on here. I don't know how, you know, it's not maybe
Starting point is 00:34:47 competitive with everything, but you can see that you're getting a little more sophistication than just like, you know, here's the input form and then go run the model for me. Yeah. So I'll tell you, obviously I use this. If you go to jupiter.tube, you can see how we use this. For me, it's like generating stock photography instead of paying for some stock photography website, I just generate my own now. But I also, I've used this now to generate backgrounds for my kids' devices. So they have like, you know, a custom background from dad. My wife has used it to generate like her perfect wallpaper for her iPhone. So she has like this photo that she just loves to look at that is something that she kind of created and then put it in there and created it for her iPhone. So she has like this photo that she just loves to look at. That is something that she kind of created and then put it in there and
Starting point is 00:35:26 created it for her. And there's just ways you can use it that are not necessarily, you might think of at first, but the more you have these kinds of things and you think, Oh, it could kind of be nice to have a little piece of art there. It doesn't have to be all that important. It just needs something to add a little visual splash.
Starting point is 00:35:38 It's just great for that kind of stuff. And they have a bunch of other things on here too. Of course, like one, I did not get a chance to play with, but I have it installed installed is a text gen which as you can probably guess is a web ui for large language models to spit out different stuff and it supports a bunch of the different models and i have that installed on my system but i just didn't get a chance to play with it but again using the nixified ai method it's so straightforward and because it's not NixOS dependent, you can be on Ubuntu and be using it
Starting point is 00:36:06 or you can be on macOS or WSO. I think that's pretty powerful because then you get a consistent experience across all your different systems. And of course, you know, it is still modifying your system. In this case, it's downloading models. It made a whole directory to keep a bunch of stuff and store its settings. But, you know,
Starting point is 00:36:22 it's not, especially in that case, it didn't have root permission. It was running as a regular user. So, you know, you's not, especially in that case, it didn't have root permission. It was running as a regular user. So, you know, you can make a whole separate user to run it as if you'd like. And then everything else, the stuff you built via Nix, that's all just in the Nix store. And when you go to collect garbage, when you're done,
Starting point is 00:36:36 you didn't like that model, you want to do something else, or you're just going to deploy it another way. It's cleaned up. It's cleaned up. I think our opinion is, is that it's easier to use. You get reproducible builds, and you have better organization It's cleaned up. I think our opinion is that it's easier to use, you get reproducible builds, and you have better organization of these really complex projects that can have a lot of pieces.
Starting point is 00:36:53 I have a question on what's reasonable hardware to expect this stuff to even be useful on? Because I know some of us are running it on fancy hardware, but can anybody just run this on their work laptop and give it at least a try? I mean, obviously the performance is going to be different, but. Yeah. How painful, Wes, would you say the image generation was on the CPU? Decently, decently painful.
Starting point is 00:37:17 Yeah. I mean, probably like four-ish minutes. Yeah. Versus the GPU, it could be a lot quicker. Yeah. So it depends on how impatient you are if you have patience right uh you know and it is nice like it gives you the little um you can set it so it'll show you the in progress sort of pictures as the model is refining and diffusing
Starting point is 00:37:35 um and uh in the command line or in your logs it'll also have a little progress bar uh so you at least get some pretty decent feedback especially compared to some of the commercial systems in terms of like, how long am I going to be waiting? But you are probably going to be waiting. I think, you know, there's a lot you can start with. I think invoke AI is a fun one. If you're listening to this and you're thinking, I do want to try some of this local stuff. I do want to try to help these projects grow. I would like to understand this a little better. If you get invoke AI deployed, the building blocks you use to do that will work for all these other ones as well. There'd be little individual little tweaks you might make if it's depending on what the software is, but like the method and the
Starting point is 00:38:12 approach and the documentation is all going to be the same at that point. So it's, it's a journey, but once you learn it, you'll have that path well-treaded. And I think it's, I mean, I think it's like, we're just going to leave it. We're just going to keep it. I mean, it's going to keep updating now. It's like, we're going to be current with the projects as they release stuff. And it's great for us. Like we've just set it and forget it. Collide.com slash unplugged. Well, you've probably heard us talk about Collide before,
Starting point is 00:38:39 but did you hear that Collide was just acquired by 1Password? It's pretty big news since these two companies are leading the industry in creating security solutions that put users first. For over a year, Collide Device Trust has helped companies with Okta ensure that only known secure devices
Starting point is 00:38:54 can access their data. And, well, it's still what they're doing, but now they're doing it as part of 1Password. So, if you've got Okta and you've been meaning to check out Collide, now's a great time.
Starting point is 00:39:04 Collide comes with a library of pre-built device posture checks, and you can write your own custom checks for just about anything you can think of. Plus, you can use Collide on devices without MDM, like your Linux fleet, contractor devices, and every BYOD phone and laptop in your company. Now that Collide is part of 1Password, they're only getting better. So go check out collide.com to learn more and watch their demo. It's K-O-L-I-D-E dot com slash unplugged. Watch the demo, support the show, collide.com slash unplugged.
Starting point is 00:39:40 Well, conference season is coming up quick, and we have some updates specifically around Texas Linux Fest, which is happening April 12th to 13th. Chris, can you catch us up on those updates? Yes. Well, mainly the time slots and the schedule are now up on the website. So if you've been trying to look at what's going on, the talk titles and the speakers have been filled in as we have them, and more will be coming up there as they get confirmed. And it sounds like there's at least some planning in the works for a party Friday night. Oh. Potentially. And there may be others.
Starting point is 00:40:08 They tend to be a little more kind of just planned at the last second. And I think the good news for anybody that's considering going is we do have a hotel group rate link. So I guess normally like this hotel is like 320 a night. And Texas Linux Fest group rate is going to be 250 a night. So it's not bad. And you're right there. You know, you don't have to go far. So that's good.
Starting point is 00:40:30 And again, that's April 12th and the 13th. And then just shortly after that, Linux Fest Northwest 2024 is the 26th and 28th of April. And the update here is the schedule is posted. You'll see some familiar names on there. I believe there's a Wes Payne on there. And if you go to Wes's talk on Sunday at Linux Fest, just stick around in that room because then we're going to do Linux Unplugged after that right there. It's going to be a whole day of fun. Yep. So you can just hang out. Pretty great. Shout out to the core contributors. They
Starting point is 00:40:59 are participating in a membership program that finances the show directly by the listeners and gives them access to additional content. There's an ad-free version of the show, but I think the real winner is the bootleg version because the pre-shows have just been bangers. Some spicy pre-shows recently. Some spicy couple of pre-shows in there. Maybe it might be worth checking out just for that.
Starting point is 00:41:20 You do get a little of the sacred sauce. And now it is time for the boost. Thank you everybody who boosted in this week. We got a whole big batch of Fountain FM feedback and I, well, with Wes's help, bundled it all up and sat down with Nick and we went over all of it. And if you've been boosting from Fountain,
Starting point is 00:41:41 I think nearly all of them, Nick has been replying to directly. So go check your Fountain FM app. So I'm not putting all of it in the show because some of it's getting handled directly. But I will put some of it that I have answers for. And our baller boost came in from User38 who sent in 51,000 sats. Hey, rich lifestyle! They write, my original donation was intended to be 150,000, but I had some problems with Fountain.
Starting point is 00:42:09 Maybe they could fix the custom donation feature, which was broken. Yeah. Yeah, really. Come on. You still managed, even with that bug. We appreciate it. You managed to become our baller this week. Hybrid sarcasm comes in with 43,000 sets.
Starting point is 00:42:24 I hoard that which your kind covet. I would like Fountain to allow me to build playlists from subscribed show feeds, not just individual show episodes. I like to listen to my podcasts in groups, you know, like tech, news, dad stuff, etc.
Starting point is 00:42:40 And just can't be bothered to peruse show episodes individually themselves. I want to know what dad stuff he's listening to. I don't have any dad stuff podcasts. Would you boost in and tell me what you're listening to? Give me a recommendation. So I wanted to put this one in the show because I talked directly to Nick from Fountain about all of these. And this one he said is doable today.
Starting point is 00:42:59 You can tag the episodes and then you can build a playlist based on the tags. So that is how you solve that problem. Sarcasm, it is available to you already. And I wanted you to know that. Thank you for the boost. Ian clearly boosted in a mega row of ducks. Things are looking up for old McDuck. For the Docker Compose versus Arian discussion,
Starting point is 00:43:17 I ended up on the Ansible Docker Compose for Ansible vault. I didn't understand secrets when doing it with NixConfig. I get that the config says what file you need, but doesn't store the content or secrets. Am I missing something? We've done this. We've had definitely configs where we refer to a secrets file. What are we doing different?
Starting point is 00:43:37 Well, I mean, secrets in Nix is a whole topic. So that depends on the approach that you want to do. There's the simple, or maybe it's not simple, depends on what you think. But there's the approach described here that we've done There's the simple, or maybe it's not simple, depends on what you think. Okay. But there's the approach described here that we've done and we did in our example Nix cloud setup where you just manage secrets separately.
Starting point is 00:43:50 Whatever you'd like, you could use Ansible Vault, you could use, you know. Which we did link to, so he may be able to refer to that. And that just means you need a separate process that's going to set secrets up where you want them and then the Nix stuff is just set up to refer to those and expects them to exist. All right.
Starting point is 00:44:04 Or you can deploy stuff in Nix. There's a few different options. I mean, you could just embed them in there if you weren't worried about that. But you do have to be careful if you're worried about security or you're on multi-use machines, because depending on what method you use, you probably don't want your plain text secrets in the Nix store, because anyone can just go and read that. But two popular ones, there's AgeNix, as well as Sopsnix, S-O-P-S. And then there's a bunch of other projects as well.
Starting point is 00:44:30 So depending on the approach, the complexity, what your actual needs in terms of deployment are, there are a bunch of ways to do it in Nix. But that's not something that we built in or planned around our solution. And of course, Ansible and Ansible Vault is a time-honored way to do it. Yeah. I mean, if it works. Sounds like a nice setup. Thank ian appreciate the boost sir lurks a lot comes in with 2048 sats hey sir nice to see you so i like hearing your experience with vr headsets like the quest chris but i wonder how well they will work for people with poor eyesight do you wear glasses inside the headset or does it have optics to adjust to focus for each eye? I don't think it has individual eye focus, but you could wear glasses.
Starting point is 00:45:07 I know some people do. For me, I'm, as I age, becoming nearsighted. And so since the lenses are just like right in front of your eyeballs, for me, it's fantastic. I, in Immersed, also realized that you can adjust some of the encoding quality and things. So I didn't actually have everything turned all the way up to like its pristine quality. It's looking even better than it was for last week. I used it for prep on the show. And one of the things that I've also done now with my Immerse setup, which I can't believe I didn't think of this sooner, is I've replicated the monitors like I have them here at the studio. So I have a vertical screen here at the studio and I now have that same on that same
Starting point is 00:45:44 position. It's a vertical screen and I have the, and I now have that same, on that same position, it's a vertical screen, and I have the same horizontal screens. And so it's very much the same workflow for me, in immersed or in the physical world. And I'm liking that a lot. VT52 boosts in with Aerodux. Reusing NixOS.org search for NixOS options. Little pro tip, try searching over at mynixos.com. It searches across options and packages at the same time. So you'll find eg Nix packages slash packaged slash tmux
Starting point is 00:46:15 and Nix packages slash options slash program slash tmux in one search. That's cool. The UX is a little nicer too. You can browse among the options hierarchically instead of one flat list. Bonus, it also searches Nix packages and Home Manager. Yeah, this is a great one. I kind of forgot about it, but I have run into it a few times,
Starting point is 00:46:36 you know, just when trying to search for Nix options and not already at the main search site. It's great to have other options. Thank you, VT. MyNixOS.com. That is great. Thanks VT. Really appreciate that. I have a little pro tip to add to this that I've implemented this week. So I've been searching the Nix packages, um, through duck, duck, go bang implementation.
Starting point is 00:46:57 So you bang Nix packages and then you just search for what you're looking for and it sends you directly to the Nix package search search but i've one upped that this week on kd plasma and i'm using the plasma launcher which has the same functionality if you look deep enough in the settings and i'm going searching directly from my desktop it's amazing yeah nice yeah i need to use krunner more it is the best runner out there i feel like it's gotten a little pokey. I'm hoping with plasma six, which comes out in a few days, I'm hoping to fall in love with K runner all over again. Well, Gene Bean boosted in with a row of ducks.
Starting point is 00:47:35 Hey, regarding fountain, it's missing the dynamic playlists that are used in castomatic and overcast are the show stopper for me. For example, this show is in a playlist called Linux and Tech that contains the shows I subscribe to that fit that particular topic. The playlist automatically shows me the unplayed episodes for these collections of shows. This allows me to see what I've heard and haven't heard yet in a given topic. Tags, labels, and manual lists just don't cut it for me. Hmm. Tags, labels, and manual lists just don't cut it for me.
Starting point is 00:48:11 How many people are doing this where you're like you're play listing out the topic like Linux and tech for Gene being here? I just always have and I'm curious about you guys. I always just have like my main list of chronologically released podcasts and I just kind of scroll through that and pick what I'm going to listen to. And I don't ever create a playlist or anything like that. I think at most I might create like a, you know, per session, like I'm going for a drive and I want to keep up a couple, something like that. But I do have, I do have road trip playlists. What about you, Brent? Do you just stack your, your list or do you have playlists? Do you go all organized? No, I think I would love this because typically when I'm reaching for my podcast players, because I'm in the mood to listen to a certain type of content. And so I end up Chris, like you browsing through the list.
Starting point is 00:48:48 But I find that more to be friction than inspiration most of the time for me. So I think I would love this feature. LazyLocks comes in with 10,000 sats. It's over 9,000! All your talk of virtual displays has got me wanting to check out links in the show notes. That visor looks promising. I hope it delivers. Looking forward to seeing you guys at scale.
Starting point is 00:49:08 Yeah, Lazy Locked. Heck yeah. So visor.com, the deal I think is over now. But it was like 400 bucks for VR headsets that are just designed for working. Really high quality displays that look like sunglasses. So you're not wearing big old ski goggles. I think people like that. Master Rebo with nine thousand cents just following up on a telegram message about the comparison between quest 2 and the quest 3 if i look dead center everything seems fine on my quest 2 but if i stray just a little
Starting point is 00:49:41 things begin to look fussy can you tell me if the Quest 3 is like this as well? The Quest 2 is my first VR, so I don't really have anything to compare it to. That's a great question because you would think this wouldn't really be viable if you looked at this through the Quest 2. I tried the Quest 2. I saw what you're seeing. I tried it this weekend with my daughters. It's one thing for a game. It's another if you're working in the terminal,
Starting point is 00:50:03 you're composing an email, and you're reading text for hours. Huge difference, yeah. Yeah, with a 3D game, you don't really care. But when you're trying to read text on a monitor, big difference. So on the Quest 3, you have a much, much better field of vision. And I noticed, because I had probably, I mean, my screens were at my, I zoomed zoom my screens to test this for you at the maximum width of my peripheral vision. And at the edges of my peripheral vision, the monitor does kind of have like a little kind of a blurry effect to it at the very, very edges. But I'm talking like if you held your hands to the edge of where you can see them, that's where it starts getting blurry at.
Starting point is 00:50:39 But then the field pretty much from in there to the center is looking really sharp. And if you use immersed, make sure you mess around with the encoding options, and that also makes the text look better, I've learned. So you got to – the defaults are optimized for Wi-Fi performance, and they assume your Wi-Fi is crap. But if you've got good Wi-Fi or you hook it up over USB-C, you can step up the quality and the text looks even better. Bear 454 boosted in 5,000 sats. Well, hey there,
Starting point is 00:51:06 Bear. I've tried Fountain a few times, but just keep going back to podcast. And there are a few reasons for that. Number one, the home screen makes it feel a lot more like a social media platform than a podcast player. Let's take these one by one. So the home screen making it feel like a social media app than a podcast player, I agreed with too and was the biggest friction point I had with Fountain initially. Then I discovered like six new podcasts that I never would have found ever and they're my regular listens now and I've changed my tune a little bit. Then I also – I started following a couple of other JB listeners and they started following me. Oh, that's fun. And now I'm seeing their comments and what they're listening to,
Starting point is 00:51:50 and they see my clips and now it makes more sense. But bear, I definitely had to take, I don't know, man, three weeks of just being like, I really wish you would just open to my library. I really wish. And now when I launch fountain, I actually stay on the home screen because I know in the, in the library screen, it's just downloading my podcast anyway. So while it's downloading my podcast, I just scroll through the home screen, see what other listeners have been boosting or listening to. So I know it's not exactly what you want to hear, but you might just give it a reframe of thought. Bear says, number two, I find it really hard to see new unplayed podcasts.
Starting point is 00:52:18 Even when I play an episode, it stays on my episodes list. I would appreciate an option just to see new unplayed episodes. Yeah, or auto-delete maybe when you finish an episode. So they've added swipe to delete, that makes that a little easier. But yeah, I think right now by default, unplayed are still chronologically listed, they just are now marked as played. And bears number three, my biggest gripe is I really would like an option to make a good use of bigger text. The text in the app is just small for me. Right.
Starting point is 00:52:49 I mean, I wonder. I know on iOS you can pop up the zoom up. I don't do that on my Pyzelle, but maybe there's an accessibility option there. I'll chat with them about that particular. And Bear, I know you had some Albi issues too. If you hit me up on Matrix, I can help you troubleshoot that. That's good feedback though. And those are things we will be talking about in our next sit down next Thursday.
Starting point is 00:53:09 Presley Weep HD. I'm going with it. Comes in with 2,669 sats from Fountain. And he says, I got a math problem for you guys. It's the sats times 17 equals my zip code so he's doing a little math game with us did you bring that map oh you did this is why we're weeping in hd i make sense now uh yeah 2669 times 17 that would be uh 45373 and that seems to be a postal code in Miami County, Ohio. Ah, hello, Ohio.
Starting point is 00:53:48 With cities like Troy, Alcany, or Staunton. So they got a county called Miami. And they got a city called Troy. This is great. This is so great. I mean, you're just like, where do you live? I live in Troy. In Miami.
Starting point is 00:54:03 You know, Ohio. Thanks, Weep. Appreciate that. Boost. Difficulty adjustment, Boost-in, with 10,000 cents. It's over 9,000!
Starting point is 00:54:15 Love your work, gentlemen. A video option on Fountain would be a game changer. Little Birdie tells me that may be a possibility. I think a lot of the plumbing is there, and a lot of the underlying
Starting point is 00:54:24 libraries support it. We may get there with our LIT support for our scale trip. We shall see. So would that be taking advantage of like alternative enclosure setups? Yeah, you got it. So podcasting.to.spec has the alternative enclosure and in LIT too you can do here's my primary stream and here's an alternative
Starting point is 00:54:39 stream. You basically put like HLS or RTMP URL in there. And then the client player would just grab that and figure it out. So it could happen. Difficulty adjustment. I like that username too. Barnminer comes in with 5,555 sats. He says, tell Fountain I switched to Breeze Wallet for boosts since it's self-custody.
Starting point is 00:54:58 I actually did. And Nick's response was, yes, that is on the list. It's been on the list for a while. It's still on the list. Well, Network Rob sent in a Spaceballs boost. So the combination is 1, 2, 3, 4, 5. That's the stupidest combination I ever heard in my life! So I'm boosting from Fountain today for a couple reasons.
Starting point is 00:55:19 One, giving a little extra value back to my stream sets, and also, number two, scale! I know it's a little, but I know every little bit helps. Number three, keep up the great work. Love the value to value. And number four, well, since I'm behind on episodes, giving a small hello to future Rob when I get caught up. Eight lines of shimmering cement running from here to Pasadena.
Starting point is 00:55:42 Thank you very much, Network Rob. Appreciate that catch-up boost. We look forward to hearing from you when you do get caught up. So excited about scale in NixCon. So, so glad we're going. I feel like we don't give enough attention to the other events that are going on, but this is the big one that kicks off
Starting point is 00:55:57 the year, right? And we only have so much space in our tiny little brains. That is probably really what the problem is. Ben the Tech Guy comes in with Road Ducks. Says Steam Link is already in the Met little brains. That is probably really what the problem is. Yeah. Ben the Tech Guy comes in with Rodeux. It says Steam Link is already in the Metastore. I mentioned I sideloaded it. And it does support proper VR, but only on Windows and Mac OS.
Starting point is 00:56:16 They're actively working on Linux support. But in the meantime, ALVR works pretty well. Okay. I didn't even look for it in the Metastore. I just got so in the workflow of sideloading APKs, I was like, I'll just go get it. I don't need no store. I don't need no store.
Starting point is 00:56:30 This is Sightel, former iOS user, folks. No, don't you think I would have defaulted to the store? Well, I mean, you're so excited about the ability to sideload at all. Yeah, you might be onto something there. Mr. Pibb boosts in with 12,345 cents. 3, 4, 5. Yes. That's amazing. I in with 12,345 cents. My issue with Fountain is I use an auxiliary cable to listen in my car, and the sound quality always gets distorted at the necessary volume. I use the iPhone app, and I'm always fiddling with phone volume and car stereo volume trying to get it
Starting point is 00:57:03 loud enough without distortion. I have the same problem on Cast-O-Matic. I enjoy the JB shows for the professional-sounding audio quality, and it's great with my Bluetooth headphones or a speaker, but just not with the auxiliary cable in my car. But if I use Spotify or Apple through the aux, it sounds just fine. Now, what could be causing that? Distortion when you're using certain apps, but not when you're using Spotify or the Apple
Starting point is 00:57:29 apps. Are these other apps, you know, normalizing, boosting the audio to a level where, you know, there is some distortion at those higher levels, but you just don't have to get to the higher levels with these other apps? So it's something like a, you know, a louderizing sort of. And it only happens when you're using an aux cable. That's interesting. Yeah.
Starting point is 00:57:46 If anybody has any ideas on this one, let us know. Also, remember, we're collecting your boost ideas for AI resources and tools that we should be looking at for the show. So please boost those in. Thank you, everybody who did boost in. If we didn't read your Fountain FM feedback, don't worry, Nick and I covered it,
Starting point is 00:58:01 and we're still collecting all of them. And check your profile. He's probably replied to you directly. We had 24 boosters come in, and we stacked 222,219 sets. Almost a row of ducks. A big old row of McDucks. So I'm going to give it to us. Thank you kindly. I'm going to call it a big old McDuck.
Starting point is 00:58:18 Appreciate everybody who boosts in. It's a great time to get a new podcasting app because we are working on new features. We've already rolled it out to our members. They're loving it. And they're going to be coming out to the main feed very soon. And some of them are going to start as we head off to scale. And we'd like you to be there with us and keep podcasting weird and decentralized. Podcastapps.com, Podverse, and Fountain FM, and Castomatic are really the favorites of the show.
Starting point is 00:58:42 You pick the one that works best for you and then boost in with a message and you can support us directly that way. No middleman, all using an open source peer-to-peer network. And thank you everybody who streams Sats to us as well. We appreciate you. And a big shout out to our members out there for using their Fiat Fund coupons to keep us going on the regs every single month. We really appreciate them too.
Starting point is 00:59:03 You mentioned that a lot of these large language models and whatnots are coming out with a paper. And I think one of the best ways to just wrap your head around what models have come out recently and what problem they're trying to solve and if you even need to pay attention
Starting point is 00:59:19 at any of that is probably the daily papers over on HuggingFace.co. And every day, and it is every day, it's pretty wild. I mean, every weekday, they're posting just one white paper after another. And it's, I mean, I don't know, Wes, it's a fantastic resource for just kind of watching the development of these. You must have taken a look at this earlier.
Starting point is 00:59:40 It's a, I mean, HuggingFace.co itself is a resource. I'm going to assume most people listening to this already know about it, but as a higher level recommendation, hugging face.co, if you're into any of this kind of stuff for the community around these open source models and data sets and yeah, they host a bunch of, a bunch of stuff like that. And then they have these daily papers and that's at hugging face.co slash papers. And these are the white papers that accompany the announcement of these large language models or whatever they're working on. And I think just you can look at the headline, you can look at the summary and know if it's something that interests you or not and pretty quickly stay on top of all this stuff. I love resources like this.
Starting point is 01:00:18 If you have any out there, let us know. Yeah, and you maybe take a look and see just how fast it goes from paper to something you can actually use. Yeah. Yeah. Yeah. I would say if you quit Reddit recently, you can go here. Even if you don't want to dive into the papers, just looking at the names of these projects is such a treat. So just scroll through the names. I'm enjoying that.
Starting point is 01:00:36 Yeah, I feel like this is our moment in history as a community to keep this stuff open and to keep the open store stuff viable, to keep the community vibrant. You guys probably saw that on the back of all this Gemini news, Google also announced though a massive, what was it, like $60 billion deal or something like that with Reddit to pull in all of Reddit's data to train AI. That cost and those kinds of deals, those are going to always be relegated to the absolute top, right? The big tech companies that can afford to write 60 bill or 30 bill or 10 mil or whatever it is, 30 mil. Yeah. I mean, JB tried to buy the Reddit data, but they weren't interested. Yeah.
Starting point is 01:01:17 It turns out they wanted more than a row of ducks. But like you look at things like Hugging Face. You look at Invoke AI. You put it all together together and we still have a very, very viable shot here. And I think ultimately if I could be a dreamer long term, I'd love to hear stories of listeners that are implementing this stuff in small businesses or in their workplace, maybe even large businesses. Because that's where I fear the most that we're going to get locked into corporate AI. And it's going to be this risk averse, not very useful, overly talkative, watered down crap tool. And the reality is, and the BSD folks have always been right about this, is a powerful tool you can shoot yourself in the foot with. And you can do good and bad with rm-rf. And I feel like it's the same kind of general principle with this stuff.
Starting point is 01:02:09 So if you're out there in the field, you're deploying this somewhere where you've got more than a couple of users using this stuff, boost a write-in and tell us how that's going. Yeah, are you relying on AI-powered stuff for your business? Would you? Do you feel differently about it if it's an open-source thing versus something you can only interface with by API? Yeah, I sure do. Yeah, I sure do.
Starting point is 01:02:30 It's just a wild new frontier. And we're really just kind of at the beginning. I thought we'd see how NVIDIA went. And depending on how NVIDIA went, we knew how much momentum there was in the industry. And NVIDIA popped to the high end for sure. I think there's still a lot of legs left in this. for sure. I think there's still a lot of legs left in this. But where it actually ends up ultimately and what kind of end result we get,
Starting point is 01:02:51 that is not so clear at this point. And we, I believe, still have a chance to shape that future. We'd love to have you join us live. You can get in that mumble room, which is popping this week. Lots of folks in there. We always hang out with them and they can pop in during the show or before and after the show is really when we just get to chatting. It's great to hang out. It gives us that live vibe. We've got details on our website for that, and we do the show on Sundays at noon Pacific, 3 p.m. Eastern. Links to what we talked about today, linuxunplugged.com slash 551 for the tools and the links and the resources, all that kind of stuff.
Starting point is 01:03:22 If you thought this show was useful or maybe somebody else should hear it, please share it. Word of mouth is the number one way podcasts get marketed. It's not really through anything else. It's not going to be Google ads, I'll tell you that. So I appreciate that too. Thanks so much for joining us this week. See you next Sunday. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.