Grey Beards on Systems - 173: GreyBeards Year End 2025 podcast
Episode Date: December 29, 2025Once again in 2025, AI was all the news. We are seeing some cracks in the NVIDIA moat but it's early yet. Broadcom VMware moves and DRAM sky-high pricing sort of round out the year for us. Listen to o...ur YE podcast to learn more.
Transcript
Discussion (0)
Hey everybody, Ray LaCasey here with Keith Townsend.
Jason Collier here.
Welcome to another sponsored episode of the Greybirds on Storage Podcasts,
show where we get Greybirds bloggers together with storage.
This is the vendors to discuss upcoming products, technologies, and trends affecting the data center today.
This is our annual year-end podcast where we discuss the years technology trends and what to look for.
for next year. Keith and Jason, what would you like to talk about?
You know what? How can we not talk about the AI bubble? Like, this is crazy.
Yeah, the AI is going crazy. I mean, the whole agenda is driving it, more Gen. AI,
more AI in the enterprise and stuff like that. It is going insane.
I don't know if I'll make an argument that we're seeing the AI actually in the enterprise,
but we are seeing it on the training side.
Yeah.
We are seeing it on the enterprise side.
So it's been one of those things where we're seeing
and basically increase,
and we're actually seeing kind of a shift
in basically the going from training to inferencing models.
We've seen a lot of folks basically taking
and almost establishing what I'd call this like three-tier model
of where they've got an LLM running as a base,
and then there's a kind of a smaller language model
sitting on top of it,
and then they're doing a retrieve
a long-minute generation on top of that.
And that's starting to basically then feed into
what kind of a lot of the agent-agetic AI
for enterprise types of programs.
Yeah, so it's been interesting.
We have definitely seen a shift in folks going,
from looking specifically at doing training to the point of,
hey, there are certain established like large language models out there
that we can run on top of VLM, Olamma, any other types of platforms,
and then kind of start to develop applications, you know,
on top of that that are using those as the base.
So you're saying a full deployment then of, I would call it the AI stack,
but I mean, it's more than just that, right?
Yeah.
And a good example, you know, of kind of how that works.
You know, the LLM, like people are just going out and they're taking the big ones,
taking like, you know, a GPTOSS or or basically deepseek or whatever you want to put as LLM.
Actually, we're seeing also a combination of people using many different types underneath that,
depending on the application.
Certain LLM support those a little bit better.
But then they're also going through and looking at more specialized models.
This is where we're seeing kind of training and enterprise.
is they're training, you know, basically those smaller language models that are very, what I'd say, you know, vertical application specific.
So things like, you know, if it's manufacturing, if it's legal, if it's medical, right, they're doing like, you know, say if it's legal, they're taking the Lexus Nexus database, train it on that and then also train it on local wall, depending on the city, the county, the state.
And you've seen all this happening in on premises, not necessarily in the cloud, but it's on a little bit of both.
A little bit of both.
And I'd say a lot of those kind of the tier two folks, think of, you know, your core weaves, vultures and companies like that that are, they're doing a lot of the training for those.
But then also we're starting to see it in the enterprise, you know, some of the Fortune 500s that are wanting to deploy, you know, some of those bigger stacks.
And they're actually wanting to do, you know, run a lot of the inferencing stack.
So I'd say those tier twos are getting a lot of training business.
And then the enterprises themselves are running a lot of the inferencing stuff.
on-prem.
On-prem.
Keith and I were talking with Articulate a couple months back,
and they've got a lot of, I would say,
vertical-specific models that they're selling as a service, actually, right?
Yeah, so very much that motion of saying,
okay, we're going to take a small, large language model,
train it specifically on a domain,
or build one from scratch,
depending on the domain and the application.
And then run it in a dedicated inferencing cluster, whether that's on-prem or in the public cloud.
But the idea is pretty well established.
And we're starting to see, I don't know, Jason, if we're going to actually see the inference flip next year.
One of my predictions for my only prediction for 2026 is that we're going to see a shakeout of all the evaluation.
bubbling that's happening with these smaller AI companies on whether or not
inferencing is going to be adopted at a pace that justifies the the frowniness of this
investment.
Yeah, I mean, AI startup valuations are going bonkers.
I'm going to a roof.
I saw one that was been, I don't know, a couple months old, and half a billion dollars.
It almost makes me want to go back into AI in a big way.
Yeah, I think Open AI as of this writing, they just got an investment this week or last week that valued them at $900 billion.
This is a company that's burning to what, $10 billion or more at whatever pace that they're doing it.
Yeah, yeah.
My overall guess, and the reality is that training piece on doing those large language models, doing those really just the ginormous ones that those guys work on.
just the sheer amount of compute the cost that that is associated with it is I mean I think the last time I think it was like roughly estimated at $10 billion was the
training costs in basically power cooling compute resources necessary to train one last GPT yeah yeah the GPD 4 GP5 I don't remember what right right but it's I mean it's a lot of resources and when you think about it
You know, and enterprising isn't going to have that many resources to do it.
So this is where I think, you know, the big companies and the companies are going to be the hyperscalers that are between the, you know, the open AIs, the, you know, Open AI, GROC with X.AI, Microsoft, and Google with Gemini, IBM with Granite.
Those are going to be the types of companies that are going to train those LLMs.
and I would guess, and here's my prediction,
I think they're probably going to be about 12 to 24 of those companies
that are going to be training those models.
Yeah, I question, I question how many of the models of those models
do, does the industry actually need?
Yeah.
Because, Jason, I'm seeing that motion that you mentioned,
which is you take a open source, a large open source general,
purpose model, put a smaller model on top of it, some RAG associated to it, some data optimization,
and that's what agentic AI workloads are being built on, not these huge models that are
great general purpose or general knowledge models, but not necessarily great for, you know,
these structured tasks within a specific domain.
Absolutely.
Yeah, I mean, especially when you look at, I think there was some, there was some statistic on the, that I think 60% of, of the knowledge it was used to train one of the models.
And I don't remember which one it was, but 60% of it came from Reddit threads.
And then I'm just like, wow, that's a, that's some really trustworthy information right there.
That would explain my experience with Claude.
Yeah.
Yeah, that's, that's, that's bizarre.
I think the other thing that seems to be occurring to a large extent, where,
starting to see a little crack in the NVIDIA armor with respect to GPUs.
I mean, there are, you know, Google's made, ah, forever really, been talking about
TPUs and their TPUs are more energy efficient and, yeah, it's almost like they're starting
to commercialize that as a solution rather than just internally use it in the hyperscaler.
But also like ML-Perf, we saw that AMD GPUs are starting to make a splash there.
yeah i mean i think the gp u yeah definitely the proliferation i mean you know invidia's had a really good head start
with right you know i mean that's i mean they kind of you know they've got a big head start on that
and you know i'd say probably the uh and you know the google tp u is great and then you know what
amd's doing with with the instinct uh components are are also good and you know i know there's
been a lot of investment on both both companies to really kind of up the software
software game and I think it's really kind of that GPU proliferation is really coming from basically both of those folks up up in there the software game Google's really good at doing software right but the I'd say the thing that that they're a little bit behind on the TPU I mean it's basically it's very domain specific and it was very has been very domain specific to running on Google's cloud for a very long time so how does that you know how would that translate to on-prem and
is it just going to be running the cloud?
And then also it has a pretty small developer ecosystem at this point, right?
So it's, and the reason that, you know,
Nvidia has had such a success with that GPU proliferation is the fact that
Okuda will run on a software, software, software.
It'll run, I mean, it'll run on, on, you know, this college kids, you know,
the college kids graphics card, right?
Yeah.
That he's playing, he's going to be playing World of Warcraft on it one day and then
right and you know an AI PhD thesis on it the next right yeah yeah there's a strong argument
for like RTX 3 3,000s which are you know a couple of generations old because they
still do have plenty of RAM high speed memory and the tensor cores are fast enough
Jason I agree with you the when this whole TPU met and it started to hit the the kind of investment
market. I scratched my head thinking, like, yeah, AMD is probably way ahead of Google when it comes
to enabling software outside of its ecosystem. I always picture the TPU is being tied to vertex
and moving it away from vertex. I don't see this tied to vertex, but I mean, they've got their
own software stack associated with it and stuff like that. But I've used it in the past for some
deep learning activities. Again, mostly in Google Cloud and stuff like that.
But, yeah, that's my, that's my point is that it's tied to, when I say Vertex, I mean,
the Google services. Yeah. So they abstract the way. So you're going to, you know,
it's curated models, et cetera. And you might be using their GP, their TPUs. But that's way
different than saying, you know, going to hugging face, pulling down.
uh quant 16 model and then run it directly on a tPU that's a whole different level of software
enable yeah whereas you can just basically pull it down and run it on like you know any like in
video graphics card right now right and then and it wasn't until honestly and it wasn't until
amd came out with basically it was the 7900s the rtx 7900 as far as a consumer card
that had the capability of running the rockum stack on it right so and it's one of those things like
I said, the best way to predict how a lot of these applications are going to be built in the future is, well, if you give the kids, guess what? Those kids are going to graduate, you know, college. And then 10 years later, they're the director of software engineering and what are they going to use, what they do in college, right? And having the ability to have been, you know, doing this for the last 10 years on pretty much any NVIDIA card that popped out the door. They already know the stack, right? They know the stack. They know how it's built. And, and, and that's a,
That's been, you know, one, I'd say a big, when you were talking about the AI bubble as well,
and you look at Nvidia and its valuation, it's like you've got, like I said, all those, all those
kids that are now directors of software engineering that are going out and building, you know,
a lot of these, like, kind of AI platforming tools.
They're building it on what they knew.
And then Nvidia was kind of, you know, they were the first mover in that game.
When you look at the AI bubble, like, Nvidia has a market capital.
capitalization larger than the entire pharmaceutical industry yeah yeah well you know it pays to be
first especially with everybody in the world wants your technology yeah so keith you mentioned something
about what's going on with uh apple yeah so apple released rduma support over a thunderboat which uh which
for the Macs
Studio
with, you know, up to
I think it, I think it goes up to a
terabyte of RAM. I can't remember.
Oh God, yeah, yeah, on the studio.
Okay. But imagine, you know,
it's shared RAM. It's shared RAM between the GPU
and the PROS. And yeah,
the GPU and
the regular system. But
this isn't an amazing hack.
One, it's super
fast. I want to say it's like
833
gigahertz, something in that frame.
So they've seeded a bunch of YouTubers with a cluster of four of these, and they're getting
50% the performance of a H-200 and basically what equates to a $40,000 cluster.
But the problem is still, you know, I've watched these, and these are pretty good system
admins and AI
enthusiast. And the problem
is software. MLX, if
the model doesn't run,
isn't optimized for MLX, it just
simply doesn't run. And MLX
is Apple's, you know,
proprietary framework for
AI.
And, you know, we all go back to the
problem. Yeah, Nvidia, we don't want
Nvidia to win
inference like it has training,
but software has been
the barrier. Yeah.
Yeah. Yeah. It's, it, and it's been their, I don't know, moat, right? Their deep moat that surrounds their, their whole world to some extent. And it's, it's been a long time coming. I mean, Rockham's made significant, significant efforts to try to be more, more complete and such. And, you know, you can see that in the ML Perf numbers and stuff like that.
Yeah. And then it's also, you know, a big part of that, too, is also the, you know,
with hip, that's been a, basically hip and hip-sickle and stuff like that.
That's been a real kind of driver and being able to take, you know,
like kind of existing coup de code and move it over.
But then I think also it's like developments and stuff, you know,
in those software stacks where it's enable, enablement for basically the higher level languages, right?
So where you're talking like basically like Pi Torch?
Yeah, I mean, that's primarily most of the AI developments being done.
with PyTorch at this point.
So you know, I know AMD's put a lot of effort
into, you know, really kind of pushing the,
basically Rockman Pie Torch kind of communities
to pull together.
And then also when you're thinking about, you know,
spanning it out, you know, the RDMA and connecting,
when you got all these large language models,
you know, there's only so much you can fit
in a single system, but a lot of these models now
have actually like basically spanning,
across and that's what really the RDMA is being utilized for is doing uh you can take and do
model splitting across multiple yeah so this whole model parallelization right you can split the model
at different layers you can split the data you can do both i mean it it almost takes both when
you're talking trillion parameter models right yeah absolutely i mean you have to have you you got
basically be able to split it and then you get into all kinds of other fun things and
synchronization networking like it's like and it's the same stuff we used to have
have when we were doing newmo when we're doing like multi-core stuff right yeah yeah only across the
network now across network right so that's that's uh there there's like a thousand or one problems
multiplied by a thousand so yeah yeah yeah i keep you know keep people keep asking me are we in an
AI bubble is it is it going to end what's going to happen i i you know it's it's a tough thing
to to talk about but i think they're that we're starting to see some evidence that that uh
things are not nearly as rosy as they seem.
And when these big companies start going into, you know,
the debt market to try to fund their data centers.
And these data centers are huge.
I mean, talk about billions of dollars, you know.
And it's, it's a question of whether the market is there for it.
I don't know.
I have had a lot of conversations with,
I was actually having a conversation with basically a senator
and the executive director of the Department of Energy
about, you know, because the reality is, I mean,
there's such a power constraint, too, for this.
Oh, God, yeah.
I mean, those, you know, an example like the Helios rack
that's going to have, like, 72 GPUs,
and it's in the, basically the OCP ultra-wide form factor.
I mean, it's going to, it's going to draw over 250,000 watts of power.
and that's like a rack now imagine a data center full of those things right and how many like
you know like we're talking like the megawatts and gigawatts that are going to be pulled in these
AI data centers so there's a big concern about one the how do you build it how do you power it
how do you cool it right right and um you know and those are all liquid cooled uh pretty much all
liquid cooled systems they got to the power i mean from a power perspective too this is very
interesting. Actually, I found out, so Purdue University actually started the, uh, the very first
accredited program, uh, on, uh, basically building, maintaining SMRs.
Hmm. So I think the small modular reactors and there's, you know, it's been in the news lately
too, you know, they're reopening three mile island, right? Oh yeah. And Microsoft bought the whole thing, right?
I know. Yeah. And it's a yeah, they've got, they've got a spinoff company that says basically doing it. And, uh,
but it's going to require, it's going to require a lot of power.
I mean, I think the estimates that I heard were at 30% of the capacity we need to be at by 2030.
All right.
Well, we could talk AI for an hour and a half, I think, maybe longer.
The other topic of interest I thought might be, you know, what's happening with VMware?
I've been a tech field day guy for, God, I'd say not quite 20 years, but close on.
and VMware has always been really, really active in tech field day events.
And lately, I've seen Broadcom networking, but I haven't seen VMware.
What's going on there?
Yeah, I actually did a buyer room on this where I, you know, get a bunch of customers
and put them in a room with a sponsoring vendor and then have a conversation around
what these customers are doing.
And the insights that I'm getting from these customers, pretty clear.
There are three types of VMware customers, VNWR customers who are exceptionally hesitant to stay VMware customers,
VMware customers who have just said that, hey, you know what, I am going to bite the bullet and not worry about hyper, you know,
not worry about finding or replatforming hypervisors and just pay the VIG that, I don't know if that's a New York and Chicago thing only when you say big, but pay, pay, pay,
pay the tax, the Broadcom tax to stay a VMware customer via, you know, kind of the same
IBM Z type of setup. And then there's the third customer. That customer who is spending
$100 million a year in cloud, and they see VCF as a way to save, eventually save 20% of that
spend on-prem. So why not give Broadcom five million bucks a year?
That's the customer that Brockcom is now focused on.
And I don't know if that's the typical tech field day content consumer.
This is hot time calling up the CTO of one of these cheap, as Jason, you know,
alluded to one of these cheaply priced from a stock perspective pharma,
Fortune 50 pharma customers and convincing them to put all their R&D on premises.
on VCF cloud.
Right, right, right, right.
I mean, VCF's still a, still a package that it can be, you know, quite, quite easy to use.
Oh, no, it's not easy to use.
It's not, it is, it's not just a package.
It is the package.
Right.
If you want to buy, if you want to buy VMware v. Spear today by itself, you can't.
Yeah.
You have, and that was the Tech Field Day audience, like the audience that wanted to buy
v-spear. If you were to install your bits, what you're enabled for today from
Broadcom, it is going to, by default, install the entire BCF suite. You may not use it, but
it will install it because that is the product. I thought they have a smaller license
for just the V-SPIR solution. They just, it was a, I think it was called V-SPIR Foundation
or something to that effect.
They just announced that Europe,
that they've sunsatted for Europe,
is no longer available in Europe.
So that's the indication of where this is.
Brodcom wants the VCF customer.
If you're stuck and you want to continue to just operate your V-Spore estate
the way you have in the past, okay, fine.
You're going to buy V-CF, but you'll get to continue.
use just v-spear if you want, but you're going to buy VCF.
Everything. Yeah, yeah, yeah, yeah.
Well, I mean, and the value really is there.
I mean, if you're running a VCF stack and bought into the whole thing,
you can you could save money rather than running on a hypervisor someplace.
That's the theory.
Especially at that level of expenditure.
Yeah, that's the theory.
The insights that I've gotten from both folks to deploying,
the software and these executives is that it takes an entire reimagining and transformation of
your staff because now you have to run a cloud and running the cloud means that network storage
compute these silos security they can't be silos anymore they have to be they have to operate
as one team and that is a much much more difficult uh challenge
then how to install and get a cloud running.
It is operating the cloud that has been the barrier to adoption.
Yeah, I heard somebody the other day.
It was kind of made a funny joke on a podcast,
and they said, you can't just basically change a guy from basically an IT t-shirt
and put a DevOps T-shirt on them and have him figure it out, right?
And that's kind of what it, that whole, like you said,
the kind of that cloud shift is it's a shift in it's a shift in fundamental skills and a shift in
fundamental um you know honestly the the the way that a lot of things are run it's interesting
i was actually at cube con um was that like back in november oh yeah i think it's back in
november early december something like that and it was uh i was talking to stew at red hat right
and it's been really really interesting you know the uh you know he that they're you know also
seeing a shift of, you know, folks that are looking for it and how they're, you know,
taking those virtualized workloads and, you know, a lot of, you know, and having, you know,
conversations with him, a lot of the folks are looking at really kind of re-orchestrating their
things into a more cloud-native style of a fashion as far as their app workloads. So they are
getting more DevOps-centric in the way that they're looking at it. But then they're also
looking at, you know, the things like OpenShift, where you can do, you know, you know,
effectively utilizing cube vert to do virtual machine conversion on top of it as well to
move over to a different platform yeah yeah yeah well i mean red hat's been a been a player in that
space for quite a while you mentioned that the ibnz model it's sort of interesting i mean that
broadcom is is taking vmware into that mode of operation where you lock customers in and
you continue to provide value to a certain segment of the market and hopefully you're not
growing your actual overall market right here you're not trying to add to customers you're just
basically trying to basically retain current customers and put them in a headlock right yeah well yeah
and and it's working the renewal rates are exceptionally high like they're not getting small
businesses so you right when you see renewal rates between 70 80
percent people are like that's horrible for software but that 20 to 30 percent that are not renewing
are the customers that Brockcom doesn't want yeah yeah and you can see that that from basically
their channel strategy as well I mean what they did after the acquisition I think they cut like
3,000 channel partners or they cut 3,000 channel partners and then even some of the bigger channel
partners. They, they ended up, you know, dissecting relationships they had with Fortune 500s
and took all the Fortune 500 to direct. Yeah, yeah. Sounds like IBM more and more, actually.
Yeah, sounds like IBM more and more. It's interesting how they play, how those plays keep rolling
out left and right. All right, well, I think we played out the VM's discussion so far. The other
thing that's of interest lately is the price of hardware seems to be going through the roof.
DRAM is being significantly affected here.
Yeah, I think this hark is back to the original conversation.
We focus on these growing data centers.
These growing data centers, you're putting something aside of them.
Yeah, but it's compute, it's power, it's cooling.
I mean, there's lots of stuff in there, but it...
Yeah, but those components are the same components that I need.
You know, I'm looking at my invidia,
spark that's on my desk now and i was i made the argument that the stricts was a better value
at nearly half the cost because of the price of the components the the primarily ram the the
the strict is now the same price as the or if not more than the invidia spark yeah yeah and it's a
deram pricing in deram pricing i mean deram pricing is going up because there's like when you think about
You think about one of these standard AI boxes that's going in a data center.
I mean, the stuff that's, you know, kind of on the truck today and then think about the stuff that's, you know, coming next year.
But the stuff that's on the truck today, you've got, you know, you've got a system that's probably got one and a half terabyte of RAM in it just for the CPU component.
And then each one of those video cards, you know, is probably going to have anywhere from, like, you know, probably about, you know, the 96 to 192 gig, HBM on it.
And, you know, a lot of the kind of the retooling, I mean, that's basically where the big margin, that's where the margin dollars are and that's where the markup is, you know, in that industry, right?
Because you're going to sell one of those systems, you know, you're going to buy one of those AI systems that consumes like four years of space has got, you know, when you add it all up, like there's, you know, like three, four terabytes worth of memory in that thing, all the way from, you know, DRAMs.
basically the DRAM, the LPDDR, HBM, all of those components are going in there,
but it's going into a really, really high-margin product.
And because one of those boxes is going to sell for like half a million dollars.
And then you've got a data center full of them.
And you got probably, you know, like you got a minimum of eight of them in a rack,
and then you got a data center full of racks, right?
And so, but it's an exceptionally high-margin value product.
So where, if you were a memory manufacturer, where are you going to spend your, where are you going to spend your production, production time?
Like, what are you going to fab?
You're going to fab out stuff that's going to make you more margin dollar.
But usually that, you know, that drives a vicious spiral.
So it certainly starts to, as the volume production starts to go up, those prices should start coming down again someday, eh?
Well, where's the value of production?
Yeah, there's no more.
volume of the production. That's the problem.
Like, you know, our friends at Solidine are probably producing as many, as many, whether
they're big, their big drives, there are 122 terabyte drives.
Yeah, yeah.
They're probably producing as many of those as they can sell.
Yeah, yeah.
Yeah, it's going to be interesting because, I mean, honestly, I think this does open up,
this opens up an interesting market opportunity on ways in which you could, you know,
potentially do some type of memory
tearing technologies
from a software
perspective. I've been talking about memory
tearing for decade now.
Well, this might be a driver.
So we're saying this in the
enthusiast community,
XO with the
with these Mac clusters and these Linux
clusters, memory sharing.
They're loading these big
models and using basically
memory sharing technologies
to do that.
I see a few.
That memory between the GPU and the CPU makes it a little bit more effective, more efficient, I guess.
Yeah, and I've seen a few startups as well that are actually doing, you know, memory tearing from basically like faster NAND and flash components, you know, utilizing, you know, specific, you know, kind of kernel, you know, special kernel stuff that they can do to effectively create, take a system with like 64, like 64, like 64.
4 gig of RAM and basically make it 128 gig by basically doing, you know,
intelligent hot blocking of what's going into the memory and what goes on to SSD.
And this hark is back to the old skill that's coming an new networking.
Like, you need ultra-low high bandwidth networking for any of this to work.
Yeah, one of the things that impressed me the most about that spark box is the fact that it's
got a Connect X-7 with two 200-gig ports on it.
Yeah, and I've seen people, and they're limiting the last thing I read,
my friends at Kamawazza told me this is that they're limiting
Nvidia in the OX and the Spark OSS on the boxes,
limiting it to two boxes.
But I've seen, again, enthusiasts take four of them,
and basically it's hardware.
You can throw your own OS on it and basically go at it.
This is basically a 50-70 with 96 gig of RAM in each box.
Yeah.
And I've seen, I've also seen, oh, he was like Patrick Kennedy.
He actually took one.
He basically took one of those, you can get these microtick switches now that are like the QSP,
QSFP 56s and can hook like five of those sparks up together.
Yeah, those, that switch is like 500 bucks.
It's 500 bucks.
Yeah.
Yeah, 500 bucks for 200 geeks, which.
Yeah.
That's crazy.
That's nuts.
Thank God.
I think I need to have one in my basement.
You may have one in your basement.
I wouldn't be surprised if they start putting these in home routers.
I don't know.
It's just crazy.
Jesus, Jesus.
Yeah.
So the other thing, so, you know, the other topic of discussion was, you know, where is Ethernet and where is Infineban and all that stuff?
I don't know who, I saw Jason, you must have gone to supercomputing this year.
Actually, I did not.
Oh.
But a bunch of people on my team did.
So what's the story on the networking these days?
Well, it is interesting.
I think there's some more, there's definitely some more kind of open initiatives that are actually out there that are starting to, that are starting to basically make themselves present in the industry.
one of those is a UALink and like AMD kind of kicked off UALink and then what's actually happened is a lot of the
a lot of the other folks in the industry like hyperscalers and and folks like that have actually got on board with the whole UALink initiative and the whole point of it is to have more of an open standard for interconnection of kind of enrax
systems. So, you know, kind of a way where you can use like PCI fabric, but then hook it together. And, you know, it really kind of started out as like the, us open sourcing components of kind of the infinity fabric to do kind of interconnection of some of these components. And it's actually gotten quite a bit of traction. So with that, you know, it's basically like an open, more of an open version of kind of what Infiniband's been providing.
So, yeah, Infinity Fabric was your interchip within a chip.
And that's how interconnected these things have gotten, you know.
And I mean, it's, you know, it's like the Infinity Fabric is when you look at the like our MI 300-based systems where they're hooked into, you know, these OAM boards where there's, so there's eight of those instinct GPUs on, on this single board.
but they're all interconnected via that infinity fabric.
And, yeah, start talking about infinity fabric and you could put like a, you know,
a memory tearing solution within that structure and stuff like now.
I was talking real performance.
Yeah, and that's not the only option.
So think about if someone gets serious about the thing that consumer machines are starting to come with,
Thunderbolt 5, this is what?
Thunderbolt 5's bandwidth is something like 280,000.
gigabits per second somewhere i know it's over 200 gigabits per second imagine if there was a
thunderbolt five's networking switch yeah right right right now we're talking uh enterprise class
yeah yeah the uh on a stop that that so an interesting one to check out so yeah checking out like
it's a ua link consortium.org is the uh that's basically and they've got the uh the the 1.0 specifications up
there for that but yeah it's all about low latency high bandwidth interconnect um and like I said
it's primarily you know for for basically you know communicating into basically switching in
the AI AI compute pods that's actually what's going to be driving like that that helios uh
that big helio system that the the we're putting out with a 72 GPUs good and that's a for
use solution no that's that's a that's a rack that's a it's a it's an OECDs
OCP wide form factor.
So, I mean, the thing's literally, I think, three feet wide.
Ah, almost as wide as I am.
Yeah, but you know what?
I just need two more racks.
I have my, I took the rack from the, at the close down,
I closed down the data center earlier this year or last year,
I forget when it was.
And I couldn't get rid of the rack, so it's sitting in my garage.
I think I got a use case for it now.
You're also, you're also going to need a reinforced floor
because I can tell you to the thing weighs back as much as a
Ford F350.
Well, I have F-250 in the garage.
I'll just pull that out.
Pull it out.
You just mount it in there.
Oh, and don't mind the fact that you need a quarter of a megawatt to run it.
That might be the real problem.
You can have an SMR in the backyard maybe.
Yeah, it's right.
Well, they actually, they got the micro SMRs too as well.
And those will push about 10 megawatt, and you can fit three of them on the back of a
semi-trailer.
You know, I was just about to say I was on the road trip the other day.
I think I was headed to Tennessee.
I saw one with three on the back of the semi.
So absolutely, it was probably heading to Memphis,
and you can tell you can.
Yeah, you know where that's going.
You know where that's going.
But my wife was like, what is that?
And I'm like, oh, I know exactly what that is.
Why would you need a generator that big?
Yeah.
Yeah, basically to power the Tesla megapacks that are powering the colossus.
Yeah, yeah, that's bizarre.
All right, Jens, I think we're about running out of time here.
It's been great.
Any last comments from anybody?
Well, it's been, I think, a very interesting, very interesting 2025.
This is definitely like AI has made a lot of kind of fundamental shifts in, you know,
the way that we've been thinking about, infrastructure, data centers, interconnects.
It's been a phenomenal year of, I think, a lot of innovation from a lot of different companies.
And can't wait until 2026.
It's going to be an exciting year.
Yeah, certainly not stopping there.
Keith?
Yeah, I'm amazed at how quickly this has evolved.
I think just from a, not just from a training and how good the models have gotten, but how
how consumable AI has become for enterprises we're just now understanding kind of the failure
modes and the architectural patterns that that will scale 2026 I think is going to be a really,
really interesting year for AI adoption. Yeah, yeah, yeah. And then there's always the whole
discussion about AGI and what happens down the road from that. So all right, gents, this has been
great. Thanks, thanks for being on the show. Thanks, Ray. And that's it for now. Bye, Keith. Bye,
Jason. Bye, Ray. See you, Ray. Until next time. Next time, we will talk to a list system storage
technology person. Any questions you want us to ask, please let us know. And if you enjoy our
podcast, tell your friends about it. Please review us on Apple Podcast, Google Play, and Spotify,
as this will help get the word out.
Thank you.
