In The Arena by TechArena - Where AI Inference Hits the Memory Wall
Episode Date: April 28, 2026ZeroPoint’s Nilesh Shah explores why data movement, compression, and memory bandwidth now shape AI inference performance, and where heterogeneous systems and quantum may fit next....
Transcript
Discussion (0)
Welcome to Tech Arena, featuring authentic discussions between tech's leading innovators and our host, Allison Klein.
Now, let's step into the arena.
Welcome in the arena.
My name is Allison Klein.
This is a Data Insights episode, so that means Janice Narowski is with me.
Hey, Janice, how you doing?
Hi, Allison. I'm doing well.
How are you?
I'm doing great.
It's the first day of spring.
Very exciting for those of us living in the Pacific Northwest.
because it means that at some point the rain will end.
This is an exciting episode,
and I just wanted to ask you who you brought with you
and what the topic is for today.
Yeah. So today I brought a friend of my Go Away Bath.
I brought Nalesh Shah, who there's a lot of great things.
You probably see him all over if you're not on LinkedIn.
But Nalesh is the VP of business development at zero point.
So we're going to talk a little bit about what do you?
he's up to it zero point, what he's really working on in terms of engineering systems and just
get his take on AI and how do we scale this stuff? So, Nalesh, welcome to the program.
Thanks for having me, Alison and Janice. This is so much fun, always fun to chat with you.
Now, Nalash, I see you all the time at conferences and you're always in the heart of where the
industry is focusing its most acute innovation and you're always on to the next thing. A couple
weeks ago I saw you in meshed and quantum. And I know that's a passion that you're going to talk about.
I know that you work in the memory and storage space as well. But why don't we just start with
0.0. Point. And can you introduce what you're doing at 0.0.0.0. Technology is a company
based out of Sweden. It's a startup and it's about compression. The company was founded on the
premise that the need for data and memory will explode. And one of the ways to tackle that
challenge is through lossless memory compression. So what I do at ZeroPoint is I drive business
development at ZeroPoint Technologies and then bringing and matching of the technology to pretty much
across the spectrum of hyperscalor chip companies. And ZeroPoint Technologies is an IP licensing company.
So we develop IP that can be integrated into any piece of silicon. So it could be an ASEC chip,
I think inference or it could be a GPU or it could be a memory controller,
which could be, for example, a CXL controller.
But right now, our key focus is, of course,
AI inference systems where the need for memory and the pain is real and people are feeling it.
Now, Lash, at a high level, given that information you just gave us,
at a high level, why is this data movement becoming so critical for you,
specifically with modern computing systems?
What's the big deal in your opinion?
The big deal out here is if you look at what it takes to null data all the way from storage out to, let's say, high bandwidth memory or LPDDR memory, and then into the silicon itself, into the heart of the silicon.
So like an S-RAM memory. And if you compare that movement of data to the actual computation that you perform on that dealing without data, it's ridiculous.
It's almost 10x more expensive in terms of power to move them.
that bit of data than it is to actually run a computation on it.
So from that perspective alone, that's why if you look at all the insurance chip innovators
out there and look at cerebrus, look at Brock, look at Sambanoa, rebellions, I can rattle off
tens of different names.
But if you look at the core innovation of all of these companies, it's not really about
the compute as much as it is about data movement and memory hierarchy. For example, GROC is
all about S-RAN. We've seen with Cerebras, it's all about WIFA scale S-RAM. If you look at other
companies like someone over there, tearing the memory. So really what that screams out at me is
the pain point is clear. If you want to scale inference, it's all about the memory hierarchy.
That's where zero point comes in by compressing that memory and enabling greater bandwidth
and also capacity, which is what you need if you're about cranking.
call more tokens per second per walk.
And you have to do it efficiently because, like I said,
the data movement is the expensive piece compared to the actual computer.
Nalish, when you think about the fact that data is distributed across data centers,
both in the cloud and on-prem at the edge, and you talk about that data movement,
do you think that we're constructing AI workflows correctly for that management of data?
and how do you see that changing as enterprises start scaling inference into different areas of their business?
I think one key thing about is we look at the problem we're trying to solve.
It's really agentic AI is entering the workflow.
So, Alison, we were at the Synopsis Conference just last week, and then before that, I was at D.B.Con.
So when you think about EDA tools that people require a lot of memory and a lot of data to rank out new technology.
chips, what we are finding now, or at least what I heard from these conferences, is now these
chip designers want to integrate five, six, maybe even ten different LLNs, all focused on a very
specific task, for example, chip verification or error detection, or think of all kinds of
RTF verification flows. So it's becoming obvious that now you have these domain-specific
inference solutions that are required for performing, let's say, even BDA operations and the same
thing. So at that point, what happens is the way enterprises will need to think about data.
You've got to need the data that you need to design the chip, but then you're going to need
this extra intelligence there as well in order to enable agentic AI to make that process
even more efficient.
So as data continues to become a challenge, and certainly, as you said, the compute is important, but the data is becoming more important.
As you see new computing paradigms emerge, right, where do you see memory bandwidth becoming a bottleneck?
Or do you see that inherently becoming a bigger problem in the future, and will there be a new change to address it?
Yeah, I think with these agentic workflows, if you break down inference, really take any transformer model, and there are two sections to it.
There's a pre-fill stage and there's a decode stage.
So the pre-fell is pretty compute intensive, but the decode phase is extremely memory intensive.
So if you break it down further, it's really what's limiting your tokens per second.
And if you need instant answer, responsiveness at scale across the enterprise.
I'm not talking about one user interacting with one chat agent, but I'm talking about an entire
company of, say, 100,000 people interacting at that scale and then across the multiple
streams of data that you might have in an enterprise.
It's really that decode space, which has become the bottleneck.
And at GTC, that was pretty evident if you look at what the key notes were talking about,
it's really breaking down that decode into heterogeneous architecture.
So it's not all about GPUs anymore.
You need these highly specialized P6s that can do the decode phase extremely efficiently.
And that's where with compression, you can actually get more efficiency.
If you can compress that data by, say, 1.5x or 2x.
Now we are moving lesser data.
And at the same time, what that does is it increases your band.
and the effective capacity that your decode engine actually sees on the silicon.
I think those are some of the trends shaping the industry.
It's going to be heterogeneous compute.
It's got to be pipelines.
Inference is no longer just a single kind of operation.
It's got to be a pipeline going through heterogeneous process node.
And all of that needs to be done efficiently.
Now, you talk about heterogeneity, and obviously that's been something that the industry is
looking at with all of these different accelerator technologies and how we deliver
Bruntforce compute to some very challenging workloads. I guess the other question that I have
for you, because I do know that you are diving into the quantum space, is when does conventional
computing platforms run out of gas and we need to move to alternative architectures like quantum
computing? Where do you see that line? Is there a line? I guess where is it? And what
types of workloads and challenges is quantum uniquely suited to address?
Yeah, I'd say maybe let's talk about where does quantum enter into the picture.
So really the big question is, first of all, is quantum computing required?
What is the chat CPT moments for quantum computing?
That's the favorite question I like to ask.
Where does or does quantum computing intersect EI of data center?
because the AI data center is currently in the trillions of dollars, the market.
If you look at quantum, it can be measured maybe in the billions of dollars.
So obviously, the immediate point at which an attached point would, that would make sense,
would be to attach quantum processing units into these AI data centers to efficiently
offload some of the work.
So it may not replace the entire, let's say, inferencing, but there might be some efficient
search space algorithms, for example, that could be offloaded to quantum processes.
Yeah, I agree. And I think you're right now, as we're in the early throws of it, right?
What is quantum computing? What does it mean? What does this ship have to take place? Is it next
year? Is it five years? Is it 10 years? I'm just curious, from your perspective,
and all the different companies you've been working with and learning from and consulting with,
What industries will quantum computing really impact?
So going back all the way to Intel, Genes, when we work together and if you remember,
we worked on a project Optane memory technology.
And at that time, this was back several years ago.
At that time, quantum computers were still at the toy level, there were like five, six,
cubit.
Cubit is the measure of quantum computing as you have bits.
And at that time, the big focus was on simulating quantum.
computers, for example, of 50-cubit simulation required 2 to the power 50 memory capacity.
So that's where we worked on opt-in.
And the very first use case actually was with Alan Astero, music was a leading quantum chemist.
At that time, I'm at Howard, but now he's got huge three-digit million funding at the University
of Toronto to investigate quantum chemistry.
So now, fast forward, even now the use cases that I hear about.
are quantum chemistry.
That seems to be a natural fit and drug discovery for the types of processes that quantum computing
is designed for.
And because a lot of these mechanics in the quantum chemistry space are quantum in nature.
So they seem like a good fit.
Now, beyond that, of course, there are other applications which are non-data centers.
So I think space, those kind of applications where you can use quantum.
devices, but of course, I'm biased towards the data center.
So I keep coming back to what are those use cases where quantum can help in the data center.
I strongly believe the unlock of potentials of quantum computing is primarily in the data center.
Of course, there are really interesting use cases for space.
For example, sovereign type of capabilities where quantum effects can help develop like a secure type of
devices that go into space, but I'm looking primarily at data center and the application.
So right now, I would say quantum chemistry, there's maybe some finance applications that I saw,
at least some of the banks are deploying early quantum computers, primarily to test out what a
slow would look like if you were to have serious quantum computing capabilities.
And then maybe a third use case might be for encryption and making even more secure,
encryption protocols because we know people are actually hoarding data on the internet today
in the expectation that in five years they'll have a quantum computer. So your data might be
encrypted today, but in five years, if they gather all these data, they can decrypt it five
years later. So I think that becomes an interesting use case for quantum to actually come
up with stronger encryption. Now, no, actually you said something really interesting earlier about
it's about the data. And I want to go back to conventional for a second. I know that you're working
on the lunatic fringe of what can be done with memory and storage. When you think about the fact that
we've got leading edge SSD technology, like what Solidime is delivering to the market, we have
standard DRAM, we have HBM, we have S-RAM that you talked about. Do you see anything on the horizon
that will break through some of the bottlenecks that we're seeing and introduce a new way to manage
memory and storage, and where in the industry are these things being discussed?
I think a couple of things. We're seeing a lot of innovation or at least funding going into new
innovation, just simply because memory and actually even storage are sold out. So looking for
denser solutions for storage and then denser solutions for memory. I think that's where
a lot of investments are going alternative technologies to DRAM. But DRAM was invented in several
decades ago, it's not fundamentally changed that much. And so people are looking at new technologies,
you know, think Engram or other persistent type of memories that may be denser and may have
better power characteristics and also may not be limited by some of the manufacturing limitations
that are actually now showing up because of the heavy demands from AI. I think there's maybe three
different vectors. One is looking at alternatives memory technologies. The other vector is looking at
alternative interfaces, so how you get those memories to communicate to the compute engines. And then I
would say entirely new paradigms in terms of networking, not just because now the conversation is
not about a single chip or even a server or even a rack. I would argue now it's really about
data center scale. So what I'm seeing companies do is, now, I don't care what do you put in your
memory chip or your storage chip because the whole thing moves as one unit, like one data center.
Or I would say, what is the new unit or the metric for compute? It's going to be megawatts.
So I give you one megawatt. You put whatever you want in it, storage, memory, compute, and then
you crank out these AI tokens.
So I think that's how a lot of memory and storage innovation will be driven.
And a lot of the things we know today, like phone factors, let's say an SSD forms factor on a PCI card or say DRN memory going over LPDDR or HBM.
There might be some completely dramatic new technologies where you'll be thinking at data center scale rather than chip or server or even racks scale.
I think those are the three vectors.
I see a lot of innovation or at least startups looking into these spaces to come out with a breakthrough.
And as these technologies evolve over time, what do you think are people's biggest misconception about it?
And what will it take to make next generation compute systems actually usable in practice?
Yeah, yeah.
I think a couple of things.
One is, like I said, in terms of you mentioned misconceptions.
I think one misconception is that in order to produce more tokens, you need more power.
So it's like we don't have an infinite supply of power.
So that's one misconception that, oh, well, AI is going to grow.
So let's just keep putting gigawatts power.
So I think that will be one of the, I expect, a surprising breakthrough is someone will come up with an entirely new style of physics,
which will break that linear kind of assumption that people have
is going from 100 LLNs to a million
or going from a million users to 100 million,
just multiply the megawatts of power.
I think that may not be valid in the future.
Something will have to fundamentally change in the physics
in order to break that assumption.
Belash, final question for you.
When you look at Accelerated next week,
can you look at the companies that are gathered there from foundational component players to
infrastructure providers to neoclouds and service providers and others that are prognosticating
on what's happening next in AI? What do you expect to emerge in that conference? And how do you
articulate your own vision for where AI will go by the end of 2026?
So, Chair, this is probably a fun concern. Like you said, neoclots, hypers, neochip companies,
even quantum computing companies, they'll all be at the same place and having the same conversation.
So I think a key thing that I hope will emerge from here is clarity, actually, for,
and by the way, there will be enterprise end users at this conference and the storage innovators like
Sauridime and others will be there. So I'm expecting that people walk away with a better vision of what it takes,
what are the components and ingredients that go into making an AI system?
And how is that different from just building your traditional data centers?
We've been doing data centers for you.
So what's different now?
I don't think everyone clearly appreciates that.
Besides, yeah, you might have heard of GPUs.
But there's a lot more like we discussed.
There's memory, there's storage.
And these things need to work in synchrony.
Otherwise, you're not going to have an optimal system.
And then even the compute itself is getting disaggregated.
It's getting heterogeneous in nature.
So I think people will walk away with that understanding.
And the fact that not all tokens are created equal,
one tokens for a text prompt is not the same as a token for, let's see, a video generation.
So what does it take to support each of these different types of tokens
and different use cases are not all mornings are created equal?
Now we're saying for enterprise, you may not need,
these gigantic hundreds of billions of parameter models, but you might need tens of really small
models that are highly efficient for laser-focused tasks that you need for enterprise.
So I think all of these conversations will emerge.
I'm excited about the conglades.
A lot of people may not be going that are listening, though, to FNELESH.
So hopefully those sessions will be recorded.
But where can folks learn more about you and maybe reach out to you and get more information
just in general.
I think one great place to start is, of course,
you can reach out on Zero Point Technologies on my email
and then also via LinkedIn.
And then, of course, if you're at these conferences,
love chatting with people and learning from them.
So, yeah, that's always an option.
But yeah, I hope people reach out and exchange ideas.
Awesome.
Thank you so much.
It's always a pleasure, Nalash, to have you on the show.
I'm going to be accelerated as well.
The Tech Arena team will be there.
I know, Janice, you're going to be there as well.
Should be a fantastic conference in New York.
But with that, I just want to say this is another episode of Data Insights,
and it was a pleasure to have both of you with me on this journey.
Thank you, Alison.
Thank you, Nalesh.
Thanks for once.
This was fun.
Thanks for joining Tech Arena.
Subscribe and engage at our website, Techorina.
All content is copyright by Tech Arena.
