Utilizing Tech - Season 7: AI Data Infrastructure Presented by Solidigm - 06x10: Using AI in Modern Applications with Matthew Wallace
Episode Date: April 22, 2024As we consider the impact of AI on modern applications, we should consider the ways that this technology will improve these products. This episode of Utilizing Tech brings Matthew Wallace, CTO of Kami...waza AI, to discuss the adoption of AI in application development with Allyson Klein and Stephen Foskett. Large SaaS providers were the first to add AI-powered features but this technology is rapidly coming to market across the spectrum of applications. Matthew likens this to the evolution from spreadsheets to SaaS tools and cloud, which was a similar revolution. He mentions tools like ?, Cursor, Llama2, Mixtral, and more. We also discuss retrieval-automated generation (RAG), which enables LLMs to bring in external data at run time. We must also consider the source of the data, both in training and RAG, and questions of sovereignty, privacy, copyright, and safety. Looking forward, Matthew expects companies to customize their own models based on use case specific data. Looking forward, we expect new frameworks and models to be adopted rapidly to bring maturity and reliability to AI in enterprise applications throughout 2024. Hosts: Stephen Foskett, Organizer of Tech Field Day: https://www.linkedin.com/in/sfoskett/ Allyson Klein: https://www.linkedin.com/in/allysonklein/ Guest: Matthew Wallace, CTO and Cofounder of KamiwazaAI: https://www.linkedin.com/in/matthewwallaceco/ Follow Utilizing Tech Website: https://www.UtilizingTech.com/ X/Twitter: https://www.twitter.com/UtilizingTech Tech Field Day Website: https://www.TechFieldDay.com LinkedIn: https://www.LinkedIn.com/company/Tech-Field-Day X/Twitter: https://www.Twitter.com/TechFieldDay Tags: @UtilizingTech, @SFoskett, @TechAllyson, @KamiwazaAI, @TechFieldDay, #AIFD4, #AI, #UtilizingAI,
Transcript
Discussion (0)
As we consider the impact of AI on modern applications,
we should consider the ways that this technology will improve products.
This episode of Utilizing Tech brings Matthew Wallace,
CTO of Kamawaza AI, to discuss the adoption of AI
in application development with Allison Klein and myself.
Welcome to Utilizing Tech, the podcast about emerging technology
from Tech Field Day, part of the Futurum Group.
This season of Utilizing Tech is returning to the topic of artificial intelligence,
where we'll explore the practical applications and the impact of AI across the range of enterprise IT.
I'm your host, Stephen Foskett, organizer of the Tech Field Day event series,
and joining me this week as my co-host once again is Allison Klein. Welcome back.
Hey, Stephen.
It's always a pleasure to be here.
So last week, Allison, we talked to my colleague, Paul Nashawati, about the impact of AI on application development.
What did you think of that?
And where do you think we should go this week?
You know, we've been doing podcasts for a couple of years together, Stephen, and that talk with Paul has got me thinking in a way that very few other episodes have.
I thought some of the things that we talked about were really interesting as he was looking
at how are enterprises actually going to adopt AI and large language models and introduced
tons of new topics that I would love to continue to unpack here.
Yes. And the three of us, Paul, you and me, were part of AI Field Day back in February. And one
of the companies that presented there as a guest of Intel was a company called Camoaza. And so we
have decided to invite back on the podcast this week, Matthew
Wallace, to talk a little bit more about implementing AI in enterprise applications.
Welcome to the show. Hey, thank you so much for having me. I'm excited to be here.
So tell us a little bit about yourself. Why should we care what you have to say?
Well, you know, I was an early drinker of the AI Kool-Aid. So I've been, you know, doing
this now longer than ChatGPT has been around, which I think is what marks the beginning for a
lot of people. But, you know, I have a really broad set of experiences in technology, right?
And some of the folks who've been in IT for a long time, you know, I worked at VMware, I helped
design their first cloud product and launched their service provider program.
And I've worked for networking companies like Level 3, where I ran the cloud provider relationships on the technology side.
I've done multiple startups. And that kind of range and then diving into co-founding Kamiwaza, where I'm the CTO, I think gives me a lot of perspective, not just for the application level.
But what does this mean in terms of operations, in terms of infrastructure?
I think some of that thinking is certainly what led into the product that we're building and what it can do.
But also, I think it gives me a great opportunity to talk about what this means for the future of both software and IT kind of alongside each other. When you look at that challenge and you think about the enterprise
landscape, where do you think we are with enterprises really adopting AI in mass across
all of their business lines? And where do you think the industry needs to go to assist in the forward progress of that adoption?
Yeah.
You know, obviously there's a sort of bell curve of adoption that you find people doing.
And it's the same, I think, with this technology as with any.
And out on the leading edge, of course, you have like the technology companies themselves, right?
Hyperscale cloud providers serving customer needs.
Other businesses, especially large SaaS providers that could
muster the talent to be really early adopters. And of course, we've probably all read headlines
about folks like Netflix offering people $900,000 salaries to do AI work. And I'm sure that plays
a part in some of these businesses being able to poach talent. But outside of that bleeding edge,
what I find is that the vast majority of enterprises
are somewhere between rough POC tire kicking, and we're really excited about it, but we haven't
figured out what we're going to do yet. And so that's a pretty big range. And it's pretty
nascent if you think about where we think this will go. And I am a pretty firm believer that
it's going to touch almost
everything that we do as you got a number of years it's hard to say how many years but for sure i
think it's going to be really pervasive and so i think and this is back kind of backed up by surveys
but it's still early days but because this is going to be i'm very convicted
a huge huge impact on the industry I'm definitely recommending that everybody really
kind of jump in with both feet because the time to learn about it is now.
Yeah. I want to bring back what we talked with Paul about last week. He had a bunch of the
numbers that back up what you just said, and his numbers are absolutely in line with that.
Although the adoption of AI in the enterprise is expected to be very high.
Right now, it's still very low.
But I think that you professionally at Kamawaza and you personally,
as somebody who's bringing AI to market in all these applications,
I think what you're experiencing really shows what you said and what I said to Paul last week, which is basically when we next do this survey, those numbers are going to be very different.
How quickly do you think that AI and large language models are going to come to market across the spectrum of enterprise applications?
I actually want to draw a corollary to what happened in
cloud, because I think there's an interesting point to make here about adoption, right? If you
think about cloud, a lot of that was about how you deployed and scaled an application. And, you know,
it either was a new way to do it as a startup, or it was something that replaced traditional
infrastructure. But at the end of the day, it was infrastructure. Now, obviously, things like serverless took you to a place
where you were replacing infrastructure with an API
that could do the things for you with code that infrastructure would do.
So I want to draw a contrast to generative AI,
because if you think about gen AI,
I think it touches so many different use cases, right?
From the coding assistants to the advanced marketing
to the copywriting to the truly,, to the copywriting, to the
truly, and where we think this is going to go is to a place where we actually automate a whole lot
of communication and thinking and processing, really to kind of put the world of information
at your fingertips. I think it's really hard to kind of imagine how pervasively it's going to be.
I mean, it's hard to go back in time and think, what was it like when somebody did bookkeeping on a physical ledger, right? We
have had spreadsheets for so long. And I think that the kind of pre-AI, pre-generative AI world
will someday look very much like that. We will be looking back at today and going, it was like we
kept paper ledgers. But I think that what we're going to see is we're going to see patterns emerge
where somebody has a good paradigm for doing something really well, backed by tools that are available, whether they're open source or commercial.
They'll kind of evangelize something that they've built, a new way of doing things.
And then you'll see people get excited right across enterprises that go, we can do this too.
They'll kind of glom onto it.
They'll put their own spin on it.
They'll get it deployed and there'll be that evolution.
But it'll hit many different areas, I think, at different paces. And it's hard to say both when
that starts and when those different kind of things land, because it really is up to all
the individuals that are kind of working on those things. But I think you're going to see a much more
scattershot adoption, I think, to get to the final state, if there is a final state. I love your analogy of the spreadsheets and paper ledgers. I also think about calculators. You know,
I'm old enough to have had math classes where they wouldn't allow calculators. And now it just
seems somewhat, you know, archaic to do something like that. They just are part of life. One question that I have for you is actually about tools. You know, we're in early stages of adoption, but are there any tools on the market that have come out where you've been really impressed or have you seen any that seem to be gaining traction as being the leading adoption candidates amongst enterprise?
Yeah, boy. First of all, I mean, it depends on the kind of tools that you're talking about.
Right.
And because there's things get divided into different buckets here.
Right.
And there's different phases of the kind of life cycle of levering generative.
I'm right.
So there's things like
where am I going to get a model?
How am I going to deploy it?
How am I going to evaluate it? How am I going to evaluate it?
How am I going to build an application that uses it? You get into things like,
how am I going to fine tune it? How am I going to connect it to my data with RAG,
retrieval augmented generation, or DAG, data augmented generation, you'll also hear.
And I think those questions, they all have sort of different answers. I would say, you know,
for tooling, like from a developer standpoint, I'd say there's a really broad adoption,
a shockingly large adoption of, say, Microsoft Copilot, GitHub Copilot right now. But I think
what I'm seeing from the tools and the development side out there is unbelievable. And of course,
timely, this is, you know, only a day or two old as we have this conversation, but like the Devon launch from, I think it's Cognition is the name of the startup, right? Pretty amazing. It's like a baby AGI auto GPT type engineer,Sphere. It's a kind of AI first fork of
Visual Studio Code, right? These are fantastic tools. And I have no qualms in saying that I feel
truly like the set of AI tools that I use day to day truly makes me about 10 times, if not more,
as productive as I was before. I think I'm maybe like the perfect target market for it,
but I still think it's pretty incredible. And then by
not sleeping, I can do the work of 20 people. But it's really fascinating. And then if you flip over
to the other side on Gen AI, though, there's this incredible evolution going on. So Meta deserves a
lot of credit here for open sourcing Lama 2 in an almost completely open way, right? They kind of
carved out a few of their competitors. But there has been such an incredible evolution
of model fine-tunes, model architectures,
both on Lama and then also, credit to Mistral.
Mistral and their mixture of experts model,
Mistral have really driven just tons of evolution.
You see lots and lots of models
getting closer and closer to the big private models.
And so I think the set of tools has never looked better.
And yet, if you think about where I think it could be,
I think there is so much research coming out.
It's hard to even comprehend.
I've read more research papers in the past year
by a factor of 10 than the whole rest of my life
because the research is absolutely breakneck.
So I feel totally confident
in saying the best is yet to come on both those accounts. And so you start thinking about what
this means for us day to day, right? What is it going to look like when an AI can independently
go and introspect all your cloud infrastructure, turn it back into CloudFormation templates or
Terraform's configs that you can just go and
apply on a new provider or even swap between things like Terraform and CloudFormation.
You know, what does it look like when you can say, I need an app that does this and
you walk away and come back 12 hours later and it's, you know, looked up hundreds of
web pages, read docs, you know, coded for hours and hours in a sandbox, it's done.
You don't have to touch anything because it can see the screen. It knows what acceptance criteria look like.
We are not to that place yet, but I think that's low hanging fruit. Like this is not the far
distant future. This is something that's coming very soon. So I think it gives you like a broad
overview of how crazy I think things are going to get. Yeah. I want to zoom in on RAG for a second
because that came up last week and it's come up a few times here. And frankly, I want to zoom in on RAG for a second, because that came up last week,
and it's come up a few times here. And frankly, I'm really excited about that, because I think
retrieval automated generation is how stuff should be done. I'm sorry. If you're just doing
generation, where does the data come from again? I mean, don't you want people to go look stuff up?
I mean, that's what my teacher told me. How do I, you know, how do I know something? Look it up. You know, it seems as that just becomes how stuff is done instead of
a specific trend. That's just sort of the right way to do it. Am I off base here, or do you think
that that's just one more trend? No, I think that you're spot on. I mean, I will say this.
I think that I'm going to go devil's advocate in a small area, but then I'm
going to bring it back. The truth is we are very, very, very naive, like across the industry about
how to kind of incrementally and easily fine tune a model, right? So the idea that you could take
ongoing information and kind of slowly blend it into an existing model or a fork of an existing
model, right? That could be specific to a group or a user or a use case, et cetera,
inside of a specific enterprise. We're nowhere near, you know,
tapping out on that. In fact, people are really far.
That's a further step down the journey than rag.
And so we may see some rag use cases shift to that.
And the reason why is because if you can integrate the knowledge or the base
model,
then if you can accurately generate and you don't have to go do lookups, you're basically fine
tuning it once to get that knowledge integrated. And then you get all the inference is exactly the
same forever. Inference cost doesn't change once you've blended that knowledge in, but it's
trickier than it sounds. Now, the reality is you still need RAG, because the truth is there's a limited amount of information that a model can hold. And that's
especially true for smaller parameter models when you fine tune them. And so being able to kind of
tap into any specific thing and just use it on the fly, it's always going to be a thing to some
level, right? Whether it's about freshness or whether it's about the long tail of data, you know, having talked to people who have over a hundred petabytes of data,
there's no way you can't meaningfully fine tune that into a model. It could never learn it. It's
just not capable of holding an information today. But there's a really interesting angle because,
you know, retrieval augmented generation, we know comes from taking a lot of documents and
information and we generally make
it retrievable searchable right classic ways you create embeddings but those embeddings in a vector
database one of the kind of naive look how early days it is is like this growing evidence that you
know vector database is not all an embedding is not always the right way to do things like classic
pm25 you know bag of words reverse index searches can be better. Sorry to get all crazy
technical. But one of the things you realize though, too, is that RAG is a very close cousin
of memory, right? And having an AI be able to do complex things requires you to prompt it over and
over and over again. Like speaking of Devin, right? This engineer, well, it's off coding for 16 hours. How many times do you think it runs an LLM? Thousands, tens of thousands, a hundred thousand.
It's really kind of unbelievable how many times you're going to have to prompt it. In fact,
even if you have a fairly simple query, you actually get much better results. If you can
prompt the model over and over again, you'll say, give me your top three answers to this question.
And then you'll go back to it again in a totally new conversation with no history and say,
which of these three best answers the question. And then you'll change the order of all those
things, right? There's a lot of techniques, but they require you to kind of go back to the well
over and over again. All those kinds of things require you also to store those kinds of
intermediate memories, right? If you say, go code something and you want an agent to come up with a
plan and then flesh out each step of the plan and then execute step-by-step,
that requires a lot of data. And that's a form of RAG. You're kind of generating it with the model
and then also retrieving it. Then of course, there's like the classic RAG, right? Where you're
going and looking up customer records or FAQs or technical documentation, which is immensely
valuable, honestly, because AIs are really good at sifting that out much better than us kind of searching bad search pages.
So that's powerful all in its own right. And yet there's like so many different ways it's going to branch out. How all of these objectives across line of business and IT start hitting things like privacy policies, ethics policies, different things that usually aren't associated with the core of IT. around how to implement these new technologies while being mindful of some of the decisions and guardrails that they put in place and how their organization holistically runs?
I'm going to say I love that question because, honestly, I don't feel like it gets asked enough.
And one of the things that I saw was that two years ago, one year ago, even before we founded Kamiwaza,
I saw a lot of enterprises really struggling with the exact questions you're asking.
It doesn't even need to be generative AI. If you want to have a data scientist and data analyst, there's data all over the place.
It's in disparate catalogs or sometimes no catalog.
You know, sometimes it has a schema, but sometimes it's a parquet table that's literally labeled field one, field two, field three of the columns.
And some poor data scientist is literally trying to disentangle what it means without having any sort of like reference answer key.
And I think those are big challenges.
Access control is a huge problem.
And it's a little bit intermediated if you're purely at the cloud, because a lot of people, you know, find some easy cloud success, just leveraging like IAM type, you know, functions to control access. Databricks does
really well there too, right? Snowflake does really well. But when you start getting into
these cases where you have information that's all over the enterprise, and there's sort of
different rules, different access controls, different security levels, sometimes it's in
different countries, there's different regulatory regimes that apply to it, right? Those become very, very large challenges. And I think at
Kamiawaza, that was something we were really keeping in mind while we were thinking about
our distributed data engine. The fact that you as a developer, as a data scientist, you don't want
to wrestle with those things. And even if you know the answers to some questions like, who should
have access and can we move this data out of the country? Is there a tool set that really makes it easy to implement that? It's not great. And
certainly there's nothing specific to generative AI. And I think that really drove us to think
about how we can make that easy. I'm really passionate about seeing enterprises be able
to adopt this technology. I was using AWS in 2007, right before EBS volumes existed. And it's like, if I try to play that timeframe for the slow adoption of cloud on Degenerate
AI, I just feel like humanity is missing out, right?
And so we're really driving to make it easier because honestly, I want to see enterprises
running trillions of inference calls so that we can just automate all the kind of grunt
work of our day and get on to doing bigger and better and more powerful things over time. It's interesting that you bring up the
question of sovereignty, though, when it comes to data. What about model sovereignty? I'm curious
about how that's going to hit us in the future. We've already seen a lot of instances where
companies, you know, I'm not trying to throw stones at them, but some of these AI companies have been less than forthcoming about the training data that they used. end up having to use different models in different jurisdictions based on the laws of, for example,
the EU versus the US versus Saudi Arabia or someplace like that. I mean, very different laws,
very different legal scheme, and very different acceptance of training data.
Yeah, that's such a fascinating question. And I don't even know, by the way, that I've never
fact-checked this outside, but I did have a colleague that was doing a lot of work on AI for a particular use case way pre-generative AI.
This is more classic, although it was still neural network driven. And he actually told me that to deploy their model into China, they actually had to train it in China.
And they actually had a whole separate training process to also train the model in China.
What we see going on right now with the AI Act
being passed to the EU, another fantastic example, right? And it's frankly, very early days. I think
that there are going to be some very serious pitch battles over everything from, you know, model
safety to regulatory regimes that need to be in place for all kinds of, you know, risks, real or
perceived. I think the question of privacy,
the questions about copyright,
which is like really dramatic,
gets a lot of press,
probably kind of hashes out to not a lot, I think,
but it's still, there's a lot of risk factors.
Even beyond those things,
which are the big like prime mover,
like nation state problems.
It's like, one of the things that we bring up is like,
let's say you deploy a model off of Hugging Face.
Let's say you wake up tomorrow and somebody replaced a file, changed the config.
Maybe they thought it was a bug. Maybe they changed it for their own reasons.
But then your whole stack collapses, right?
And like that's one of the things that we tried to solve really early on was kind of providing some of that operational stability.
But I think it's a great question and it's one of those things you don't necessarily spend a ton of time thinking about when you're in that, like, let's kick the tires
POC phase. But I think that number one, people are really going to end up, you know, wanting to
fine tune models of their own as they kind of grow in maturity, right? Because it's going to
fit their company, their use case, whether it's tone, whether it's the guardrails, whether it's
the use case, you could just get so much
better performance. Like fine tuning for a specific use case makes models almost superhumanly
good at that, right? And so you can actually surpass something like a GPT-4 when you're
training on that use case. We're going to see a lot of that. The skills are just not there
for people to do that at scale yet. But then you get the win, right, of you're going to be able to do much cheaper inference
at larger scale, have full control,
you know, actually derive, you know,
the changes that you want to see in the model behavior.
And so I think guaranteed,
you're going to want to see some of that model destiny.
But then to your point, as you have that model,
there's going to be questions of,
are we getting it commercially?
Is it open source?
What do we have to do? You know, you, even as a large enterprise, if you're multinational, you may going to be questions of, are we getting it commercially? Is it open source? What do we have to do?
You know, you, even as a large enterprise, if you're multinational, you may have to have
different copies of the model that are trained in different ways with different data because
you're not allowed to use data trained on X, Y, and Z when you put the data in the EU.
Like a great example there, will you be able to train on data that you got based on the
customer records in the EU and then push the model over to the United States? Even if you say, oh, it's not personally identifiable, you could see a regulation
where they say, no, no, it's not safe because you might accidentally leak data that would be
protected, which is very easy for me to picture them making that claim. So yeah, great question.
Definitely a lot of challenges and a lot of unknown unknowns ahead, I think, in that arena.
When you were talking, it made me think of,
if we have multi-jurisdictional models, do we get into regional truths based on different AI models?
And that gets very dystopian very quickly. When you look at 2024 and you look at some of these
topics that we're talking about, things are moving so fast.
Do you think we're going to be still in the same place of enterprises looking for implementation and still, you know, seeking those tools, seeking the path forward?
Or do you think that we're going to quickly move into this feels old hat?
I think both because that whole bell curve concept, right, there will definitely be some laggards and we'll actually be having that like much as we said, cloud leader, cloud laggard,
you know, so many years ago, that same kind of paradigm is going to apply. On the other hand,
I would bet, you know, dollars to donuts that there's going to be a couple of things that just
drop your job this year. Like, you know, the figure being a robot being powered by open AI.
I knew that was coming because I follow them both really closely. But I think that type of thing where you go, wait a
second, like what is really possible here? And your mind just kind of explodes. We're going to
see a few of those things. We're also going to see just a really huge rank and file improvement
across the board, you know, in frameworks, in tooling. We're obviously trying really hard to
contribute to that at Kamehwaza,
but I think that there are just so many tools
and so much research.
And actually, so many things
that I've read about in research papers
aren't really, quote unquote, in the wild yet.
And so I think it's going to be a breakneck year.
Do not expect it to slow down for sure.
But by the same token,
you're going to see the maturity extend
and advance quite a bit.
I absolutely agree with you, Matthew. I feel like the world is going to change dramatically in 2024. The thing that
strikes me is, to your point, that we spent 2023 sort of playing with a new toy of, you know,
let's see what we can do. Let's see the cool stuff. I mean, I'm not seeing as many
people posting on social media, look at this funny thing OpenAI wrote, right? Look at this funny
thing ChatGPT wrote. Look at these funny pictures as sort of an end discussion. Now, I think what
we're going to see is people say, okay, that shiny new toy, you know, we've played with that now.
Now, let's see where it goes from here. And I do think that there's going to be some radical new use cases. I think it's going
to be implemented everywhere. And as I said to Paul last week, and I think that what I'm hearing
from you too, is that these models and frameworks are going to be deployed all over the place in
2024. And we're just going gonna see bang, bang, bang,
new applications and new use cases for AI this year
that are gonna just blow us away.
And that's pretty cool.
Pretty excited to see where that goes.
So thank you so much for joining us
on the discussion today.
It's been a short discussion.
I mean, this is a short podcast.
Where can people continue this conversation with you? There's lots a short discussion. I mean, this is a short podcast. Where can people continue this conversation with you?
There's lots of great places.
I myself do a podcast called AI Every Day,
and people are welcome to check it out.
I'm on LinkedIn.
You can follow me on Twitter at Matt Wallace is the handle.
So it's easy to find.
Of course, please check out our website,
kamiwaza.ai, which is fantastic.
And I'm also going to be presenting on the topic of enterprise AI at SW2Con,
which is from the organizers of GlueCon.
It's this May.
In fact, you can use the code SPEAKER15 to get early word registration
if you're catching this early enough, 15% off that.
However you find me, I'm pretty easy to get a hold of, but reach out.
I mean, I definitely do love talking about this. I like to tell everybody, this is the thing that wakes me up
at five in the morning and I can't sleep because it's just too exciting. So I feel really privileged
to kind of find something that is so awesome and so motivating and exciting for me personally,
right as it seems to be the tsunami that's just going to wash over the industry. Absolutely. Absolutely. And of course, we got to hear a presentation on Kamiwaza at our AI Field
Day event, which Allison and I were both part of. So I would recommend people search your favorite
search engine for Kamiwaza and AI Field Day, and you'll find those videos. Otherwise, you can go to techfielday.com
to see them. Allison, what's new with you? I'm still publishing on the techarena.net,
Steven, on AI, 5G, and from the cloud to the edge and everywhere in between. I'm going to be at GTC and at MemCon.
So you can check my page out for the latest in terms of industry innovations.
Thanks a lot.
Yeah, I do enjoy the podcasts as well that you record with notable people in the industry.
And I urge people to check those out too.
Thank you for listening to Utilizing AI, part of the Utilizing Tech podcast series.
You can find this podcast in your favorite applications
as well as on YouTube
if you're interested in seeing the video.
If you enjoyed this discussion,
please do leave us a rating or a review or a comment.
This podcast is brought to you by Tech Field Day,
home of IT experts from across the enterprise,
now part of Futurum Group.
For show notes and more episodes, head over to our dedicated website,
which is utilizingtech.com, or find us on X Twitter and Mastodon at Utilizing Tech.
Thanks for listening, and we will see you next week.