Utilizing Tech - Season 7: AI Data Infrastructure Presented by Solidigm - 2x02: The Last Mile - Connecting AI to Enterprise Applications with Monte Zweben

Episode Date: January 12, 2021

There are many “last mile” items on the enterprise checklist, and companies are struggling to connect everything together. In this episode, Monte Zweben, CEO of Splice Machine, discusses feature s...tores with Andy Thurai and Stephen Foskett. Data engineers maintain data pipelines, data scientists maintain the data store, and machine learning engineers are trying to create models and package them so they will be useful. One idea is to store a model in a relational database, store records in a feature table, and enable the database to trigger a model based on this data. That’s what Splice Machines is implementing - in-database ML deployment. SQL is making a comeback in ML, with scale-out solutions providing a more familiar and usable environment than leading noSQL databases. Monte believes that SQL will be the dominant data paradigm for machine learning, modeling, experimentation, and deployment. After all, SQL is the dominant language of enterprise data scientists. Guests and Hosts Monte Zweben is CEO of Splice Machine. Find Monte on Twitter as @MZweben and Splice Machine as @SpliceMachine. Andy Thurai, technology influencer and thought leader. Find Andy’s content at theFieldCTO.com and on Twitter at @AndyThurai. Stephen Foskett, Publisher of Gestalt IT and Organizer of Tech Field Day. Find Stephen’s writing at GestaltIT.com and on Twitter at @SFoskett Date: 1/12/2021 Tags: @SFoskett, @AndyThurai, @MZweben, @SpliceMachine

Transcript
Discussion (0)
Starting point is 00:00:00 Welcome to Utilizing AI, the podcast about enterprise applications for machine learning, deep learning, and other artificial intelligence topics. Each episode brings together experts in enterprise infrastructure to discuss applications of AI in today's data center. Today we're discussing the last mile. Basically, how do you operationalize AI in the enterprise? First, let's meet our guest, Monty. Hi, I'm Monty Zweden. I'm the CEO of Splice Machine,
Starting point is 00:00:36 and I'm really excited to be here today and talking about that last mile. And I am Andy Thorai, co-host of this podcast, founder and principal at thefieldcto.com, home of the Unbiased Advisory Services. Check us out at thefieldcto.com and on Twitter at Andy Thorei. And finally, I'm Stephen Foskett. I am the organizer of Tech Field Day and the publisher of Gestalt IT and the host of this podcast. So you can find me on Twitter at S Foskett. So Monty, this is really right to the heart of what we've been talking about on utilizing AI since the very beginning. And essentially, how do you bring enterprise, how do you make enterprise AI real? And one of the things that we've been talking about is feature stores, and you're literally writing a book on feature stores. So talk to us a little bit. What is it that the enterprise needs in order to make AI a real thing?
Starting point is 00:01:26 Well, that's a great question. And that last mile is fundamentally about model deployment and creating repeatable training of models, fast serving of models, and lineage and traceability for governance. And I believe that these are the enterprise requirements, but there's one really important piece that has to be woven through those requirements, and that is scale. Many of the approaches to operationalizing that last mile has created a loosely coupled architecture. And what I mean by that is that the data layers and the machine learning layers have been separated. And what our philosophy is, and I think many people's philosophy in this space
Starting point is 00:02:19 is starting to merge the machine learning capabilities with the data layer. So a feature store that can, in one component, be able to serve features to models in milliseconds, and at the same time, be keeping track of features that are changing over time so that you can train models and be able to keep that consistent and also make that reusable through the enterprise is becoming a critical component that's necessary. That's an interesting point, right?
Starting point is 00:02:58 You talk about the last mile of the AI or ML modeling. So that's not all that different than the last mile of enterprise obligations that enterprises used to have. If you were to look at it, things like governance, security, compliance, making it production grade, there are so many issues that get solved in that last mile. Basically, you're taking an innovation piece. How do you make it production-ready kind of thing, right? So AI is kind of moving into that space. But then again, I see a lot of companies, they're still struggling with the, maybe I can call it as the first mile of innovation with AI.
Starting point is 00:03:39 They're struggling to create the proper models. I mean, do you see a lot of companies already starting to move into the last mile and or what time or in what phase of their innovation should they be thinking about moving towards the last mile? Yeah, that's a really interesting question because it's really all about the maturity curve. Almost every company has a couple of data scientists now, major fortune, let's say 500 companies, global 2000 companies. They have a few data scientists, even if they're very early on their evolution on the journey to AI. Those data scientists now can capture data from data warehouses very easily. They can build models very easily on many people's platforms, my company's platforms, many platforms that are out there from other vendors and from the cloud vendors themselves.
Starting point is 00:04:35 Then when they go to put something into production, when they're taking a model out of the lab and really want to weave it into the enterprise workflow, then everything becomes very difficult. Because now you've got this blend of persona that need to cooperate with each other. You have data engineers that are maintaining all the data pipelines. You have the data scientists that are more mathematically oriented that are building models and making sure they statistically converge with predictive accuracy. operations and trying to package up those models so they're useful for the operational applications that they're going to impact. And right at that first deployment, they start to realize this is a lot harder than I thought. And then they get that into production and they get to their second model deployment and they're like, oh my God, this is never going to scale
Starting point is 00:05:44 across my enterprise if I keep doing it this way. And so're like, oh my God, this is never going to scale across my enterprise if I keep doing it this way. And so that's when it becomes apparent that model deployment and feature stores and the whole MLOps world becomes something that's not a nice to have, it's a must have to scale. And I think they feel that pain upon that first model deployment, Andy. That's actually a really key point, I think, because I think a lot of AI applications come in sort of as one person's idea over here in the corner. It's kind of a skunkworks project, or it's something that that guy knows about, or the data scientist, she put that together. I'm not sure. It's all about basically
Starting point is 00:06:31 operationalizing these things and making them not just function, but making them scale, making them part of the overall enterprise application infrastructure. And I think that that's really the, that is the last mile, am I right? Yeah, I think so. And this is best maybe explored with a real example of how you can architect to facilitate that last mile and bridge some of these gaps. As I mentioned earlier, people have separated the
Starting point is 00:07:07 data layer from the machine learning layer often, but what happens when you merge them? And as an example, in terms of model deployment, a traditional approach to model deployment would typically be an endpoint deployment. A data scientist says, I've got this model. I think it works. It has a pretty good F1 score area under the curve metric. And I'd like to put this into operation. And they have to go get someone else usually to wrap that model with a RESTful API probably, and maybe package it as a microservice and potentially containerize it, push it out as an endpoint. And those are traditional software engineering practices that are necessary to do that.
Starting point is 00:07:55 But an alternative approach is to literally put the machine learning model into the database. What does that mean? Well, imagine if you had a low code way of saying deploy this model as a data scientist, maybe a Python function or even a click of a button on a UI. And imagine the machine learning framework able to take that model, store it in a relational database, a traditional relational database, meaning it's in SQL, but a scalable relational database. And there are many on the market. And imagine if you could deploy a model now by simply doing the following, insert records into a table, let's call it a feature table where each of the columns of
Starting point is 00:08:46 that table reflect features of the model. Perhaps there's some metadata in that table too as to who's asking for a prediction, maybe a timestamp for the time. And imagine if the database was able to recognize, wait a minute, there are new records in this feature table. Let's trigger a model and let's figure out what the predictions are, the inference, put them into the records with the features and automatically memorialize that just as a relational table. And now users of the model would just be accessing SQL to get those predictions. There isn't a container. There isn't orchestration.
Starting point is 00:09:30 There isn't a RESTful API. There's no creation of that endpoint. It's simply low code, one line of code, deploy, and now you can just query a database with those predictions. This is the future, I think, of the last mile. This is, we call this in-database model deployment. This is, but this is something that I think is going to become standard in the marketplace as you look at the maturation of machine learning and data layers. But what this does is it facilitates that last mile where you don't need as many special skill sets or special processes, and you can rapidly deploy. And you have governance, too, because databases are extremely well controlled.
Starting point is 00:10:16 There's access controls and backup and resiliency capabilities. So I think this facilitates the maturation of the deployment of machine learning models into that real-time fabric of the enterprise, but with governance, with controls, but without having to have all these moving pieces in a technology that people know already, which is SQL. So that's an example of, I think, how this space will mature. There are like 20 points you made in there. I want to double click and drill you on, but let's start with the first one. So you were talking about using a SQL for the digital enterprise. That's almost like, you know, it's like you touched the kosher word, kosher thing that all the digital enterprises are moving away from SQL because to no SQL and other, you know, time series and whatnot, because the sequels become this, you know, only monoliths can handle it in the back of the enterprise.
Starting point is 00:11:19 It doesn't scale all of those issues. Right. scale, all of those issues, right? So now you're saying not only that you could have the digital enterprises use, you know, equal and off that, you know, the SQL database itself, but also you're suggesting the AI ML models can use that even as a future store or one-click deployment, which sounds too good to be true. It sounds too good to be true. Okay, tell me about it. But the fact is, if you look at the trends, there definitely was. There was a move away from SQL, especially as, you know, Google published the MapReduce paper and the big data architectures
Starting point is 00:12:00 from the Hadoop world emerged, there was a tremendous push into a different data model. But if you look at the trends, SQL is back and SQL is big right now because the other languages just simply were not representative enough or comprehensive in their behaviors to really support what's necessary for what we're trying to achieve with AI. And if you look at the trend, in the beginning, we saw MapReduce as being a way to scale. And it was a tremendous paradigm breakthrough. It allowed mere Java programmers to be able to handle thousands of computations in parallel without having a PhD in distributed systems. That was really important. But then quickly thereafter, we started to see things like Spark come into the picture, Apache Spark,
Starting point is 00:12:58 where you could not have to work with these esoteric mapping and reducing functions, but just work with sets of data and apply transformations to that data. Much more, I would say, programmer friendly at that point. But then what happened? Every piece of the big data architecture needed SQL. Everyone, including Google, built SQL stores. What am I talking about? On Hadoop, it was Hive first. And then there was Impala. And of course, there's Spark SQL. In the Google world, there's BigQuery. And then there was Spanner. There were even systems before that. And if you look at the leading capabilities now, and even look at the tremendous successful
Starting point is 00:13:47 IPO of Snowflake on the data warehousing side, this is all SQL that's returning. So what's happened? There was a major turn away from SQL, but now people are coming back and it's back to stay. What happened was exactly what you said. The flaws of the first generation of SQL databases have been solved. And those flaws had to do with architecture, not SQL, but the architecture of the database. One piece of it was the fact that the database was centralized. It worked on one server pretty much or a loosely coupled set of servers. It wasn't a scale out architecture. It typically was a scale up architecture.
Starting point is 00:14:34 That's been blown out of the water now by a number of the relational database and data warehouse vendors out there, including Snowflake, including Spice Machine, including Databricks, including Google, and many others. But that just now gives, that just gives you infinite scale. And it auto shards the data so that you don't have to worry about what server you put your data on. It's spreading the data out, distributing it out in an efficient way. So that took away, I would say, at least 75% of the criticisms of the old relational database
Starting point is 00:15:14 model. The other piece of it had to do with flexibility of representation, changing columns, changing schema, and so forth, being able to provide even unstructured data in storage. All of the modern databases have made it very easy for you to change schema and to be able to be elastic where now you can turn things on and off. Imagine having a petabyte of data inside of a data warehouse and you just turn it off when you're not using it. Now we all can do that. So yes, I would argue SQL is back.
Starting point is 00:15:53 And the reason why it's back is because it is the most comprehensive data language. It does support transactional and analytical workloads in very big ways. It's been around for 40 years, so it's 30, 40 years, and so it's really been made robust, but now it truly scales. So the answer to your first question is that I think you're going to start to see much less of the schema on read or the no SQL world being dominant. I think SQL is here to stay. And yes, my second point, and we can talk about that, I'll let you guide our conversation, is now not only is SQL really powerful as a data platform in general to power apps or to do insights for analytics, it is going to be the foundation of the machine learning world.
Starting point is 00:16:45 It is going to be so tightly knit with the modeling environment, the experimentation environment, and the MLOps environment. And I think you'll start to see this from many different capable vendors out there merging these worlds. Let me ask you a follow-up question on that before we move on. I get that. I mean, look, SQL, there's a reason why Oracles and DVDs of the world have been our own and even charging an arm and a leg, right? A ton of money for the enterprises.
Starting point is 00:17:19 And enterprises completely rely on that. I'm not questioning that. But the AI and ML models seem to be a force fit into SQL, but we can have another podcast talk about in detail. But are you seeing really the major enterprises starting to use SQL for all of their AI ML modeling? Particularly, I'm asking this question because when you do an edge distribution of the models, it becomes really complicated. And how do you store and distribute that when you use the SQL database? What percent of the enterprises do you see is moving? Okay. So that's a great question. So on deployment
Starting point is 00:17:56 and in terms of the proper architecture for the deployment of a model, and I made the argument that in database deployment of a model is very powerful. You brought up the edge and the edge is an alternative case, especially if you're talking about an edge that's disconnected from the internet, at least intermittently connected. Perhaps you're on an oil rig and you have some disconnected capabilities. Even there, you can put racks of computers. So that's not really an edge. But a driverless vehicle might be a good example of the edge where you don't have the time to get to the cloud or you just for safety reasons can't rely on the cloud. I was talking to another customer about cruise ships as an example where there's spotty connection.
Starting point is 00:18:54 In cases where you're not connected, you do need to have a loosely coupled scenario where you may have an endpoint deployment and push that out to the edge. And then as the model is making its inferences, that gets queued up for essentially transfer to a cloud repository that can be the feature store and keeping the historical set of features that have been used so that you can continuously retrain offline or in a centralized location. So I do think there's a role for edge deployment. But on the other hand, and this is an observation, in financial services, in healthcare, even in process plans, petrochemical, oil and gas, energy, things like that, the large majority of deployments have access to the data store all the time. And so
Starting point is 00:19:58 while there are edge exceptions, being able to leverage your data store inextricably linked to your inferencing is something that I think is way the majority of the case. And edge point deployments, edge deployments where you have intermittent connectivity is a very specialized case, and that you have to accommodate that case. And I think it's important also to remember that AI is still a nascent product category in the enterprise, and the vast majority of enterprises and data scientists and data analytics folks, well, SQL is their native language. So if you are a data scientist at a major corporation, you worked there a long time, you're starting to think about bringing in AI or ML into the database, you're trying to think about how you can leverage that to better use your data. SQL is just what you speak. And so by having a SQL-based ML solution, I think that's a lot more attractive
Starting point is 00:21:08 in the enterprise than it might've been maybe in the academic space or in the hyperscaler space. Well, that was certainly what was on my mind when we started our company. And we can even talk about this at an even more extreme level than just the skill sets. In one of our customers' circumstances, a very interesting situation where they're trying to modernize an old application. It's an insurance company. They've got a robust insurance application that handles many insurance functions. In fact, all insurance functions, including underwriting and claims and disbursements and so forth. And this was an older application. It was originally written in a legacy language on a mainframe, and they've converted it to a modern
Starting point is 00:21:56 language. And they're moving it to a cloud. They're on a cloud migration. And the question was, can we incorporate AI into that old app? And one of the things that we tried to do, even more than just support SQL, was to try to support legacy SQL. So that companies can not just be relegated to writing new apps in SQL for AI. But imagine if you could take those legacy applications and literally run them on a different SQL database than they were originally designed for. They're probably designed for IBM DB2 or perhaps designed for Oracle. Imagine if you could easily migrate them over to a scale-out SQL system
Starting point is 00:22:48 and still run that old app with the same functionality without the rewrites that have been required over the years. And now on that scale-out platform, start injecting models into it where you can just simply add a few more pieces of logic in SQL to that application to leverage that AI, but without having to re-architect or rewrite that application. So not only is SQL prevalent in the skill set, but SQL is prevalent in billions, if not trillions of lines of code out there that run organizations everywhere. I believe that those applications are not going to remain static. and bringing them to a more modern architecture, you'll start to see enterprise applications moving off of older frameworks, maybe even mainframes, into the cloud and actually being replatformed onto scale-out SQL, and then with AI models being injected into them. In fact, we're banking on that as a part of our business model.
Starting point is 00:24:05 And I think that you're seeing in every SI organization, significant practices on application modernization, cloud migration and application modernization. That's not just about bringing something to the cloud. That's about bringing an application to a modern architecture and then extending it and enriching it with AI. And I think it's very exciting times because we don't have to wait for people to build new applications to make those applications smarter and intelligent with machine learning models. You literally can inject machine learning into old applications much easier than ever before. I saw a couple of use cases you guys do using machine learning, particularly the one that you mentioned about doing a live leak detection. Tell me a little bit more about that. That's
Starting point is 00:24:59 very interesting because obviously places like water and chemical plants, especially when humans cannot go, this could be a huge use case. How is that applicable in the industrial environment? I'm really excited about industrial problems that we've most recently been applying some of our technology to. I've been exposed now to problems in petrochemical plants like ammonia plants, desalination plants, and other industrial frameworks where there is an order of magnitude more data than let's say in a bank or an insurance company or healthcare company. You've got these distributed control systems, DCS systems that are generating readings from sensors at significant velocity at great volume. Maybe in the larger plants at a million tags where a tag is a value of a sensor reading at a particular timestamp, maybe a million tags
Starting point is 00:26:07 a second. Okay, that's big. And typically, this data flows out of DCS systems into what are called historians. And these historians have been kind of the old school data store that is analogous to the SQL data stores in the business systems. Historians are the time series databases of the industrial systems. And we really haven't made a lot of progress of being able to inject AI into these industrial scenarios like leak detection or maybe mechanical failure, or even just doing some predictions about load distribution and utilities, because the data has just been too large. And these historians haven't been able to scale in a way that provides the ability to do both the storing of time-stamped time series information,
Starting point is 00:27:01 as well as the analysis on it. So the next generation of historians, I think will look more like the big data architecture, where you're able to ingest very large amounts of data, SQL databases like ours, or even key value stores that have been used in the big data architectures, things like HBase and other time series databases, Influx and Prometheus as examples. Some of these systems now can handle tremendous scale of time series data, but they'll be very tightly coupled in our case on the same platform, in other cases, through communication with analytical frameworks that can truly scale. So now you finally have the ability to be able to ingest
Starting point is 00:27:52 data from pressure, volume, and temperature sensors at grand scale, be able to, you know, do, you know, trending on that data in split seconds, be able to run that through models in situ and literally provide an operator with a dashboard that not only has real time sort of panels of what the sensors are reading, but right next to what the sensors are reading are predictions of failure states, predictions of
Starting point is 00:28:25 leaks, predictions of remaining useful life, predictions of mechanical failure, so that the operator is now not just looking at dials and at real data and trends and saying, oh, I think something's out of the ordinary. That kind of anomaly detection is literally now proposed right in front of the operator on screens because of that real-time stream of AI that's going on. This is a new world. And this is happening quickly. There are a couple of companies that just went public, or one is going public, that's C3 AI. That's led by Tom Siebel, who was on a couple of my boards in the past. He's a great mentor of mine.
Starting point is 00:29:10 Palantir, very well-known organization. Those are kind of black box, big framework approaches to this problem. We take more of an open source SQL data platform approach that's just much more modular and open source oriented. But what we're all doing is putting this big data architecture into the industrial frameworks that exist out there and providing that real time AI. think it's a very exciting time. I remember when we first did this back at NASA. I didn't mention some of my history. I helped co-run NASA's first AI lab. And I remember when mission control went from green screens to intelligent expert systems. I remember when my project took scheduling systems for the refurbishment of the space shuttle from little pieces of magnets on big boards and rooms to try to get people to prepare and turn it into real-time planning and
Starting point is 00:30:14 scheduling of refurbishment of the space shuttle. This was a major change that took place in the space agency. The Defense Department has seen that. Now industry is experiencing that with real-time AI. Applying AI in the industrial world or even real world, that's actually a good kind of trying to solve some problems, but also it goes to an extent beyond at times. For example, I was reading this, one of the, either a Scandinavian country or a Baltic country, the traffic was so horrible in the mornings around from seven to nine. So they wanted to stop that. So they put an AI algorithm and then based on that, they were charging tolls, which could be 10 times as much as a normal toll they were charged to go through the city center,
Starting point is 00:31:01 right? And then the reverse of the problem had to happen because people, when you force them, don't go out there, they won't listen to you, but you charge them 10 times more. And they were trying to recruit people now because nobody was driving through the town center through seven to nine in the morning. So yeah, I mean, predicting the patterns and usage and making recommendations in the real life scenarios, whether it's industrial
Starting point is 00:31:26 traffic or whatnot, that's where AI need to come up to speed fairly quickly to solve the problems that we can't solve otherwise. Yeah, I have to agree. And is nothing more satisfying than seeing some of this technology hit real-world problems, helping people with very tangible issues. One of our customers is using a real-time AI model to predict the trajectory of neurological disease and using population data across networks of neurological clinics to try to predict what's happening to a patient who has multiple sclerosis to get not just what one doctor thinks, but what the whole network of doctors
Starting point is 00:32:13 have performed and the outcomes that have historically been observed and using that to predict the trajectory for a particular patient. I think we're going to start to see the applications here injected into real-time processes across every element that we are involved in, where there are human experts doing things. And we will see, in many cases, very positive outcomes for people. And of course, we have to be careful, make sure that things are ethically done and governed properly. It's all other conversation. But the new capabilities of merging data and machine learning together to deliver on predictive real-time processes is really emerging in a very real way in industry today.
Starting point is 00:33:09 So, Monty, I have the joy of doing something a little fun here at the end of the podcast. So for season two, basically, I've got a set of questions that I'm asking the guests. And I've got a number of different questions. I'm just going to pick a couple of them and quickly ask you. So this is the lightning round. And don't be surprised if you don't know the answer. Just make something up. Okay. So here we go. Here we go. Ready? Question number one. How long will it take for there to be a conversational AI that can pass the Turing test and fool the average person? I think that's already happened on the average person with a conversational AI.
Starting point is 00:33:49 I think that happened a while ago in some ways, but GPT-3, the newest model that has been talked about lately that's an extremely deep transcoder model can do that today. All right. Another question. In your opinion, does it drive you crazy? Are you nodding along in agreement when people use machine learning, deep learning, and artificial
Starting point is 00:34:11 intelligence as synonyms? This is, it drives me a little bit crazy because AI is a lot larger than just machine learning. And machine learning is a lot larger than just machine learning. And machine learning is a lot larger than just deep learning. So I think it's important to make that distinction. There's a whole host of AI techniques, planning, scheduling, reasoning, logic, non machine learning oriented approaches to natural language like computational linguistics approaches.
Starting point is 00:34:43 There's a tremendous amount more to AI than just machine learning. And machine learning happens to be the most successful right now. So I think it's getting the limelight. And clearly deep learning has been very effective on unstructured data. So yeah, I think we need to try to be a little bit more disciplined in our terminology. Where can people connect with you and follow your thoughts on artificial intelligence and other topics? Well, that's great. So my company is at www.spicemachine.com. I happen to have a podcast as well called ML Minutes, where practitioners of machine learning have to answer our questions in one minute or less, I clearly wouldn't have been a good guest on that particular podcast. And also,
Starting point is 00:35:33 I'm writing a new book called Feature Stores for Machine Learning to be published on Manning Publishers. And I'm very excited about that coming out soon. And if people have thoughts on feature stores, I'd love to hear them. All right. Well, thank you so much, I really appreciate it. If you enjoyed this discussion, please do remember to subscribe, rate, and review the show on iTunes, since that really does help our visibility. And please do share this show with your friends. This podcast is brought to you by gestaltit.com, your home for IT coverage from across the enterprise. For show notes and more episodes, go to utilizing-ai.com
Starting point is 00:36:10 or you can find us on Twitter at utilizing underscore AI. Thanks and we'll see you next week. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.