Utilizing Tech - Season 7: AI Data Infrastructure Presented by Solidigm - 06x01: The Bedrock of AI is Data with Nick Magnuson and Clive Bearman of Qlik

Episode Date: February 19, 2024

Data is the foundation on which AI models are built, and integration of enterprise data will be the key to generative AI applications. This episode of Utilizing Tech brings Nick Magnuson and Clive Bea...rman from Qlik to discuss the integration of data and AI with Frederic Van Haren and Stephen Foskett. Enterprises sometimes worry that their data will never be ready for AI or that they will feed models with too much low-quality data, and overcoming this issue is one of the first hurdles. Another application for machine learning is improving data quality, organizing and tagging unstructured data for applications. The concept of curated data is an interesting one, since it promises to elevate the value of enterprise data. But what if a flood of data causes the model to make the wrong connections? If data is to be a product it must be profiled, tagged, and organized, and ML can help make this happen. The trend of generative AI is driving budgets and priorities to make data more useful and organized, but even unstructured data streams can be valuable. The application of large language models to structured data is promising as well, since it enables people to query these data sets even if they lack the background and skills to construct queries. Hosts: Stephen Foskett, Organizer of Tech Field Day: https://www.linkedin.com/in/sfoskett/ Frederic Van Haren, CTO and Founder of HighFens, Inc.: https://www.linkedin.com/in/fredericvharen/ Guests: Nick Magnuson, Head of AI, Qlik: https://www.linkedin.com/in/nick-magnuson-0a253931/ Clive Bearman, Senior Director of Product Marketing, Qlik: https://www.linkedin.com/in/clivebearman/ Follow Gestalt IT and Utilizing Tech Website: ⁠⁠https://www.GestaltIT.com/⁠⁠ Utilizing Tech: ⁠⁠https://www.UtilizingTech.com/⁠⁠ X/Twitter: ⁠⁠https://www.twitter.com/GestaltIT⁠⁠ X/Twitter: ⁠⁠https://www.twitter.com/UtilizingTech⁠⁠ LinkedIn: ⁠⁠https://www.linkedin.com/company/Gestalt-IT

Transcript
Discussion (0)
Starting point is 00:00:00 Data is the foundation on which AI models are built, and integration of enterprise data will be the key to generative AI applications. This episode of Utilizing Tech brings Nick Magnuson and Clive Bierman from Qlik to discuss integration of data and AI applications. Learn how data is supporting applications, learn how AI is helping to improve data quality,
Starting point is 00:00:23 and learn how AI applications are gonna improve data quality, and learn how AI applications are going to make data a lot more accessible. Welcome to Utilizing Tech, the podcast about emerging technology from Gestalt IT, part of the Futurum Group. This season of Utilizing Tech is returning to the topic of artificial intelligence, where we explore the impact it will have on technological innovations moving forward. I'm your host, Stephen Foskett, organizer of Tech Field Day and publisher of Gestalt IT. And joining me this season as my co-host for this episode is Mr. Frederick Van Heren. Frederick, welcome to the show. Thank you for having me. You know, Frederick, we spoke in the last episode about the need to bring together AI and data.
Starting point is 00:01:08 I think it was you, actually, that mentioned the fact that, you know, generative AI and large language models are but one of the modules required to build a complete artificial intelligence solution. And I remember what I said was that people get carried away because this is such a credible and capable component that they think that that's sort of the beginning and the end of the discussion when it comes to AI, but it's really not, right? Right. Yeah. I think generative AI is just showing the maturity of where artificial intelligence is coming from, from the last couple of years. And I think that we now can start looking at applications that build from a modular standpoint, applications for generative AI. And of course, we're still talking about data. Data is still the basics. But I think the generative AI is bringing a level of maturity to the AI market. is to get these things real. And one way to do that is to have them be populated with structured or unstructured data. When I mentioned that, I actually had in mind a conversation that I had
Starting point is 00:02:32 at AWS reInvent with Qlik. And so I am very, very happy to welcome to the show today, not one, but two guests from Qlik to talk about how to integrate data with AI. So welcome to the show, Clive and Nick. So my name is Clive Behrman. I'm the director of product marketing at Qlik. And I've been messing with data for a very long time, over 20 years now, despite my youthful good looks, of course. Yeah, Nick Magnuson here, head of AI at Qlik. Glad to join everybody. I'm responsible
Starting point is 00:03:09 for developing and executing our AI strategy, which is really predicated on helping our customers be successful in implementing and adopting AI of all varieties. So Qlik is essentially the data company. I mean, that's what the company is all about. And so it seems like this is a huge opportunity to bring data into or bring it to the table and not just have, well, a lot of the aspersions that people cast on generative AI, you know, the hallucinations and things like that. Is that really the strategy here? Is your plan to basically have data join the party? Yeah, I'll start with the reality with data and AI is they're interrelated. I call it a symbiotic relationship, even though I think that type of a term has a lot of loaded consequences. But the reality is data is sort of the foundation on which you build your AI models. And it's important that that foundation be established so that when you do so, you have a high degree of integrity and a
Starting point is 00:04:17 high degree of conviction and confidence in the data under which that AI model is built. But at the same time, as you build AI models, it inevitably happens where it tells you where you have data problems and you start to see those things borne out in the actual execution of those models. And so it should inform your data strategy conversely. So I've seen a lot of companies where customers are, where we've worked on the first day I model with them. And, you know, even before they get to production, they get an output from that model that tells them, OK, I've got a data quality problem over here that I need to go and look at. Or there's a lot of insight coming out of this particular signal from this particular aspect of the model.
Starting point is 00:04:58 And then therefore, there's an area for them to go and investigate, you know, further exploiting that opportunity. So I think it's a symbiotic relationship. The data is at the heart of either side of it. So when we talk about data, I mean, people understand the difference between structured and unstructured data. But in the end, people have to make data work. And sometimes we use terms like data quality and data pipelines and so on. So what are kind of the challenges you see when you have conversations with people around data?
Starting point is 00:05:36 What are the biggest roadblocks and what are the biggest concerns? Is it terminology? Is it understanding the data pipelines? So what do you see in the market? The challenges that we see with data are a couple of fold. One is you have an outcome that you want to achieve with an AI model and you go and you start that process and you identify we actually don't have the right data to solve for that. And so again, that's a, I think a positive evolution, right?
Starting point is 00:06:04 In anyone's journey, because now they know they need to go and collect that data. Conversely, it could be, yeah, we're collecting the data, but as we get into the modeling process, and again, AI models are very unforgiving. You start to understand that the robustness and quality integrity of that data is suspect. And so again, the action there is to invest in refining that data so that it can be used for AI systems. The other piece of this that I think is really important to understand is AI excels when there are a variety of inputs. The more variety, the better. We call it veracity. Essentially, AI can detect patterns in data that a human cannot.
Starting point is 00:06:42 And at scale is where we really derive the most value out of it. So as you're thinking about that data, it's not only about are we capturing it, is it robust, but is it across a variety of different inputs that can detect patterns that we wouldn't otherwise as humans be able to find. So the challenges, of course, are how do you get there? And I think there's some organizations that tend to over-rotate, meaning they never feel like their data is ever ready for AI. And you start to lose the benefit of that first step in the journey when you bring that data that you do have and you start to learn how you can improve upon it.
Starting point is 00:07:18 So I think there's a couple of misconceptions there that you're not ever ready. Well, that's not always true. And then two, that it doesn't form your data strategy as I talked about in the symbiotic relationship. Right. And I think when we talk about data quality, I mean, it's really a nonstop event, right? People collect data all the time. And so the definition of incoming data is my data of better quality than what I have today. So Clive, what do you suggest organizations do when they have streaming data and they're trying to identify the quality of their data? Although, you know, the quality of the data is a topic by itself,
Starting point is 00:08:00 but how do you see organizations kind of deal with data quality and measuring data quality, if there is a way to measure it? So for me, there's a spectrum of data quality, right? From an automated perspective where I can do a lot of very simple things in the data pipeline. So for things like uppercasing fields, lower casing fields, all the way moving, you know, we generally tend to put the data sort of sets on the left. And when we draw pictures, we put the sort of consumers on the right, right? So on the left, you can be a lot more automated because you don't have people involved. On the right, it's much more akin to data wrangling or data prep,
Starting point is 00:08:45 where maybe for a different type of AI, whether I'm doing maybe something like ML, where the data scientist might be using a tool to prep the data. So I think you're absolutely right. It's an ongoing process, but it's a spectrum from automated from the data source, correcting that data, if data is missing, if I'm not tagging the data correctly. And then, you know, when I'm moving it into towards being consumed, whether that's for an LLM module or an LLM to be used for RAC or indeed for a machine learning type instance. I apply different types of data quality techniques and rules. So it's not one. It's very it's many across that entire spectrum. I'm really not surprised to hear you talk about data quality
Starting point is 00:09:45 and the effect of machine learning on data quality because that's one of the things I think that was presented to me by my friends in the data space as one of the key applications for machine learning that they're most looking forward to, basically setting their unstructured data sets in front of a machine learning model that would be able to go through and do the things that you've addressed. So things like making sure that the data is properly
Starting point is 00:10:14 formatted, making sure that data is more structured, making sure that it's tagged properly, and using pre-existing smaller data sets as an example to allow a model to assist in that. Is that realistic? And is that the sort of thing that's happening? Does it work well? Should that be something companies are excited about using machine learning for? I was going to say, one of the things, you know, AI, whether it's machine learning or generative AI, is the latest sort of evolution in the massage of data, right? Over the last 20, 30 years, we can take a lot of the lessons we've learned with data warehousing, right? Build those pipelines that we can massage the data to a particular sort of format ready for consumption, whether it was a report or something like that. We can take a lot of those lessons and bring them to the fore and use them with AI and our AI processes, for sure.
Starting point is 00:11:18 Yeah, I heard some new terminology. Well, maybe it's not new. At least it's new to what I hear in the market is people talking about curated data. It's kind of a step up from quality data, right? It's data that kind of has been pre-processed by one of more organizations to the point where the data quality has been set
Starting point is 00:11:41 and easy consumable. And I think generative AI is really key there. But I do have concerns there because in traditional AI, the data is your IP, right? So once you start sharing your data, you're sharing your IP. So where's your revenue? And I think that's kind of a shift towards applications. But do you see that? Do you see people kind of selling or distributing curated data, if I can call it like that? Yeah, I've got an opinion on that. I'm glad you asked. My background, I spent about 15 years of my career, early stage in the investment area, and I saw a very profound evolution of using, it wasn't even called AI and machine learning back
Starting point is 00:12:24 then. It was just like the most advanced math and statistics that we could apply at the time. And it went from no one using it to people starting to try it out. And then it became pretty ubiquitous because it was so easy to get your arms around so much data, so many different inputs that can inform an intelligent investment decision. And at that point, it actually changed from who's using what math, which is kind of the stage I think we're in right now. Hey, are you using generative? Are you using traditional?
Starting point is 00:12:52 Are you predictive or whatever? To what data do you got? In fact, the most sophisticated hedge funds and so forth were at that point investing in teams of people that would go into the field and collect unique data. They'd be in a store, they'd be collecting data about what people are buying, how many people are going in the storefront, what, you know, what sort of buying behaviors they're exhibiting, because that unique data was more valuable than all the math that you could possibly come up with. And so I do see an evolution in this field where people will become astute enough to realize that, you know, the AI that they use is great, but the data that's underlined is actually the real medium that has value.
Starting point is 00:13:32 In fact, if I think if the IFRS, you know, the accounting standard boys got their stuff together, they would realize that this should be on the balance sheet of every organization. Like what kind of data do you have? Because that ultimately is your biggest, in many cases, your biggest asset. So again, I think you'll see organizations start to realize that they can monetize their data in a way in which doesn't compromise their IP, because it is something that is going to be a unique thing that can be brought to bear on the map that I think will become relatively, you know, not unique anymore. It's interesting that you mentioned that, Nick, that the value of this data, and I think that it would be hard to dispute the incredible value of data and the fact that machine learning and deep learning models are able to extract notice
Starting point is 00:14:21 patterns and extract patterns and qualify data sets in ways that humans not only couldn't do, you know, just sort of physically, but also might not even have the logical ability, the cognitive ability to sort of make connections and so on. But in a way, what you're describing sounds almost the opposite of curated data. So if a company, like an investment company, basically gave a flood of information, you know, it's not, it's not curated. I could see that there could be problems with that too, where, you know, maybe it would make the wrong inferences as well. You know, my background, I worked in retail. We talked about that last season on Utilizing Edge. You know, I worked with a bunch of retailers and one of their concerns was that they would make, you know, sort of wrong inferences based on the data that was being fed in. Because, you know, sometimes there's correlations that that they don't want the system to learn, that they don't want to capitalize. Is there a concern of sort of a data
Starting point is 00:15:26 flood in a way that causes problems because people are just saying, oh, machine learning can do anything. Let's have it look at all the data. Yeah, I think the data certainly needs to be germane to the use case or the intended outcome, right? I mean, I've seen early stages of generative AI where you put a bunch of unstructured data that you think is, it can confuse one versus the other. And then you ask a question of expecting it to answer out of this repository and it gets confused and answers out of the other repository. So I do think there's an element of curation to ensure that the right data with the right level of quality and integrity to it is present. Otherwise, you will have these instances where it goes, you know, the AI can't detect the right pattern to get the right
Starting point is 00:16:18 information. I also subscribe to the belief that organizations want control over the quality of that data. They want to know that in certain instances when you ask a question of a system that it's going to give you the right answer. Because otherwise, like you said, there's the possibility that it gets it wrong. And when it gets it wrong, there could be unintended consequences that come from it. So I think it really boils down to having control. Control can come in the form of, I'm only feeding it this input that I know is very restricted to a very specific domain or use case. Or having control over managing the quality of that data and seeing that if that data changes over time, I'm able to detect it, I'm able to be alerted to it, and I'm able to take an action so that I can prevent AI inadvertently using the wrong input
Starting point is 00:17:11 to make a decision. Yeah, and I think it's a real problem. I mean, to Stephen's point, when I started with AI in the early 2000s, I mean, our problem was not necessarily compute. I mean, of course, compute was a problem, but the bigger problem was the lack of data. I think today it's a little bit different. You can buy a lot of compute,
Starting point is 00:17:33 but with streaming data, you have so much data that the difficulty is to figure out what data brings you value. And the challenge with AI is you're not going to know unless you process it. So I do remember from my days working on AI projects, we were actually just collecting 30% of all the data we could collect.
Starting point is 00:17:55 But we had actually no idea if we were collecting the correct 30%. And when I say correct, I mean the one that has value. Do you see any data management tools today in the market that kind of give you an early warning saying, I got a pile of data, I can only use 30% of the data? Is there a way to quickly get value out of it to trying to figure out which ones to keep and which ones to kind of place a pass on you know you yeah you know you you've talked to when I said before about you know we shouldn't throw our processes out the window right one of those great processes is data lineage knowing where the data came from right whether that's from a structured perspective feeding feeding data from our structured systems into an AI system. Or in fact, if I'm going to train or fine tune a generative model with unstructured data, I've got to know where that data comes from. And existing data quality tools do a good job for both those things, right?
Starting point is 00:19:08 You can profile data sets or as we move more towards a data product approach, right? The data product will inevitably have a sort of trust score, if you like, or a score of how good the quality of that data is. But again, it's all about knowing not just what the type of data is, but where it came from, who's accessing it and using it. And of course, the more you can automate that and get tooling around that, the better. And yes, absolutely, to answer your question, there's lots of tools. We offer a bunch of data quality tools for lineage and such like, and for profiling as well.
Starting point is 00:20:00 So, yeah, absolutely, you can get help as you build're trying to do, not necessarily if your data is on point or not. So one of the things about Qlik that I think is fascinating is that you are working with a broad range of customers on a daily basis, not just in the AI space, but overall in terms of making enterprise data more valuable. Clive, what trends are you seeing in terms of how enterprises are using data and making it more of an asset? So one of the major trends we're sort of seeing is that the thrust around generative AI is definitely driving interest and budget around trying to get data ready for generative AI use cases. I think our generative AI benchmark report on our site is something like, you can go and download today. There's a big plug. Thanks very much. But
Starting point is 00:21:26 it says, you know, something like only about 20% or 27% of folks really have a formalized strategy for what they should do with AI. So, you know, the use cases around that, I'd pick a use case around, as we've seen, you know, sort of using generative AI as a helper, as a productivity boon. So for things like customer support, and you feeding it your data around your products, and asking it questions to sort of add as an extra channel to the existing channels that you already have.
Starting point is 00:22:09 You know, you can ask it how you install certain products or whether you could use certain features. I know Nick's thrust is all about sort of using the generative AI to augment analytics and asking questions around analytics. I mean, you know. Yeah, I'll pick up on that. I mean, one of the trends obviously is ChatGBT got into every person's household virtually in the world. And now every C-suite knows the power of AI in a very intimate way. And their boards do as well. So, you know, you've got this downward pressure,
Starting point is 00:22:44 which is a new phenomenon that I haven't seen in past cycles with AI, where you've got downward pressure to see results out of AI, and you've got these new budgets that are being born to bear against it. And I think in 2023, if I were to characterize Gen AI, it was an intellectually interesting technology with like a speculative theory of different use cases that we could apply it to. And in 2024, I think that changes where we're going to see real actionable business outcomes being driven through generative AI. So unstructured data becomes one of those
Starting point is 00:23:19 trends that is, we've been in this world of structured data. Now all of a sudden we have a new tool that we can access this very large body of information that has largely not been touched. So you'll see a lot of iteration in that direction. The one thing I would say, though, is all AI matters. If a business outcome can generate this outcome and it comes from structured or unstructured data, I don't think any C-suite really cares. So my point in all that is there are a variety of use cases that tap into structured data, structured data that your organizations already work with. They already know. They've
Starting point is 00:23:57 already spent a lot of time curating, preparing, ensuring the quality and integrity of that data. And oh, by the way, the technologies that work on unstructured data like, you know, our traditional machine learning, those have gone through many cycles of iteration, innovation, and improvement. They're hardened. I would just caution, like, the unstructured stuff is super exciting, high upside. But structured data, not every organization has gotten every ounce of value out of it that they can use in AI. And so I don't want people to over-rotate too much to what the hot trend is because there's still a lot of value to come out of the other side of that coin. Yeah, I think I see two trends, right?
Starting point is 00:24:37 So a couple of years ago, AI was really driven in organizations more on the CTO side of the house, meaning more research, people that have a good understanding of how AI works, not necessarily scale the technology. What I do see in the last couple of years is a shift to the CIO, meaning that enterprises see the value of AI and then push that through to the CIO, which comes with its own challenges, right? But then I think the second point you were talking about is generative AI. I mean, generative AI is relatively broad compared to what's available today. I mean, chat GPT is really one vertical within generative AI, which is large language models.
Starting point is 00:25:26 I mean, full disclosure, my background is speech. So when we were working on large language models 10 years ago, 10, 15 years ago, large was two megabytes. Nowadays, two megabytes doesn't get you anything. But I do see those two trends kind of mixing. But I do agree with you, Nick, that the approach of generative AI made AI easier to consume by enterprises and organizations. In other words, you can type something in ChatGPT and get some interesting results. I think that's key. The real question for me is chat GPT is people using chat GPT, right? It's not really building generative AI. It's consuming generative AI. How about any other markets you see within generative AI, meaning video,
Starting point is 00:26:28 pictures, as opposed to just text and large language models? Well, I think that the trend will get us into those areas. And for companies where that's a focal point of their business, they're going to get there faster because they have the most to lose. But if you think of what's the easiest starting point it's it's text it's documents right and there's a variety of use cases that that can open up where organizations haven't really spent a lot of time there unless it was critical to their business so if you think about the average organization the average employee at that organization how much time do they spend engaging with text data, documents, PDFs,
Starting point is 00:27:06 SharePoint sites? It's everywhere, right? I mean, it could be in operations, could be in HR, could be legal, you name it. I'm willing to bet that there are big productivity gains like Clive pointed to where if you had a way in which just interacted with that through natural language, everybody gains from that productivity boost. And then you'll start to get into use cases. Well, you know, we have all this imagery in our PowerPoints. Can we harvest information out of that? You know, we have, you know, speech information that we're capturing from recorded calls or,
Starting point is 00:27:42 you know, those types of use cases will come unless, again, like that's your core business. If you're a call center and that's the only data you really capture, of course, you're going to go after that right away. But there's no shortage of, I think, use cases that we're only starting to explore around using text data that every company has. It does seem like there's a big opportunity in unstructured data simply because structured data is structured already and is probably already used by applications, enterprise applications and so on. And I would think that there's just this massive amount of unstructured data that companies
Starting point is 00:28:12 are just would love to try to tap into. So that seems like a big opportunity. But even on the structured data side, one of the things that I'm extremely keen on seeing is what happens when structured business data sets are plugged into a large language model environment for what you were just describing, Nick, which was basically using the LLM to construct queries in order to make any kind of use of that data. that people could be constructing SQL queries, as simple as that, or working with Salesforce, for example, or SAP in a more conversational way, and that the answers that they get are not AI-driven hallucinations based on nothing, but are actually human understandable answers based on enterprise data. Is that the direction you're hoping for? Yeah, 100%. And I'm going to add on one thing that I think is incredibly important as it relates to
Starting point is 00:29:30 that workflow and that dynamic, which is if I ask a question and I get an answer, I've got to be able to interrogate and understand how it arrived at that answer, right? So if you do that with chat GPT, it cannot do that. It cannot tell you why it thinks you have a PhD from Berkeley and you're a noble physicist, right? You have to be able to go to the source document in Salesforce and show the answer and where it's being derived from. And that's, to me, incumbent whether we're working with structured or unstructured data, that auditability, traceability, explainability, whatever you want to call it, that's really important because users are not going to adopt that technology unless they can trust the response. And being able to trust the response means I can see the actual, you know,
Starting point is 00:30:15 where you pulled that context from within the document set. And, you know, I think that as we go from structured to unstructured and then combine those two worlds, that becomes increasingly important because you may get answers from different sources with different contexts. It doesn't make one right or one wrong. It just means that you have to, as a user, be able to understand where those sources came from and how it arrived at the response it gave you. And I'd like to add there with Nick, you know, it's not just the structured and unstructured, but the metadata around that too, right? So when you start feeding the data into these models, you know, you can feed extra metadata to do exactly what Nick said, you know, things like the last time this was answered, the last time this was updated. So it can become so, so powerful when you add all those sort of three things together, the structured data, the unstructured data, and the metadata as well.
Starting point is 00:31:15 Yeah, when I look at generative AI, I mean, to me, it looks more like an application, right? Certainly, we talked a little bit about large language models. How many people in the world, how many organizations, I should say, in the world are building large language models, right? It's a handful. When you use the generative AI is when you bring your own data to the mix,
Starting point is 00:31:40 when you integrate that, and in some cases, overrule or improve the data of the base model. I think that's what makes it really interesting to people. And that's where you merge structure and unstructured, right? It's the ability to take a significant base model and augment and improve it with whatever data you have, right?
Starting point is 00:32:03 You don't have to rely on anybody else's data. You can bring your own data. To a certain degree, it reminds me a little bit of Google Desktop in the early days, right? You would install it on your laptop and it would just crawl or your laptop and you can look for pictures and all that stuff. It's kind of a simplified view, but it's a little bit like that, right? You bring in an application and you provide the data
Starting point is 00:32:24 and then you can query it so i think that's that's really what what i find interesting about generative ai is bringing structured and unstructured data i'm just hoping that we will do a little bit more than just text right i think i think text text is key because that's an easy way for us to communicate. That's how we store data, typically, text. But we also do a lot of video and pictures, and it would be nice to see a combination of all of this. I mean, do you have any view into what's next with generative AI
Starting point is 00:33:00 and beyond large language models? I mean, Clive, maybe you can provide us with some feedback. We're seeing a bunch of things, right? Underlying a lot of the software that we have are essentially data structures, right? So even though the visual might look like a pipeline, underneath it, there's a document. And so we can use large language models as co-pilots and assistants to augment the software that we provide. So I'm going to ask it to build a data pipeline or ask it to augment my data set or clean my data set. And it's going to produce a document back, we're going to reinterpret that and display it as a different visual. And that's the same whether it's from a data integration
Starting point is 00:33:52 perspective or in fact, an analytics perspective, right? The underlying bar chart that you see in Qlik, there's a document structure behind that, right? So, yeah, it's very much going to change the sort of modality of how we interact with our software and our world around us, for sure. So we all know data is really important in the AI world. And I think with generative AI, data is even more important. So what are enterprises doing today with their data? Concerns are data quality, data pipelines. How do we integrate the data, analyze the data? And we all know that the value of the data can only be observed
Starting point is 00:34:37 when we process all that data. We have been talking here with Click, who has been working in data for a long time. What is your summary of data in the enterprise? Maybe Nick, can you summarize this for us? Yeah, sure. Obviously, the world in a generative AI world that we're in now has put a big focus on unstructured data and all the hidden potentiality that it might hold. And so I think organizations are now having to think about a world in which, you know, it's not in tabular form, but it's in a document form or an image form that we don't really understand and don't have ways to handle change data capture and those types of things, you
Starting point is 00:35:21 know, in terms of how the data changes over time. So obviously, the scope of what data means for an organization has gotten bigger. The challenges have gotten a little bit more unique to unstructured data. But I think at the end of the day, like all that data, it all matters. It's really about understanding where you have the easiest path to getting value out of it. And for AI, as it relates to that, as I said earlier, all AI matters. You know, there's a lot of focus on using unstructured data today. I think high ceiling, high potential on that.
Starting point is 00:35:53 But again, there's a lot of use cases that organizations have yet to exploit with data that they know really well. They trust they have a foundation set for it. And so to think about a roadmap that both includes unstructured and unstructured data is really important going forward. I'll put a plug in for data integration from Qlik, right? That, you know, our strategy has three pillars, the first of which is the trusted data foundation. And we do three things with that trusted data foundation, right? We make an AI-ready data set. So, you know, that doesn't have to be generative AI,
Starting point is 00:36:28 but that could be, you know, an ML-ready data set for training or, in fact, in production. So that's the first thing, make those AI-ready data sets from enterprise data. I can create enhanced or augmented data sets. So I was talking a bit what we did earlier about improving the data quality for maybe the machine learning data that's in place. And the third thing that we do, we can provide the augmented data for fine-tuning the large language models with domain-specific data. So, again, it's creating those AI-rated data sets, enhancing those data sets,
Starting point is 00:37:14 and then fine-tuning those large language models with domain-specific things. So, you know, that's what it means to me to provide sort of that trusted data for whatever AI initiative you're sort of embarking on in your enterprise. I think overall, we can tell that data is going to be the key here. As we talked about on the last episode, and as we've sort of been walking around this whole conversation, you know, people have this incredibly powerful tool that is large language models and generative AI. The thing that makes it a useful tool is the data that's feeding it. And without that, then we're going to have all sorts of the problems that people worry about. And they're fearing, you know, the hallucinations and things like that. With proper data integration, and with, as Clive and Nick both pointed out, with the ability to check and verify and make sure that the results that it's giving are actually based on
Starting point is 00:38:19 enterprise data, well, then that's a whole different ballgame, isn't it? Because it means that certainly this property, this technology becomes a much more useful and friendly and universal user interface paradigm for enterprise data. And then that is opening a whole world of new applications. So that to me is going to be what I'm going to be looking forward to this whole season. I'm really glad that we are able to set up our season here of Utilizing Tech focused on AI by talking to a data company because I guess editorially speaking, that goes with my beliefs about what's most important when it comes to rolling out AI-based applications.
Starting point is 00:39:01 So thank you so much for joining us today on Utilizing Tech. As we wrap this up, where can people connect with you and continue this conversation about artificial intelligence and other topics? I'd point us to click.com slash stage. That's S-T-A-I-G-E. And then you'll get all the information that you need there. Yeah, and I would just point out for those that are interested in actually getting hands-on and into developing the next wave of AI technology, we are hiring. I have a very aggressive hiring roadmap and product roadmap that supports it.
Starting point is 00:39:38 So click.com slash careers is another source for those out there that are interested in being part of the delivery of these new innovative solutions. And Frederick, you and I are both going to be part of AI Field Day this week. What else are you working on? Yeah, I'm working as always, working with customers, teaching them what AI is and trying to find what the best AI solution is for them. But just like you, I'm looking forward to AI field day and also happy to do a face-to-face as opposed to always video. Absolutely.
Starting point is 00:40:13 It's nice to be back in person. We've got a great group of people coming out to California for this. Wednesday, Thursday, Friday, presentations from important companies in the AI space. If you miss any of that, you can find that live on LinkedIn space. If you miss any of that, you can find that live on LinkedIn. But if you miss any of that, just go to YouTube slash Tech Field Day, and you'll see video recordings there real soon after the event. So thank you so much for listening
Starting point is 00:40:36 to Utilizing AI, part of the Utilizing Tech podcast series. If you enjoyed this discussion, please do subscribe. You'll find us in your favorite podcast applications. And please consider leaving us a nice rating and a nice review if you're on the nice side. And if you're on the naughty side, I'd love to hear from you and figure out why. We'd love to address any concerns you have as well. This podcast is brought to you by GestaltIT.com, your home for IT coverage from across the enterprise. For show notes and more episodes, head over to our dedicated website, which is utilizingtech.com or find us on Twitter or X,
Starting point is 00:41:11 Mastodon and so on at Utilizing Tech. Thanks for listening. And we will see you next week.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.