LPRC - Episode 11 – Artificial Intelligence & Machine Learning in Retail AP

Episode Date: August 9, 2018

The post Episode 11 – Artificial Intelligence & Machine Learning in Retail AP appeared first on Loss Prevention Research Council....

Transcript
Discussion (0)
Starting point is 00:00:00 Hi everyone, welcome back to Crime Science. In this podcast, we aim to explore the science of crime prevention and the practical application of this science for loss prevention and asset protection practitioners as well as other professionals. Co-host Dr. Reed Hayes of the Loss Prevention Research Council and Tom Meehan of ControlTech discuss a wide range of topics with industry experts, thought leaders, solution providers, and many more. In today's episode, Dr. Reed Hayes and Tom Meehan will discuss artificial intelligence
Starting point is 00:00:26 and machine learning in retail AP with David Spikes and Daniel Downs of Atbers Retail. We'd like to thank Bosch for making this episode possible. Video analytics is an important enabler of any retail loss prevention strategy that improves security and delivers clear business advantages. With the intelligence built into Bosch IP cameras, you can use video data for far more than security alone. This includes identifying patterns in customer activity and recognizing retail hotspots to improve sales, as well as optimizing store layout to enhance the customer experience. Learn more about adding intelligence to your system in zones one through four of the LPRC zones of influence by visiting Bosch online at boschsecurity.com. All right,
Starting point is 00:01:03 well, welcome everybody to another episode of Crime Science. Today what we're going to do is talk a lot about what we call Zone 5. And as you know, we are all looking at how we define and handle our issues in three dimensions. And so we talk about Zone 1 as whatever we're protecting, that asset itself. We talk about zone one is whatever we're protecting, that asset itself. It could be a person. It could be an item. It could be cash, a cash register, a safe.
Starting point is 00:01:33 It could be merchandise on the sales floor. Zone two is that immediate area and what can we do there to protect, influence behavior, to understand better what's going on. Zone three is the entire interior space of our store or distribution center, warehouse, office environment. Zone four is the parking lot from the building wall to the edge of the parking lot. What can we do there, again, to understand the dynamics and better address ways to keep people safe and run offenders off? And again, and then zone five, of course, is beyond the parking lot in the cyber environment and into the built space, the social environment. So today, we're going to talk a little bit about, you know, fraud and other threats to our overall environment. But what can we learn out there in zone five, and then talk a little bit about a new book that's out that we're
Starting point is 00:02:23 excited about, that kind of takes a look at these things and helps all of us think more quantitatively. How do we change things into numbers and make sense of the world in that way and then convert it back? So what I'd like to do is I'm going to go to Tom Meehan, my partner in crime here from Control Tech on crime science and crime. Tom, if you would go ahead and introduce our special guest today and we'll go from there. Well, thanks, Reid. Really excited to be here to talk about data and analytics and the zones of influence. And especially excited to have our special guests, David Seitz and Daniel Downs. David, Daniel, welcome to the podcast. Thanks, Tom.
Starting point is 00:03:20 And what we're going to do is we're going to go through the zones of influence and really how data applies to retail risk and asset protection. And our guests are experts in the field, and we're going to have a lot of specific questions, but we're just going to have a general conversation like we always do. David, why don't you tell the listeners a little about yourself? Yeah, I'm David Spites. I run data sciences for APRIS. And most of the listeners are probably familiar with APRIS retail, but I actually have responsibility for the whole of APrys, which includes Aprys Health, Aprys Safety, and Aprys Retail. And then our team is probably a team of about 25, and we're focused on all three business units, but probably about 70% of our time and effort is dedicated to retail.
Starting point is 00:03:59 My background, I have a PhD in biostatistics, and I jumped into business a long time ago and sort of deviated from the biostatistics part. But, you know, my background is applying data and analytics and models, predictive models, artificial intelligence to solve real business problems. Great. Thank you. Daniel, why don't you give the listeners just a brief overview of who you are and what you do at Aperis. Sure. So my name is Daniel Bounds. I have worked for Aperis going on seven years. After my PhD in criminology, I got into LP and LP analytics back in the day with Reed. This is back when we had a little office up in the Sears building on the second floor. So I was a research coordinator for LPRC for a couple of years and then relocated out to California. And with Aprys, I do predictive modeling, analytics, and been doing that for the last seven years, working on retail and safety initiatives.
Starting point is 00:05:04 Been doing that for the last seven years, working on retail and safety initiatives. So I'll turn this over really to both David and Daniel. But I think when you talk about Aperis Retail, I'm obviously very familiar with it and excited to have you on. But could you give the listeners kind of an overview of what Aperis does and some of the products that are related to retail and actually outside the retail sector so folks can understand a little bit about what your organization does. Sure, I'll handle that one. So, APRIS retail is really the combination of the retail equation and CIS Republic, and we joined together about two years ago as a single organization under the umbrella of APRIS, which is a Louis actually, we track 80% of the opioids in the United States, and we monitor those and report them back to 42 states use us exclusively to provide that reporting back to doctors within the states. And we have a layer of analytics on top of
Starting point is 00:06:21 that. So we're trying to detect people are going to overdose from overdose from opioids you know while we're trying to fight crime and retail so we're doing quite a lot of different things and then in our safety business Aprys actually has about about the same amount of the jail bookings in the United States about 80% of the jail bookings we have 48 states that use our service in one way or another and we monitor people going in and out of jail so we can notify victims, report that data back up to criminal justice investigators like the FBI, the CIA. They're using our data to go and investigate people's arrest histories. And then we layer on top of that as well, crime analytics, recidivism analytics. We look for,
Starting point is 00:07:04 you know, likelihood someone's going to commit a specific kind of crime. And a lot of that techniques, a lot of those things we apply is exactly what we're applying in retail. And so our retail business has a number of main products there. The retail equations main product was a return fraud detection product called Verify. And that product monitors returns in real time and applies predictive modeling, machine learning type techniques so that we can detect in real time
Starting point is 00:07:34 when someone's either committing fraud or abuse or is a general outlier to the business. And then we shut that person down or we warn them, you know, they can't make returns anymore. CIS Republic, who we joined forces with two years ago, and I say we because I came from the retail equation, but now I represent all the companies. When we all joined forces, CIS Republic brought to the table a very popular exception-based reporting system called Secure. And that product we've now, my team's gotten heavily involved in. We've started to apply as much analytics to it as we can.
Starting point is 00:08:14 And we began going down the path of integrating in artificial intelligence-based, really machine learning-based exception reporting or employee fraud or abuse detection. And so I think that a lot of the questions in this conversation may revolve around that really that new step, which is to take analytics to or take machine learning and modeling to detecting employee fraud. Great. That's really helpful. Actually, I learned a lot just in that brief overview.
Starting point is 00:08:43 Would it be fair to say that would you classify yourselves as a data company? I would like to call ourselves a data and analytics company, but we are definitely a data, like as a whole, as Aprys. I mean, we warehouse data for 42 states for opioids and other 48 states for jail data. And for retail, we're probably a little bit of both. We're warehousing the data to solve problems, really. The data is just there so that we can say yes or no to a return so that we can help identify employees that are fraudulent. We actually have products that do marketing as well, where we're trying to figure out what incentives to give to individuals. And all that's really more of an analytics play. Got it.
Starting point is 00:09:28 So I wanted to talk a little bit about your book. And Reid and I both share books all of the time. And the book is titled Essentials of Modeling and Analytics, Retail Risk Management and Asset Protection. And from my perspective, not only is it a very good book if you're into data and analytics and asset protection but I think it's the only one that covers the vast sides of data analytics and modeling in relationship to risk in retail I kind of when I when I went through the book I read it about six weeks ago and went through unrelated to this call, was actually looking at it.
Starting point is 00:10:07 And one of the things that came up is integrating analytics into loss prevention teams. There's a section about that. And you talk about people. Can you share a little bit about what your thought process was with our listeners there? Yeah, I really wanted to. And I see an evolution of LP teams trying to go down this path on their own to some extent. And to get, you know, our team has 10 PhDs and nine more with master's degrees. Getting to that level takes a long, long time.
Starting point is 00:10:38 But to get started, I was trying to give the sense of the skill sets needed to just get started. And some people, their first instinct is, oh, I'll find a new grad student. We'll bring him in and he'll solve all our problems. But in reality, the very first analytics person you hire should not be the random grad student. You've got to hire somebody that has some business acumen and can do both. They need to be able to solve business problems and talk to humans and not just be math guys and also be able to do the math, too. So, you know, our take on it is you kind of need somebody with a lot of these skills. And so I listed out in the book a variety of skills.
Starting point is 00:11:18 There's five different skills there. And in your first hire, you want to have some level of all of those skills. They need to be able to communicate might not manage their own projects team lead etc and otherwise you're going to get what happens is people don't know how to manage analytics people and they have no idea what to call on them excuse my language but to say are they really progressing the way they need to progress are they solving the problem in the way they need to solve it? So you need kind of a proven entity as your first hire, somebody that's actually put analytics into work, not just a random student that's really never done it in the business world.
Starting point is 00:11:56 Especially for an LP team who, you know, I view LP people as very no-nonsense, smart thinkers, and they're practical and they want things that work. Whereas sometimes you get somebody right out of grad school and they're more living in the clouds. They haven't quite learned how to apply their craft yet to solve real problems. That's where I was heading with that. So, you know, just a quick personal story here is I guess it was about 10 years ago. I was given some data responsibility and pushed very aggressively to hire data scientists. So I wish your book was available for that. And I would say if the listeners on this call are AP professionals that are adding data analytics folks to their team, that one section, if I had it 10 years ago, would have made my life a lot easier.
Starting point is 00:12:45 So working through, I highly recommend it. I know what it was like. And I can tell you that I went through a round of hiring data scientists and all different people trying to figure out what we should do in-house versus through a third-party vendor. So kudos to you on the book, you and your colleague there. And I really think it was great. I'm going to turn it over to Reid for and your colleague there, and I really think it was great. I'm going to turn it over to Reid for some questions. Thanks, Tom. And, you know, I guess to follow up on that, that's a bonus to the book is thinking about how to use data and then maybe a little bit about,
Starting point is 00:13:19 well, this might be beyond what I know how to do, So I need to have somebody on my team that can do this. And now what am I looking for? What really, what skill sets? And like you say, you know, very often we're all in that position where we've got to bring somebody onto the team and we just don't know enough to know what kind of person and their capabilities that they should have, what their capabilities are.
Starting point is 00:13:43 So that's another bonus. I want to talk a little bit about, since we're talking about data, that's the world that you guys live in, we all do, but that you're mostly specialized in. And so how do you look at these things? I noticed that your team works on entitlement programs, for example. And I know that at UF we had a grant to work on SNAP and WIC fraud, which is essentially food stamps and so on,
Starting point is 00:14:12 and that there are these dynamics out there just like with any other program where there was fraud. And so Dave and Dan, what can you tell us a little bit about what your organization does to help people better understand the dynamics of what's going on to enable them to better address those problems? So we've done that a little with our safety division. And for those programs, the problem we're trying to solve there is fairly simple in the sense that we're trying to identify people that would be disqualified
Starting point is 00:14:45 because of their criminal activity or because they're actually in jail, not walking the streets. And so there we're trying to figure out what are the triggers that may disqualify them from those programs. Dan, you want to elaborate a little more? Yeah, I would just say, you know, we'll partner with agencies to basically identify these folks and help save money. So we'll basically, we'll look at the data, jail data, and try to identify these folks. Basically, it's a withdrawal. All right, well, that's good stuff. So I think at the end of the day, what we do at LPRC and on the UF team is try to understand human behavior. Good, bad, benign. Better predict what they're going to do, where they're going to do it, how they're going to do it, when they're going to do it, who they're going to do it with, and a little bit about their motivation and motives for doing it.
Starting point is 00:15:48 And it sounds like, and what I know about you all, your team and your different offerings, is that that's what you do. You're trying to help understand, help your customers understand externally what people are up to, where they're coming, and then help them use leveraging what you're feeding them and your skill set to reduce the problem, suppress the problem, displace the problem, and then the same thing on the internal side. But can you guys kind of talk about the behaviors you pick up on, the data that you get, how you use that, and how you make sense of that, not only for yourselves, but as you said, to get it back to the practitioners that need something that's very understandable and actionable.
Starting point is 00:16:30 Our goal generally, Reid, is to bundle up the data. And what we collect from retailers, that data is typically point of sale data and a bunch of add-on tables like store masters and item masters, employee masters, sometimes a customer database. And we're doing a number of things there, but we're really trying to create variables. But what I mean by variables is just summarizations of the data that tell something about the data. So usually we're operating on a unit. So the unit may be a customer or it may be an employee. And then a summary of
Starting point is 00:17:05 that data might be, you know, something simple about employee might be how many transactions have they processed today? How many had voids in them? How many times did they run a refund to themselves? So we're trying to summarize up that data into variables, which can be very understandable. We do the same thing on the consumer side. We look at, you know, how many refunds they do, how many purchases they make, how profitable they are to the organization. And there's ways to summarize up that data to give you those snapshots of that individual. detection models we're calculating probably three or four thousand variables and there'll be things you know there may be categories of variables that we calculate at different windowings of time they're they're just minutiae of you know we might look at things 15 minutes within 20 minutes of an activity so you look to see okay they ran a post void did they re-ring within five minutes
Starting point is 00:18:02 within 10 minutes 15 minutes we'll come up with varieties on that um and then it's really about predicting outcomes so the outcome may be they got terminated later that month or the the customer side it may be they got arrested for shoplifting or they got you know identified by the store lp team and they were taken in the back room and interviewed and we're really trying to predict those events from all those variables. That's really the main thing we're doing and how we're bundling up that data. All right, fantastic. So that's part of it. It sounds like right now the lion's share of the data that you collect, that you bundle,
Starting point is 00:18:39 and that you recode and use is coming from the retailer, particularly transactional data from the point of sale. What other information data are you all collecting, Dave, to make sense of the world that I might have missed that helps further inform you and your practitioners? It's mainly internal data from the retailer. We've gone beyond point of sale to some extent, and we're heading down the inventory path, trying to get data on inventory flows and out of the store. We're trying to collect as much of the case database as we can where we have times where they've been investigated. And with our exception reporting system, we're beginning to collect click-level data of the investigators so we know whether or not they're opening an exception, right? So if an exception comes back and says, hey, you should look at this, and they just immediately close it out and say I'm not interested without even opening it, that tells me something different than if they open it, which tells me something different than if they open it and then they place it into a case. Even if that case doesn't succeed, it tells me what they're interested in and what's catching their attention. And all those facts can be used as outcomes to model against.
Starting point is 00:19:56 All right, excellent. Can you guys provide sort of a real-world example? how, in this case, a retailer, use your data to impact a problem, whether it's an individual situation, would be helpful, and then also, you know, overall, how they're impacting across the enterprise. Yeah, our objective is to bundle the results up and hand them or spoon-feed them little chunks of information that are very pointed. So on the return side, for example, the retailer doesn't necessarily interact with the data, the variables, the model at all. All they know is we declined somebody or we approved somebody. Now, where they may learn about it is during the algorithm development process. So as we're developing the algorithm, we're going to talk them through the types of things we look at to make that decision. They're going to interact with us and say,
Starting point is 00:20:49 hey, I'm a big retailer and I have a very rich loyalty program and my top tier loyalty customers need to be treated with kid gloves. And so we may take that into account in our models and say, we actually will never decline those top tier customers. And so we interact with the retailer on broad stroke levels. They're not in there engineering the actual models. They're telling us what they want to accomplish with those models. And in the end, the model makes the decision and the algorithm makes the decision automatically. On the employee fraud side and exception reporting side, we provide a tool with secure where they can
Starting point is 00:21:27 really interact with the data. They can write their own queries. And there, it's more of a hands-on experience. So that one is probably more traditional exception reporting. With our employee fraud models, what we're trying to do is get that a little more in line with how we do return fraud, where we're detecting the employee that may be suspicious for some reason or another. And then we're playing that back to the retailer to say, hey, I think Joe Smith over in store 123 is doing something weird, something that may be fraudulent. You should take a look. And then we'll literally, as part of the modeling process, we'll play back specific things that drove that person to be suspicious in our minds. We'll say, hey, they ran this kind of transaction at this time and we'll bring up the actual transactions and we'll show them how they relate to other employees in the store, other employees in the district, et cetera.
Starting point is 00:22:24 employees in the district, et cetera, and say, look, they're five times the normal, or they ran 17 transactions where 90% of their items were line voided. The typical employee does that once a month. Those are the kind of things we're going to be playing back to the retailer, which allows them to kind of interact with the data. Excellent. So let me ask you another question, if I could. Could you go explain in a user-friendly way, we know what big data is, you know, and the term that people use are throwing around out there, and then what machine learning does, how that
Starting point is 00:23:01 helps you and others make sense of and better use data in a user-friendly way, Dave? Sure. Big data is really, you know, it's a buzzword like a lot of the buzzwords that came about, but there is a meaning behind it. What people were getting at was big data was more about pulling multiple data sources together. And so, for example, a big data play we might have is to pull our crime data, and we've talked about this and we've actually done a little bit of it. Pull our criminal justice database with all the arrests into the retail world to say, hey, is there something in the retail world where I could use that data to predict the other data? We've done it in the opioid space, too. We've brought the criminal data into the opioid prescription monitoring data.
Starting point is 00:23:45 prescription monitoring data. And we were able to predict actually the likelihood you're going to overdose on opioids is 7x greater if you've been in jail, for example, 14 times greater if you've been in jail for a drug-related offense. And so we've used that, that sort of big data in my mind is coming, bringing data together from multiple sources. Like if I bring point sale and inventory and crime data and all that, and I put it all together and I solve some problem with it. Now, the machine learning piece of it, machine learning is sort of the evolution of predictive modeling, really. Predictive modeling came out of statisticians and social scientists, and machine learning kind of came out of the computer scientist side. But they're really trying to address the same thing. They're trying to do the same thing. They take data, they turn it into variables, and use those
Starting point is 00:24:30 variables to predict an outcome. Now, the differentiator and what's got machine learning all the buzz lately is that it's been applied to some pretty cool problems like computer vision, speech detection, self-driving cars, and this has kind of gotten it a little bit more notoriety because you can enter you can see that the general public can see when alexa talks to me and answers my questions and looks up things for me uh that's cool that's like a new thing but that's really machine learning being applied to uh speech recognition now in the retail side, though, machine learning kind of comes to play is now we have complex problems we want to solve, and we're really just taking the things that we were solving with predictive modeling in the past, and we're sort of upgrading it a little bit.
Starting point is 00:25:16 And the cool thing about machine learning is it looks at very complex interactions automatically. So you don't need, I mean, you probably can solve the same exact problem with predictive modeling, but what machine learning gains you is it's going to automatically look at variable interactions that you can't, that the human may not have caught up, caught on to. And so that's really the next step. It's not much different than predictive modeling. And in fact, people that do predictive modeling now use machine learning techniques. And the fields have actually pretty much merged on top of each other. Now, the whole AI things of umbrella over all of it.
Starting point is 00:25:54 And that's usually it's got machine learning and predictive modeling, which are really mathematical techniques inherently. And that's kind of how I see everything relating together. I want to throw out another thing to you all, and that is we have what we call innovation chains here at LPRC. And so we, you know, we innovate to specific problems. This is a problem-solving community. So if we've got individuals or crews that are using sledgehammers to breach walls and go in and take Apple technology, steal them from stores or distribution centers. If we've got that's a specific issue, we're now going to try and define who, what, when, where, why, and how is driving that. And then we're going to say, okay, let's outline some specific solution sets that would address a would-be offender
Starting point is 00:27:21 or crew of offenders at each step or stage of committing, in their mind, successfully committing that crime. We're going to lay that out. We're then going to, okay, here's some tactics and techniques and technologies that we could use for step one as they ideate. Step two is a, you know, recon and initiate and so forth. Then at the next level, the next link in the chain, we're going to move from the lab environment where we can kind of do whatever we want to do in a way
Starting point is 00:27:50 to one or more locations, maybe in Gainesville, Florida, right here, where we can now adjust and dial in, adapt what that solution set is to that drugstore environment or a convenience store or department or specialty or whatever. Work together. How do we do this? Okay, now let's go to the third link. Now we're going to go, say, to this Baltimore region, three or four county area. That'll be our research lab as an ecosystem. But we'd like to look in zone five. We'd like to understand not only the data that are coming out of our point of sale and out of our item movement and other data sets, how can we look at who's taking advantage of those things? Let's say in Baltimore, do you all have any ideas, suggestions that you could share with this podcast audience on how you would look at that specific market in that area and understand the dynamics there and ways to better address them. So to reiterate your question, you're interested in focusing on a specific
Starting point is 00:28:56 crime type within a specific geography and how we go about building that up or kind of getting to what you call zone five? Is that your question, Reed? No, exactly, David. I think, you know, let's say in the Baltimore case, and we've got theft, a lot of it seems to be caused by and even driven by the ability to conduct illegal illicit returns in that dynamic? And let's say it's that Baltimore, greater Baltimore area, the three counties up there that we're interested in. Yes, is there something that big data could do to help us paint a picture as an individual retailer, but also in this case as a community of 16,
Starting point is 00:29:38 20 retail chains working together and working with law enforcement, all the agencies at the local, county, state, and federal levels. Yeah. From the machine learning predictive modeling side, you might look at if you can get your hands on actual events. Obviously, that's the best case scenario. And it's back to just the things you might put together to do a normal regression
Starting point is 00:30:06 analysis or to do some kind of analysis of an association between different things. So if you can say, okay, let's say I have a hundred stores participating in that pilot from, I don't know, 10 different retailers. And then you might say, okay, well, can I boil down what are the characteristics of those stores? And you could look at the likelihood they're going to be targeted or there was an event there. And then you're really just building models to relate those facts together. things you might be able to change in those stores to prevent or to limit the criminal activity there. Is that kind of what you're aiming for, Reid?
Starting point is 00:30:53 No, I think that's exactly it. We're looking, as you probably know, at robbery, parking lot violence, other violent events, but we're also now looking at maybe there are truck hijackings, there are issues in the supply chains that last mile delivery on the doorstep, and return fraud, what's going on with our refunds. We're trying to understand and map and measure and then affect and see what happens when we do that in a larger scale area and could big data and some of the machine learning and analysis techniques that are out there help us? We're working a lot with mapping now, but we want to continue to look for innovative ways to make sense of what's going on and then measure and observe what happens when we take action. If we implement some solutions in some test stores and not others, what happens?
Starting point is 00:31:46 What are the ripple effects? I mean, take your truck example. That's a great one. I mean, if a truck's robbed on the way to delivering to a specific store, I mean, you might, for example, that's a very specific problem. A lot of the machine learning and the predictive modeling tools are engineered to solve very specific problems where you can pinpoint an actual outcome. So there you may say, okay, well, how many robberies have we had in shipment to these stores? And then you start looking at variables about the robbery or about the shipment and say, okay, well, what about ones where they didn't have a robbery? Was there any differences there? For example, were the robberies more likely to occur if the shipment was after 6 p.m. or before 5 a.m.? Or, you know, kind of breaking down what are variables that are involved, like, you know, what was actually in the shipment?
Starting point is 00:32:39 Or what kind of truck was it? Or were there two guys on the crew versus one? was it or were there two guys on the crew versus one or you know kind of just breaking down each and every variable there and seeing how that relates to the probability of an incident or the probability of a shipment being stolen and i mean that's kind of how we always look at these problems is trying to figure out an outcome that you can measure and variables that you think just logically might predict that and then you just go about using a database method uh breaking that down and then the next level is once you can identify those sets of things you can feed that into machine learning type algorithms
Starting point is 00:33:11 to give you a little bit more edge in terms of it'll pick up on more complex interactions like you know if it's after a certain time at night and there's you know this is the particular shipment then i have a way more likelihood of a theft occurring than, you know, if you just looked at one variable at a time. Excellent. Sounds good. Tom, I'm going to go ahead and toss back to you and for more insight and questions. Yeah. So, you know, David, you talked a little bit about BI and big data. Can you give the listeners kind of a layman's term in the simplest terms, what artificial intelligence is in relationship to that? Yeah, I think the simplest would be it's something that mimics a human action. Right. So a dumb one might be like I always like an ATM is kind of a dumb example. A dumb one might be like I always like an ATM is kind of a dumb example. Right. It's not really actually new. Newer ATMs are.
Starting point is 00:34:09 But originally you typed in your you put in your bank account information and you slid your card in and money came out that that's automating a human task. Right. It's very it. But it's not that intelligent. But now it scans checks and it and it's probably looking at your face when you're standing there too to make sure you're you. There's probably a lot of other things you don't even see that's going on. That's sort of basic AI. It's taking a human task and making a machine do it. Now, everyone thinks of AI like it's a robot and it's going to talk to me and learn anything. That's what's called
Starting point is 00:34:45 general AI. General AI is what everyone's scared of. That's where you invent something that can literally do exactly what humans do, learn like humans do, act like humans do, and yet has unlimited databases of knowledge and all this stuff. That's what people, that's what you get Elon Musk saying is, that's the singularity, that's all that stuff. That's what people, you know, that's what you get Elon Musk saying is, you know, that's the singularity. That's all that stuff. What everyone's talking about today, what I'm talking about at least is what's called narrow AI. And that narrow AI is more about solving a specific problem. And, you know, whether it's playing chess or, you know, the game go,
Starting point is 00:35:21 that's a very specific problem or even Alexa is a very specific thing. It's right. It's, it's, it's engineered to sit problem. Or even Alexa is a very specific thing. It's right. It's engineered to sit on your table and listen to your voice for certain commands and answer questions. Siri does the same thing. Now, what we do is kind of this invisible AI. You can't even really, you don't even really know what's going on, right? It's taking in data and then it's just saying yes or no on a return.
Starting point is 00:35:43 Or we're saying, saying hey look at this employee they might be suspicious it's not listening to anybody's voice it's not using computer vision it's but it is using an enormous amount of data and doing things automatically that maybe a human might have done historically um but it's trying to do that using a broader set of data. That's sort of my simple answer on AI. It's not, and there's all ranges of AI. The self-driving car is probably like the coolest, most visible AI thing that's come out lately, but there's a lot more going on under the hood and have been for 30, 40 years.
Starting point is 00:36:22 And I know when you've been speaking a lot about machine learning, I'd ask, can you kind of explain the relationship between AI and machine learning to the audience as well? Because I think you mentioned before big data is a buzzword. AI becomes an all-encompassing word, and your narrative and example is probably the best I've heard where it's sobering and says, hey, it's mimicking human behavior. Can you just give us kind of a brief overview of the relationship between machine learning and AI and what you're actually doing? Sure. Machine learning is, and if you go get a book on machine learning, one of the first things in there is going to be regression analysis or logistic regression. one of the first things in there is going to be regression analysis or logistic regression. And so it's really machine learning is an evolution of what historically has been called predictive modeling or just modeling.
Starting point is 00:37:16 It's taking a bunch of variables and trying to predict an outcome. That's the essentials of predictive modeling, of general modeling, of regression analysis, logistic regression. Now you have the more fancy math in there. You have things like extreme gradient boosting, deep learning neural networks, but they're all trying to do the same thing that the predictive modeling had tried to do all along. They try to take inputs or variables and predict an outcome, whether it be, hey, I'm standing here and I, I see 30 things going on. Can I predict what's the likelihood someone's going to steal something right now?
Starting point is 00:37:50 You know, I got a guy, he's got a hoodie on with dark sunglasses and he's been milling around my most expensive aisle. Okay. Now I've taken like five variables or 10 variables there. And I'm saying just logically in my head, what's the probability now a theft occurs on that aisle? That's inherently your brain doing a predictive modeling exercise. But you can set all that up. As long as you have enough data, a model can do that. That's really the connection. Now, AI and machine learning, machine learning is a subset of AI.
Starting point is 00:38:21 It's a tool. Think of AI like a toolbox or even the self-driving car. Self-driving cars is a great example. It has lots of pieces in it that are machine learning based. So the computer vision piece, you know, there's other pieces where it's looking at sensors in the bumpers. There's, you know, probably five layers of machine learning in there that are trying to solve different problems and telling that car whether to go right, whether to go left, and they're trying to avoid, you know, accidents. They're trying to stay on the road, stay between the lines. There's all sorts of outcomes they can predict and measure there within that body.
Starting point is 00:39:02 Now, within like a returns management platform like we have for Verify, our machine learning is really simple. All it does is take a bunch of data on the consumer, maybe on the employee, maybe on the store, and it's saying, what's the probability of fraud or not fraud? Or what's the probability this consumer makes 30 more returns in the next year, but no purchases? That's the sort of stuff we have an outcome to predict. And all those relate together in an AI system. So an AI might be just something that solves a specific problem, whether it's driving a car or doing an ATM machine for getting bank transactions or saying yes or no on a return transaction. The AI is the whole wrapper. It's all the rules. It's the software. It's
Starting point is 00:39:44 everything. Whereas the machine learning would be one algorithm within there that's solving one very specific problem. Hope that helps. And I would add, just to kind of piggyback on that, statistical modeling, traditional statistics, and machine learning, the objective of both of them is just to learn from the data. And as it is implied with machine learning, there's a lot less human interaction involved in contrast to traditional statistics where a statistician might use data reduction techniques, his assumptions about the data and its distributions,
Starting point is 00:40:21 whereas with machine learning, machines are exposed to the data, they learn from the data, and like David said, they get smarter in the student-packed loop. So that's kind of a little bit of the difference between the two. There's a lot more human interaction with traditional statistics. Great. Thank you. No, that definitely answers the question. I think it's probably the number one thing that comes up for me when people ask me off the side of, you know, what's machine learning, what's artificial intelligence, and, David, Daniel, the NRF published a short overview of the top tech investments right now in retail. And not surprisingly, advanced data analytics and artificial intelligence were both listed in the top five.
Starting point is 00:41:20 Just curious if anybody had a chance to read that and any points of view on the advanced data analytics piece of it and really explaining, you know, what the difference between data analytics and advanced data analytics would be to the user with some, you know, a real world example. Yeah, I view just data analytics more like reporting, whereas maybe advanced data analytics, you start to get into predictive modeling machine learning type stuff so i the difference between advanced analytics and ai i i mean that's just i i honestly don't know it ai is an advanced analytics in and of itself um you know you a lot of companies are going to go out there and propose report ai i have ai or i have this. The first thing I look at to see if they're serious is if you don't have PhDs on staff or guys with master's degrees or advanced degrees in a data science type background field, then they're just spinning. They're not really telling you the truth. So I think, you know, I looked at that article and I saw it and people are definitely trying to jump on the bandwagon and capitalize on this being a top thing. But, you know, there's people seriously doing it and there's people that say they're doing it.
Starting point is 00:42:35 And the big differentiator there is really what tech talent you have on staff. As far as the guy's predictions about AI and what people invest in, he kind of implying, hey, you better get ready to get into AI. Make sure the juice is worth the squeeze. Well, I think LP professionals have been doing that for years, right? If you deploy any solution, I don't care what it is, whether it's a sensor or an AI solution, you're going to test it and make sure it works. And so as long as you're going through with that process, there's nothing to be scared of here. You just, because usually even, even though the under the hood part of AI is a little bit scary,
Starting point is 00:43:12 you may not know what's going on. The outcome's usually pretty simple. And if the outcome's not really simple, then you probably need to stay away from it because it's, it's usually, I'm going to find a bad employer. I'm gonna find a bad consumer. I'm going to figure out who to give an offer to if it's a marketing campaign. And those things are very simple problems to articulate. Now, how you solve them and how you bring data together, maybe a million lines CFO and say, hey, if I deploy this technology, I tested it and I'm going to I'm going to gain, you know, I'm going to decrease shrink by 10%. Like that's the measure you should be employing. Don't worry about what the technology is.
Starting point is 00:43:56 But it's data analytics is definitely a hot topic and there's some cool problems to be solved with it. We're kind of a testament to that. But the way that LP professionals should look at it is, how can I measure it? And did it really work for me? And that's all you need to do. I think as data continues to grow, you know, we'll probably see a cultural shift. But AI is not going to be a panacea. And I think, you know, people are going to start integrating it as maybe part of a larger ecosystem.
Starting point is 00:44:27 But like David said earlier, I think you need to get a really good team lead, start slow, kind of get a roadmap and develop a strategy on how to use it. Make sure you have someone at the top who has a big picture idea and also can get in the weeds and understand what's happening analytically.
Starting point is 00:44:45 Right now, a lot of people are using it in kind of customer retention, recommendations for, you know, products for shopping. But I think we'll slowly start seeing a cultural shift with, you know, LP starting to use more advanced analytics. So this has been fantastic. I really appreciate you all joining us today. So, again, APRIS doctors David Spites and Daniel Downs, Control Tech, my partner in crime, Tom Meehan.
Starting point is 00:45:20 And I want to thank, as always, our producer of the Crime Science Podcast, Kevin Tran. Thank you again, everybody. And we look forward to meeting up again. Thank you. Thanks, Reed. Thank you, everyone, for tuning in to this episode of Crime Science. We also want to thank Bosh again for making this podcast possible.
Starting point is 00:45:40 If you would like to suggest topics for future episodes or provide feedback, please email kevin at lpresearch.org. See you next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.