The Data Stack Show - 263: The End of Busywork: How AI Transforms Productivity at Scale with Alberto Rizzoli of V7

Episode Date: September 24, 2025

This week on The Data Stack Show, AI entrepreneur Alberto Rizzoli shares his journey from early computer vision breakthroughs to leading the automation of back-office workflows at V7. The discussion e...xplores the shift from bespoke model training to configurable AI solutions, the impact of automation on business roles, and emerging best practices for integrating AI into enterprises. Listeners will gain insight into how AI infrastructure is moving from labs to everyday businesses, which roles are most vulnerable or secure amid automation, and why future-proofing your career means focusing on creativity, first principles, and continuous improvement. Don’t miss it! Highlights from this week’s conversation include:Setting the Stage: AI’s Hype and Today’s Innovations (1:16)Alberto’s Non-Tech Passions: Physics & UX (4:04)The Paradigm Shift: Machines that Adapt (6:22)Scaling AI: From Niche Apps to Mainstream Use (8:23)Large Models vs. Bespoke Solutions—Power Law in AI (11:07)Evolving Roles: From Engineer to End User (14:14)Simple vs. Complex AI Implementations (18:14)When to Scale from Simple AI to Production-Grade (22:40)Capturing Tacit Knowledge: Crowdsourcing vs. Centralization (27:22)The Challenge of Unstructured Process Documentation (30:08)Practical Impact: AI-Enabled Enterprise Leverage (33:29)ROI: Time Saved & Compound Effects in the Enterprise (38:08)Redefining Information Movers vs. Information Creators (44:14)Roles at Risk and the Case for Creativity (45:30)Alberto's Favorite New Tech & Future of User Experience (46:00)Spreadsheets, Business Logic, and AI’s Next Leap (48:41)The Data Stack Show is a weekly podcast powered by RudderStack, customer data infrastructure that enables you to deliver real-time customer event data everywhere it’s needed to power smarter decisions and better customer experiences. Each week, we’ll talk to data engineers, analysts, and data scientists about their experience around building and maintaining data infrastructure, delivering data and data products, and driving better outcomes across their businesses with data.RudderStack helps businesses make the most out of their customer data while ensuring data privacy and security. To learn more about RudderStack visit rudderstack.com. Hosted by Simplecast, an AdsWizz company. See pcm.adswizz.com for information about our collection and use of personal data for advertising.

Transcript
Discussion (0)
Starting point is 00:00:00 Hi, I'm Eric Dots. And I'm John Wessel. Welcome to The Datastack Show. The Datastack Show is a podcast where we talk about the technical, business, and human challenges involved in data work. Join our casual conversations with innovators and data professionals to learn about new data technologies and how data teams are run at top companies. Before we dig into today's data. episode, we want to give a huge thanks to our presenting sponsor, Rutter Sack. They give us the equipment and time to do this show week in, week out, and provide you the valuable
Starting point is 00:00:38 content. Rutter Sack provides customer data infrastructure and is used by the world's most innovative companies to collect, transform, and deliver their event data wherever it's needed all in real time. You can learn more at RutterSack.com. Welcome back to the Datasack show. We are here with Alberto Rizoli, and Alberto, I cannot wait to dig in on this episode because you've been working in AI for more than a decade, actually. And so I think you're going to bring a wealth of knowledge to our guests. But give us the quick one-minute background on yourself. In a minute, I'm Albert Rizzoli, co-founder and CEO of V7. I have started my journey in 2015 in AI. So shortly after the first Khrushchevsky paper to demonstrate it that GPUs can
Starting point is 00:01:29 train neural networks efficiently. And so I've seen all the iterations of AI hype and then AI dehypeization. And we're obviously in the most exciting time ever in the field. And I think we have enough of a technology to change the world in the literal sense. And the way we're doing it at V7 is by enabling people to automate the most boring parts of their work. Everything that involves document understanding, back office workflows that are slowing down enterprises massively, and that usually involve dozens or hundreds of people doing work that they frankly would rather not do and consider to be repetitive and giving them the chance to develop systems to automate that on their run.
Starting point is 00:02:16 Alberto, it's going to be really fun to talk about. So I'm super excited to talk infrastructure with you. We talked a little bit before the show. There is a wide variety in infrastructure and AI today that people are selling under these same marketing campaigns, and it looks really different. So we're going to dig in a little bit into that topic. What's something you're excited about chatting?
Starting point is 00:02:38 I think I'm very fascinated about what the end stage should look like. So today, we're all dipping our toes into AI workflows with in our own businesses. some people are taking it to the extreme quite successfully, but what I'm interested in is like, what will the Fortune 1000 look like in five years if they correctly implement today's AI paradigm? And what will the perceived roles that are affected by it
Starting point is 00:03:01 need to do to adapt and successfully adapt to all these changes? And then, like, what are the things that AI is still not very good at doing that we probably should not bet our careers on? Awesome. Let's dig in because we have a ton to cover. Alberto, like I said in the intro, what a privilege to have someone who's been working really at the forefront of AI for over a decade. It sounds weird to say that because the industry is moving so quickly. But before we dig into your background and then, you know, talk all about data infrastructure and AI and what you're building of V7,
Starting point is 00:03:34 I have to ask more of a personal question. So you've been an entrepreneur since you were 19. You've been working in the forefront of AI. and that's so interesting because, you know, kind of what you've known in your professional career is being your own boss, essentially, and working with the latest and greatest technology. If you had to go get a job outside of tech, what would you do? Like, what interests you? Great question. I've asked myself that many times before. Probably hard to answer after so many decades in the saddle as an entrepreneur.
Starting point is 00:04:11 Yeah, indeed. It's effectively, if I hadn't chosen the entrepreneurial career, I would still be building something. And the careers that I considered when I was in high school was either theoretical physics, which does not involve building something. And so I scrap that because you have to spend 15 years and then maybe you get mentioned in a good journal. And maybe you make a good discovery, and then you're done.
Starting point is 00:04:37 Essentially a teaching career. Teaching and research, yeah. Yeah, if you can't succeed, you end up having to teach in any academic field. But the other thing I was really interested in was user experience design, and specifically in the advent of the personal computer revolution, I think we had to figure out all these things that we take for granted that have to do with the user experience all the way from like the mouse and the graphical user interface.
Starting point is 00:05:04 These were brilliant inventions at the time because there was no pre-existing paradigm. And why I find AI so fond these days is that it's a repeat of that. We're finding out new ways to interact with machines that have this new paradigm of an LLM or like this token predictor. And I think we're maybe like 20% of the way there in really figuring out what in which ways we will interact with computers and intelligent machines in the more distant future. So in short, I would have probably been in the product side at a tech company if I had not to pick a founder group. I love it. I actually was reading recently just doing some research on the chat interface. and some of the inherent limitations there
Starting point is 00:05:43 and other things like that from a user experience standpoint. And one author mentioned that he said something very similar to what you said, which is that LLMs present the biggest user experience shift that we've had in 60 years and is sort of a major paradigm
Starting point is 00:05:58 where you're moving towards sort of an outcome-based, you know, you're telling the computer what outcome you want as opposed to issuing a command and then getting a response and that being an exchange where there's direct input
Starting point is 00:06:10 from, you know, a mouse or a keyboard or even, you know, buttons that you click or other configurations that you make, but you're actually giving an open-ended outcome and expressing your intent to the computer, which is fascinating. It's hard to believe that 60 years ago was the 60s. For a second, I thought, like, oh, but the GUI was a much bigger fit. But actually, it's around that time. Yeah, I think that's precisely it.
Starting point is 00:06:32 We're no long, we also are now interacting with machines that adapt themselves. They're almost like shape-shifting software, and it's able to adapt. to the need or tool that we need to use at the time. And there's an avenue by which maybe the feature is software is software that codes itself and generates what the exact minimum viable code that we need to run to do something. And it's very similar to sort of a science fiction or fantasy magic tool that just turns into whatever we need at the time.
Starting point is 00:07:05 Yeah. And yeah, that brings up a lot of different possibilities and also a bit more of a winner-takes-all approach to the world of software, which is a more concerning aspect. Yeah. Well, let's talk about science fiction a little bit because back in 2015, you were working on the cutting edge of computer vision technology and translating images into text across multiple languages.
Starting point is 00:07:32 And, you know, what really struck me about that, and I hope this doesn't come across the wrong way, But when I was, you know, looking at that technology and prepping for the show, you know, in terms of, you know, the timeline, it wasn't that long ago. But relative to what we have today, it felt really primitive. I mean, it felt like a very early prototype of, you know, sort of the types of things we have today. Could you give some perspective on that? Because back then it was really cutting edge. I mean, there was, there were very few technologies that could do anything like that. You point your phone at an image, it speaks the word back to you across multiple languages. I mean, that was really wild back then, you know, but now with GPT's real-time vision stuff, you know, it does kind of feel primitive. And that happened very quickly. Yeah, it happened quite quickly, and most of that progress happened in the last three years. Yeah. So I polly was an app effectively to capture images with their smartphone in real time. So think of it as a video. And then every frame of the
Starting point is 00:08:37 video was understood by AI and spoken back at you. So someone with the visual impairment could wave their phone around and hear what they had in front of them. And it had a vocabulary of about 5,000 words and objects. And if you were to try it, it would have a similar wow effect at the time as if you were to use GPT or Gemini's real-time vision features today. But only if you were an engineer because you understood the impressive capabilities required to make something like that work. And the actual proof points is whether your average person on the street would gain a lot of value from this type of experience, which is why it was successful with a very niche group of users, which are people that have very low vision and some enterprise use
Starting point is 00:09:26 cases that knew exactly what types of objects they needed to understand and therefore you could limit its capabilities. But it was a good proof point of why AI only succeeds when it's able to cover a lot of edge cases. A good way to think about AI is that it's always a distribution of data and the AI model models that distribution. Think of it as a bell curve where the total number of objects that you interact with on a daily basis is like the very middle of that bell curve. And if you have a product that doesn't capture the edges of the bell curve, which are the least common objects you interact with, you consider it to be a bit useless. In particular, because as the average person, you don't actually care about identifying a glass of water. You care about
Starting point is 00:10:11 identifying that one component in your AC, which is broken. And we've only been able to enable that at scale. Before that, you had to spend months of data labeling time and R&D to build a custom model to understand the inner parts of an AC unit, which only made it viable to companies. that had research budgets, and we're really focused on a big problem. That has completely changed. And the main reason is that the AI models are able to vacuum up most of the internet and then train on them. And we discovered that if we gave a lot of money to a few entities
Starting point is 00:10:47 that spend billions of dollars to train big models, we end up creating a better outcome than if a lot of small companies focus on their own domain and create bespoke models for that. It's unfortunately kind of a power law distribution aspect of the rise of AI. Yeah. Can we dig into the, you mentioned that AI is a data distribution and then LLMs are processing that. Let's talk about the infrastructure and the data layer. I know John has a number of questions here, but I'll kick us off with, can you paint a picture of the differences between just the underlying infrastructure?
Starting point is 00:11:27 And it's really funny that you said, you know, this would be amazing. It would be a crazy experience, you know, but only for an engineer. For the average person, it's like, okay, I put my phone at a glass of water and it says this is a glass of water. That's cool. But an engineer, you know, he would be, you know, amazed by this. What I actually think is a side note is kind of an interesting description of AI in its best form ever where, you know, it almost disappears and you have this sort of, you know,
Starting point is 00:11:53 really fluid experience. But can you talk about the infrastructure then and what you and your team had to build and the data processing layer versus what you're doing today at V7? How is this substrate of infrastructure changed? Yeah, absolutely. When we first started V7,
Starting point is 00:12:13 the majority of our workflows that we built were training workflows. So they were teaching an AI how to identify objects that general purpose AI did not know. And that aspect was essentially capturing thousands of images of unique objects and then loading them onto B7, which would use AI to pre-label some of them. So it would segment relevant objects. It would understand what you're trying to identify there.
Starting point is 00:12:40 And then having humans complete the rest of the work, add additional tags, adjust the labels. And then a machine learning scientist would train a bespoke model that can understand these objects. say, again, the parts of an AC or the specific components in a document, like the shipping address and the consignee address and the unique reference number, and only then you could start using these models. And if you had to change anything in these models, you had to go back into the infrastructure, add additional labels, restart the GPUs, and train a new model,
Starting point is 00:13:13 and then push it back into production. Now that has become a lot more fluid. A user can go into V7, drop the document, that they're looking to automate and simply explain in plain text what are the fields they want to extract. And if there's anything that AI is struggling with, you just explain in English a little bit more
Starting point is 00:13:33 on how to find those specific fields or what rules they need to follow in order to reason through, say, what is an acceptable termination clause and what is not an acceptable termination clause in a contract. This has changed the user base from being very technical
Starting point is 00:13:47 and scientifically-minded users to people that are actually doing the job that AI was being built for. And I think that's a massively fundamental shift. And the stack or the infrastructure now has moved away from something that leads you to training a model to something that leads you to just showing an off-the-shelf model like GPD5 how to do a particular task that is unique to you or to your business. And effectively, we have moved the workflow from the end goal is to produce a model to the end goal. is you start with a model and then given the inputs that you give it, whether there are documents
Starting point is 00:14:26 or recording of this call, what do you want to get out of it? And then you're configuring this workflow so that it can call external apps, send an email, or just understand within this unstructured data context, what do we need to extract and what guidelines and rules do I need to follow to evaluate what I just extracted? This model is a lot closer to either standard operating procedures within a company or a back office. And it enables these very back offices to effectively replace the most annoying tasks with AI WorkClose. The infrastructure has therefore moved away from the lab and into the hands of general users. And there's many different types of AI infrastructure as well. This is just the one that V7 specifically caters to. But
Starting point is 00:15:15 you can think of it as like the next generation of e-commerce websites, may no longer have a process by which they hire a photographer to go and take pictures of models wearing their particular dresses. But they start by generating them. And then if a product becomes evergreen, then they go through the effort of doing a more refined job, which is getting a real human person to do that task. And whether you like this process or not from a fundamental ethics perspective of hiring a human to do the job, it is the most likely course of action that we will see. And so there's now AI infrastructure components for almost every operation within a business. And what business leads into ask themselves is, what are the most likely ones that we
Starting point is 00:16:00 can apply today? And that would ultimately be fully solvable with AI so that it's worth investing the time and effort into it today. Yeah. Yeah, I think that makes a ton of sense. And I do think I think that's going to be hard for people, right? If you're in your example, if you, You know, somebody calls you up and said, hey, I know that we normally use you, like, to, you know, demo or try on our new products or whatever. And then, like, we don't need you anymore. We're replacing you with AI. I think that's going to happen to people. And I think, and we've got, we're going to dig into that topic a little bit more later.
Starting point is 00:16:38 But on the infrastructure side, I want to talk. We've got a number of engineer types that listen to the show. I'd really like to, like, paint this a spectrum from people of, So let's start with the marketing. The marketing says, like, we have AI agents. We have agent swarms. We have, like, whatever key buzzword you want to use. And it can do all these magical things.
Starting point is 00:17:00 Like, that's the marketing side of things. But then from an implementation side, from what I've seen and what I've talked to people, there's vastly different implementations. There are some very light-handed implementations that are pretty simple workflow tools and you do a little custom prompting. You need to like, you know, stitch them together and, you know, there's that. There's some very more complex implementations where you've got each component separate and you're doing, you know, vectorization and custom tuning of the models or maybe build your own model.
Starting point is 00:17:32 It'd be fun to kind of talk through your thoughts on that. So for, so when you talk to companies, like, I imagine a lot of these people, a lot of people that are kind of into AI, like, are aware of some of the, like, consumer-grade tools, like an innate end, for example. Not that you can use it, you know, not that companies don't use it for some things. But I'd love to talk through that tooling world from like a, you know, end user that's kind of like getting into AI versus like true production AI like at scale for some of these like Fortune 500 type companies. And what's the difference and what should people think about? When should they know like, hey, I'm hitting a limitation of like my, you know, alpha project to my little, you know, project that I'm working on the side? Yeah, great question. So I think the simple stuff will eventually be eaten by the LM providers themselves.
Starting point is 00:18:23 Yeah. And simple stuff is your typical like ZAP from three years ago, which was full data from this app, applied some heuristics, push data into this app. And that is actually a lot of tasks that are being done today. And just by nature, they exist because they're simple to configure. What's much harder to configure is build a system. that helps us decide whether to buy a building or not, or like an asset manager, which normally is something that you give to an analyst that is paid quite a lot,
Starting point is 00:18:56 and they have all this inherent information about how to do the job, what the company needs, what are the way that evaluates a particular property, and ultimately who to go to if there's a decision to be made. And this just expands the number of nodes by quite a bit, and the infra ends up looking like you still have inputs, which are usually a document, reaching you by email or a person just saying evaluate this to an agent and dropping in a PDF. Then there's all these reasoning steps in the middle, which is probably
Starting point is 00:19:27 the biggest value driver at the time, both for B7 and I think of AI in general. And then there's outputs. And a lot of startups tend to focus on either straight up output generation or making it easy to like zap here to take in inputs and then push out outputs. Because It's a very renewable to consumers, and it's a low-hanging fruit for a lot of small companies. But in an enterprise, it's very unlikely that you have like an ABC workflow. You have a lot of steps in the middle before an output is generated, and that output can be an IC memo telling people, hey, we should buy this building. It can be a slide deck or it can just be a decision.
Starting point is 00:20:08 Yes, no, or like, you know, let's update the CRM to say that this is a deal to be made. But what happens in the middle of the sandwich is really complex. There's knowledge bases that need to be queried. There's chain of thoughts that need to be developed. And some of these are inherent to the process of the company, and some of them are crutches to just the way LLM's work and making them work more efficiently. The former is the equivalent of memory.
Starting point is 00:20:38 You have a vector database. You pull information from it. That helps you understand. how to solve a particular task or sub-task. The second is about breaking any reasoning task into a smaller sub-task. And the reason why this is done is because LMs are largely trained on question-answer pairs that are part of short conversations. It's unlikely that there's a lot of training data in these models
Starting point is 00:21:03 that is of the nature of, here's an offer memorandum for a building, think like an investment manager, and tell me whether you should buy it or not, and then it ends up producing a 200-page thesis document. Very unlikely. And so the best way to do it is to break it down into individual steps where every step is effectively a prompt and expected response with customized inputs. The input usually starts with the core document.
Starting point is 00:21:32 As you move along, you may use some of these previous outputs, like, hey, let's evaluate the financial performance of this building from the past five years. Then you use that financial performance analysis as a downstream, input to another property, in this case, not a building property, but a data property, to then at the end come up with a final decision or output after all this like forced reasoning and forced filtering. And if you don't force it, you probably are not going to get a really good result that is up to spec to what a Fortune 1000 will do. The reason why we like these
Starting point is 00:22:07 reasoning models so much is that they kind of do the reasoning flow for us. And for every topic that they are smarter than us at or better than us at, it feels really good. But it's very frustrating when you are a lawyer, you ask a legal question to the model, and it starts making all of these assumptions that are not grounded in the way you actually do the job. And so a lot of this is about actually grounding the model into, we are this business, and this is how we run the specific process. If you follow these steps, you will be successful. Okay.
Starting point is 00:22:40 That makes a lot of sense. So if it's a simple, so let's think about like your original example. I'm using Zapier to, you know, currently to do this thing to like tag all my emails based on deterministic rules. I want to use an LLM to make this a little bit better. Now I'm going to use baby Zapier to better categorize my, you know, email. Very simple type test. So I guess, I guess from there, what I imagine, other than the technical side, so we're going to
Starting point is 00:23:09 kind of get into this, like, AI implementation, you know, tithek a little bit. I imagine that a lot of the hard part here, there's some hard parts with, like, you know, tuning models, getting context right. But a lot of the hard part here is still working with the humans to have SOPs of, like, what needs to get done anyways? If you've ever sat down with a team to, like, develop an SOP, or even we were talking before the show about, like, clear objectives. It's actually a fair amount of work just to like map out a clear objective and like the, let's use okay ours for example, like a clear objectives and key results. Like that's actually a lot of work to even do that. So I'm curious about strategy like maybe that you guys have developed
Starting point is 00:23:51 as you help people think through. Like what are we actually doing here in a system thinking where there's a bunch of people involved and it's not just one person's work flow? It's universally hated and the rules of construction are written in blood. And if you speak to anyone in the construction field, they're intelligent people who just want to do the thing, the way they know how to do the thing. And usually processes are only documented when something bad happens. And someone reacts by saying, okay, let's implement a way to do this properly every time. Somewhere, there's also an insurance company that it forces it.
Starting point is 00:24:28 Yeah, so you can get covered. They still remain universally hated exercises. And I think one way to hate it less is that as you develop the SOP, which we see it, just as a list of prompts effectively, you're actually seeing the work get done in real time. So one way we sort of solved it is that as you develop an agent in B7 Go, you're effectively adding all these rules in English sandwich, something that looks a lot like a standard operating procedure. And every time you add something, it recalculates your workflow.
Starting point is 00:25:03 and then you can see what the end results end up looking like. So you're almost like simulating the job over and over until you're happy with covering all the edge cases through the SOP. But the problem remains that almost no process, especially in enterprise, is known by one single person. And we're sort of like biased by usually being in startups
Starting point is 00:25:24 or being in small businesses where you have that one person that kind of knows everything and knows how everything should be done. But companies are a bit more like ASML. where are large companies where like there's no person that actually knows sorry at tsmc more so than asml but probably both of them no person actually knows how developing chip actually works end to end it's sort of like this wisdom of many people that know their specific niche and it's quite similar in even though allegedly simpler in enterprises and the difficult part is to get the right people to put the work into documenting how their specific slice of the pie works
Starting point is 00:26:03 And I think the only way to do it effectively is to give an incentive for people to just go into a platform and document this thing as a set of prompts. And I think the wrong way of doing it is to have effectively someone that goes and does all these user interviews like your typical for the upload engineer until they build a full mental model of the system. Yes, it works, but it doesn't work organically. And I don't have a full answer on what the best organic model for it is. But I think there is a future by which any company's biggest asset is its knowledge base and it is the aggregation of what AI has been able to observe across any of our work streams and document effectively their own opinion on how processes are run and that can be used as a knowledge base for running any other process.
Starting point is 00:26:54 Hypothetically, if you had an AI that was monitoring your laptop 24-7, I don't think that is a viable solution. But hypothetically, he would be able to keep writing to a document to say, this is how John does business. And then if it does it for John, for Eric's, for Brooks, for everyone, then over time, it starts to develop a pretty good thesis of how processes are run. Today, you kind of have to force that process it. You kind of have to get the humans to go and do that documentation. Yeah. Yeah.
Starting point is 00:27:23 Well, that's exactly where I was kind of going with this is, you're right. like people, most people really hate making, stopping slowing down, making SOPs, making an documentation, most people really hate it. And my question was going to be along the lines, like, well, you know, how do we use AI to do that? And to your point, like, maybe there is a future state where you can essentially have, you know, the AI, you know, recording everything, right? But in the mean, but that's probably not practical yet. So in the meantime, I saw an article this weekend, and this is what made me think about this topic, where there's some of these robotics companies targeting smaller manufacturers where you're able to essentially record
Starting point is 00:28:09 with your hands, like what you'd like the robot to do, and then the robot learns from that versus having to program the robot with code. What do we have that exists today that you think is useful with that same maybe mechanism with an LM, other than like a prompt, obviously, but Are there some similar things that would feel a little more natural to people, you think? Something that we tried is to use screen recordings as training data effectively to turn it into prompts. And it works surprisingly well, but it doesn't cover every edge case. But it is a good way to, like we do have some customers that come to us, say we don't have a documented process, but we have a lot of training videos on how to do this process.
Starting point is 00:28:52 So we put the training videos through an LLM and then through that we generate effectively adjacent and that can be imported into V7 to create a workflow. And that works surprisingly well. And maybe like the closest proxy to that in the research side is that computer use within NLMs is largely being taught as this like set of steps that a human is doing, like, you know, order a burger on DoorDash depending on where my address is. And here is like a list of my preferences.
Starting point is 00:29:22 And a human would go and try to fulfill that specific past and create training data for it. I think it's a, I think it's a good way to do that, but it's still, I don't know, I think it's, it still doesn't solve the problem of like, not everyone knows how specific process is run. But one thing that we do at E7 internally a lot is we use Lume to teach a man to fish for almost any process. So if someone asks like, hey, how do I register time off or request a budget to do this particular event, then we just say, send them a loom on how to do that. And that very loom could be used to just teach a model that can then straight up answer or do the task.
Starting point is 00:30:05 Yeah. Yeah, that makes sense. Yeah, it's super interesting because I agree that we're in the early stages of is there a really, you know, sort of elegant, you know, one-click way to do this, you know, that's still to be seen. However, as I think about the types of data, or let me be more, more specific, the data formats in which those processes exist or in which they can be extracted, right? So we have a loom video, we have, you know, Notion or Google Docs, you know,
Starting point is 00:30:38 that someone creates to explain the process step by step. You could even, you know, have someone explain it and spoken word. And all of that's, you know, sort of, you know, longer form unstructured data in various different formats, which is, happens to be like, you know, a marvelous input for a large language model. And so I'm pretty excited because I think that's going to rapidly become easier and easier because the formats are not a problem, whereas before that was a major, you know, super expensive technical problem. And that's not the case anymore, right? And so I'm pretty excited about what we'll see happen in the next couple of years. Yeah, 100%. It's like the field of meta learning and research is like learning how to learn might be the next, the next big hill
Starting point is 00:31:25 decline for any of these models. We're going to take a quick break from the episode to talk about our sponsor, Rudder Stack. Now, I could say a bunch of nice things as if I found a fancy new tool, but John has been implementing Rudder Stack for over half a decade. John, you work with customer event data every day and you know how hard it can be to make sure that data is clean and then to stream it everywhere it needs to go. Yeah, Eric, as you know, customer data can get messy. And if you've ever seen a tag manager, you know how messy it can get. So Rutterstack has really been one of my team's secret weapons.
Starting point is 00:31:58 We can collect and standardize data from anywhere, web, mobile, even server side, and then send it to our downstream tools. Now, rumor has it that you have implemented the longest running production instance of Rutterstack at six years and going. Yes, I can confirm that. And one of the reasons we picked Rutterstack was that it does not store the data and we can live stream data to our downstream tools. one of the things about the implementation that has been so common over all the years and with so many rudder stack customers is that it wasn't a wholesale replacement of your stack it fit right into your existing tool set yeah and even with technical tools eric things like kofka or pub sub but you don't have to have all that complicated customer data infrastructure well if you need to stream clean customer data to your entire stack including your data infrastructure tools head over to rudderstack.com to learn more Alberto, let's take a little bit deeper into V7 in this context. So you've brought up a couple of use cases, but I'd love for you to talk about what are the examples of where V7 is creating just immense, you know, sort of asymmetric leverage for a business. And I know you work in multiple verticals. I'd love a couple of those examples. And then follow that up with, you know, here's why we do not, here's what we're not building because we don't. think that V7 could be really great and create asymmetrical leverage, you know, for these other
Starting point is 00:33:27 use cases. Excellent. Yeah. So we work in several industries and maybe in order of size of business, it's financial services, insurance, real estate, tax and legal, and then a long tail from there that includes logistics. And what all of these industries have in common is that they handle a lot of paperwork. They have a general idea of how the paperwork should be handled, and very few people at these businesses enjoy that specific task. And the use cases themselves range from high complexity, high risk use cases that we can handle very well. For example, in finance, it's a confidential information memorandum, which is basically a pitch deck for acquiring the business. Requires five to seven hours to process from an analyst perspective to
Starting point is 00:34:19 turn it into an IC memo or like something that's getting to the investment committee. And an agent bill, 3B7 takes about 15 minutes of processing time, including some review time by the human. So that's a 50 next improvement on the actual time dedicated to it, all the way to much higher volume use cases like understanding tax claims, understanding logistics slips, or even insurance slips, which usually take, you know, five to, 10 minutes to process manually, but they are huge in volume. And they're usually a big cost carrier to the actual service that the company is fulfilling. I'm calling from the UK and the NHS
Starting point is 00:35:05 spends about 40% of its budget in administrative costs. That's our health care system. All these administrative tasks by and large can be taken care of by AI to reduce that time frame quite significantly. So there's a big cost saving in general, and we're doing things rather and efficiently today when you imagine that AI is possible. But on the other side, I think one of the biggest points of our OI is not easily measurable, and it is the lead time between an intent and getting a response back. So a good example of that is that we've all been in a pinch and needing something to be reviewed by an expert. An NDA could be an easy example that many can relate to, if you're doing a deal, 48 hours is a huge time frame. And like 24-hour
Starting point is 00:35:52 turnovers for NBA review is already pretty darn fast. But if you start to consider almost every process in a company to now be a five-minute turn around time, then the full aspect of work and productivity changes for anyone who perceives this lagging indicator. And if you imagine a business where almost every process requires two days to get back to you, any IT request, and that's fairly common outside of a startup where you can tap someone on the shoulder versus a competitor who has this system
Starting point is 00:36:29 that is just as reliable, but it only takes five minutes and that you get your response back, that becomes a massive quality of life improvement to work that I think will encourage people to, it's not really necessarily about doing a lot more, of the same thing, but being able to do more things because there is a lot more room for interpersonal tasks and creative tasks. And maybe to answer what do we not do particularly well are just those that I've described. We don't want to get into the way of human creativity
Starting point is 00:37:02 and human relations. I think fundamentally, I would love for a very smart 11 labs agent to answer my customer support queries. But that's generally because I don't need to develop a relationship with customer support. But we don't want to get in the way of humans developing relationships with their own customers or with their own people or getting in the way of what is truly creative work. Whilst AI is amazing at producing creative outputs from art to images to poetry, it is that good because it's read a lot of it. And in a way, it stunts our true creativity because it's not that good at creating something completely net to new. And I think a lot of creative aspects is about wowing the world by creating something entirely new, even if it's
Starting point is 00:37:49 like designing a new way of doing something, designing a new process. It can be a useful co-pilots to validate your biases and assumptions. But I think that is the work of the future, is designing we're close in the most efficient and fun way possible. And the execution of them will be what we use machines for. Yeah, that makes total sense. And to drill in a little bit on something you said, the difficulty in measuring the ROI of, you know, I mean, okay, if you just look at the raw percentage of, I go from, you know, a two-day review on an NDA or a contract with a vendor, which anyone who's been in a large organization and has had to go through that, you realize that there are, you know, there's this sort of class of people where they're really good
Starting point is 00:38:37 at their job, you know, in terms of just the raw skills that were on the job description. but they're very effective at their job because they know how to navigate the system more with more agility than other people, right? Oh, hey, I know this person down in legal and I did a favor for them or kids play soccer together or, you know, whatever it is, right? I know their schedule really well, you know, or I've done this 100 times and so I know the process and so they're just able to get stuff done more quickly, right? But that's a horrible use of their time.
Starting point is 00:39:09 that's a horrible use of their, like, energy is trying to navigate a really complex system to do something, you know, that could be done in five minutes. But I think the reason that's hard to measure is because of the compound effect, right? Like, in an isolated one time thing, it's like, okay, well, that's great. Like, my week at work was better. But if you multiply that happening a thousand times a week across the entire organization, the time you get back, the energy you get back is it really compounds, right? That's really, difficult to measure, but I think to your point, like, well, great, if I don't have to spend, you know, eight out of the next 16 working hours over the next two days, I just slot that off for navigating the system. It's like, well, I have that time back, right? It's like, well, that's really interesting. 100%. And I think, like, if you think about a sales career, the best salespeople in the world embody that, they find other revenues to remove friction from their ability to close a deal that are not the standard operating procedure. They meet with their clients.
Starting point is 00:40:12 They, you know, remember personal details about them. They put, like, they do extra work for them that goes outside of what one would normally do. And then they eventually become very successful. And the parts that they don't want to do, though, are, like, I need to wait two days to get this deal approved or this statement of work approved. because, and like the way they go around it is to, I don't, charm their way through it, to go and bother people, to go and push for things.
Starting point is 00:40:42 But these are all things that AI is not very good at doing. Yeah, yeah, yeah. Yeah, well, we're close to time here, but one thing I want to get your thoughts on, especially is you're implementing V7 in these companies. You're removing these barriers. Who is most impacted and what types of roles are at risk? I mean, you know, let's talk about,
Starting point is 00:41:03 the salesperson that you just mentioned, you know, they're really good at navigating the system. They're really good at, you know, figuring out how to, I don't want to say cut corners, but they know the process well enough and they've built up enough institutional knowledge of how things operate in the organization to where they can just get things done more quickly than other people, you know, and unfortunately, the cost of that is whatever, 40, 50% of their job is navigating, not actually selling or building relationships. that person, this is going to be, you know, you know, sort of like a miracle drug of productivity, but they were already super productive, right? Do you see that being the case where you sort of
Starting point is 00:41:45 have this subset, let's say, you know, half of a team, you know, over indexes for being productive already? Are they the primary beneficiaries of this? You know, who's going to get the most asymmetric value? And I guess I say who's going to, who is, you know, and how does this play out practically on a team, you know, at a company that V7's implemented at. Great question. I think there's effectively three groups of people in the world that follow any process. They're the ones that have a different way of seeing the world. And it is very hard to replicate it. And they're almost like an artist and an operator. And that will be the example of a salesperson that figures out that their biggest client goes to specific golf court
Starting point is 00:42:30 so they learn how to play golf to get there. It's a bit of a crazy idea, but they're thinking out of the box. Then there's like efficient or maybe sometimes less efficient operators that are following the rules and they're doing quote well. And then there's folks that understand
Starting point is 00:42:43 the actual operation at first principles to the point where they could teach it. And the ones, number one and three are the most secure. Number two is the least secure because they are, no matter how efficiently they're following their process, which could be developing leads and then doing sales,
Starting point is 00:43:03 if they're not having something super new to the process each time and really think outside the box, they're a risk of automation. The ones that really understand the process of the first principles have a good opportunity of actually developing those automations. Software is becoming very easy to use, and so they could be the ones that can actually automate group number two. Group number one, the creatives will likely not be automated
Starting point is 00:43:27 for a very long time, if not ever. And a good proxy to this is manufacturing. We had only artisans in the past. Today, we have people that design manufacturing processes. They help both factories. Sometimes they're engineers, but sometimes not. We have super skilled creative artisans that will create something that a machine could not make. And they just have this very unique way of seeing the world.
Starting point is 00:43:52 And then we no longer have artisans that do something similar to what an automated process can do, at least in modern countries. And so, yeah, I think effectively, like, you're either a crazy person that just does things completely out of the box, or you should start to think of how to turn what you're doing on a regular basis into a machine. Yep. I love that.
Starting point is 00:44:15 One thing that we've talked about before, John and I've actually talked about this and sort of explaining, you know, different types of people within an organization and kind of how to navigate things. But this is a dramatic oversimplification. But, you know, One way you can look at people within an organization is people who create information and then people who move that information around, right? And if all you do is move other people's information around, I agree. I mean, you should be worried, right? If you can't teach, like you said, they have an understanding in the process at the point where they can teach it, but that's actually creating, you know, new valuable information. So that's, yeah, super interesting. I've looked at it. I mean, very similar framing, but essentially two jobs and that definitely overlap.
Starting point is 00:44:57 One is that, like, creativity, creative endeavor job, and one is that, like, continuous improvement job. And you can do both for sure, but those are, those feel like the, and those are not specific to, like, a domain. Like, sure, you'll be applying that in, like, marketing or operations or whatever. But I feel like, that's what I've told a lot of people, like, if, like, continuous improvement, and that's, like, you, if you've got AI that can do this, then, like, how can AI do it better?
Starting point is 00:45:23 And then for their creativity, like, that, you know, has a million different angles you can take it. But other than that, like, I think there's a lot of change. A lot of change coming. A lot of change you got me. Well, Alberto, we're at the buzzers I'd like to say, but one more question for you, what is like an app or a product that you've used recently that's maybe new to you that really caught your attention? I'm just always interested for people who work in this stuff every day, who are building it at a foundational level. And, you know, especially with your interest in user experience, is there something you've used recently, even a physical product or even an experience you've had where it just really caught your attention?
Starting point is 00:46:00 I have, I think like many people have started to use video models more, and it's sort of, as someone that's been in the research side of AI for some time, it kind of turns into a light bulb moment where you realize that a next frame prediction engine can do a lot more than just create videos of the Bigfoot, but it can actually create full demo experiences down the road. And that could be a nice little game changer that are also interactive. On the interactive side, I forgot what DeepMinds name for those virtual worlds, but there's many startups that do it. One, for example, is Odyssey Systems, and it's effectively a navigable 3D world
Starting point is 00:46:36 that is entirely aged out of it on the fly. And you can think of it as like a video game that continuously renders new things on the spot. And those will be very useful for training certain models that need to be embodied, like self-driving cars and embodied robots. But I think they will also have a tremendous impact on entertainment, on VR and they're generally just quite fun because they're exactly like the human mind. We're imagining a world in our heads and then we're navigating through it. And we use that prefrontal low experience to think about whether our decisions are going to be
Starting point is 00:47:13 good or bad. It's actually what differentiates us from a lot of animals that we picture scenarios in our head and we imagine like, if I do this is going to happen, if I do that is going to happen. and we can do it for quite a long period. And so it might even give AI a much better way of reasoning that is not like just spitting tokens out and talking to itself. And yeah, I would say probably these are the two most exciting that are completely outside of our domain.
Starting point is 00:47:38 Yeah. We don't really deal with the generation of image assets. Another thing that we are experimenting with a lot that I'm having fun with is when you look at B7, it looks a lot like a spreadsheet, but it's actually a database. It's a big table. And every row is an agent run and various steps that it's taken and the information that is recent. Awesome.
Starting point is 00:48:00 Yeah, we're experimenting. What would that look like if it was actually a through spreadsheet where you're not bound to a peer grid where you have like columns and rows, but rather something that's a bit more fluid? And I think it's the perfect scratch pad for AI to operate in, but it's a thesis that we have not yet fully validated. But I think like of all documents that are going to change quite significantly with AI, I think spreadsheets are going to be the one with the biggest impact because you're no longer doing regular expression. You're doing AI functions and app tool calls
Starting point is 00:48:32 and running agents within a cell so it comes back with a particular piece of information. So there's a lot of U.S. changes that are going to occur there. Yeah, yeah, that's super exciting. I heard a product leader one time say, you know, you could essentially boil all software down to like forms on a database. You know, that's basically what software is, right? Yeah, so sure.
Starting point is 00:48:52 Which, you know, you're like, whoa, that's overly simplistic. And then you think about it, you're like, yeah, that's actually true, right? But I think what's interesting about what you're saying is that, well, and the point the guy made was like, so it seems like you could, you know, like, well, if Salesforce is just forms on database, like someone's going to build another sales force is like, no, the business logic is where they create the lock-in, right? And so I think what's really fascinating about V7 is you're abstracting the business logic. You're abstracting that out and turning that into an. AI layer, you know, on top of the database. And so it's going to be really interesting to see what happens the next couple years. Yeah, exactly. That's one way to explain it to an engineer. We are a multi-database system like Salesforce that is built for AI from the ground up. And as we built this
Starting point is 00:49:38 thing, we started to realize that there's maybe a chapter beyond this framework because of AI's capabilities to break out of the mold of an SQL table. Yep. I love it. All right. Well, Alberto, So thank you so much for the time it flew by. And we'd love to have you back on soon to hear about what's new with V7. Thank you. Thanks for having me. The Datastack show is brought to you by Rudderstack. Learn more at rudderstack.com.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.