PurePerformance - Back to Basics: Increase DevEx in the Age of AI with Laura Tacho

Episode Date: November 10, 2025

Don't get stuck using AI to build faster horses. Instead, find the opportunities and rethink your software delivery processes! That, and only that, will help you increase Developer Experience and Effi...ciency!This episode is all about how to measure and improve DevEx in the age of Artificial Intelligence. And with Laura Tacho, CTO at DX, we think we found a perfect guest!Laura has been working in the dev tooling space for the past 15 years. In her current role at DX she is working on the evolution of DORA and SPACE into DX Core 4 and the DXI Measurement Framework.In our episode we learn about those frameworks but also how tech leaders need to rethink where and how to apply AI to improve overall efficiency, quality and effectiveness! The key takeaways from this conversation areDevEx is all about the identifying and reducing friction in the end-2-end development processTech Leaders need to become better in articulating technical change requirements to businessAs of today only 22% of code in git is really AI generated. Don't get fooled into believing AI is already betterBack to Basics makes companies successful with AI. That is: proper CI/CD, testing, documentation, observability!Here the links we discussedLaura's LinkedIn: https://www.linkedin.com/in/lauratacho/DX: https://getdx.com/Cloud Native Days Austria Talk: https://www.youtube.com/watch?v=kZ1F0-XS1l4Engineering Leadership Community: https://www.engineeringleaders.io/

Transcript
Discussion (0)
Starting point is 00:00:00 It's time for Pure Performance. Get your stopwatches ready. It's time for Pure Performance with Andy Grabner and Brian Wilson. Hello everybody and welcome to another episode of Pure Performance. my name is Brian Wilson and as always I have with me my energetic co-host Andy Grabner how are you doing today Andy pretty good pretty good actually we had a little bit of sunshine today here in Linz Austria even though most of the day it's been gray like the last last week I would say yeah yeah sun is always good for energy if you say energetic maybe that's the reason oh there you
Starting point is 00:00:53 go see I'm getting my energy because I just had my double espresso and we have cold and gray skies out yesterday and today. So it's been a little bit of a downer. But we're heading in. So people listening, I'd say, you know, it's going to be quite an experience today because if you listen to this episode and the next one, we're actually recording these back-to-back, which I don't think we've ever done before. We've done like day after day. We've never done back-to-back episodes recording. So it'll be interesting to see how much steam we have second time or if we're just so stir-crazy that it just becomes wild. We have no idea.
Starting point is 00:01:29 You'll have to tune into the next episode to see what kind of experience you have. See, Andy, I try to do a poor man's, Andy. Just a reminder, Andy is the king of transitions. So, or segways, if you will, right? He's always great at taking our dumb banter
Starting point is 00:01:47 here and leading it directly into the topic. And I tried that. The thing is, if we keep talking like this, we probably just run over and we won't have a chance to actually get our guests on the line. So that's why I think let's make sure she has a great experience as a first time guest on this show. Let's do this. Laura Taco, thank you so much for being here and being our guest at the Pew Performance Podcast.
Starting point is 00:02:13 You are the CTO at the DX, the developer intelligence platform. I had the pleasure to welcome you on stage at Cloud Native Days, Austria. It was just like three weeks ago. And when I listened to your talk, I thought I would like to have. have Laura on the podcast because you've been talking about a very important topic that I think we have covered a little bit over the last couple of episodes, but hearing it from you, it's just great. To get started, Laura, maybe a quick introduction, who you are, what you do at DX, besides obviously being the CTO, what I just said, but what is driving you, what is
Starting point is 00:02:51 motivating you, and then I have a couple of questions for you. Yeah, my pleasure to be here with both of you. I'm Laura Taco, CTO at DX. I would say, actually, I have been working in developer tools for the last 15 years. And so my role at DX, we can kind of see as like a capstone. If you're listening to this and you've been involved in Docker and cloud native for a long time, you probably know me already because I've been working with Docker, started working with Docker in the very, very early days. I've always been kind of obsessed with developer tooling, cloud tooling, and trying to find ways to make the day-to-day work of developers just easier so that we can focus more on what's important.
Starting point is 00:03:33 Of course, that varies company to company, individual to individual. At DX, what we do is we are an intelligence platform. So we're taking data from your systems. We're taking data also from developers themselves through self-reported surveys, marrying that together, and then giving organizations basically a map to understand where the friction is, where the pain is, so that they can improve the experience for developers, remove the friction, and get to better outcomes. I lead a lot of research at DX.
Starting point is 00:04:04 We're very research-backed. So we work with a team of researchers. We also have a tremendous amount of data, as you can imagine from all of our customers that we can aggregate and look for patterns and trends, which is a lot of what informed my talk that I gave back Cloud Native Austria because AI is, of course, a big hot topic. and there's lots of claims in the headlines and then having a front row seat
Starting point is 00:04:27 to how, you know, 450 companies are actually doing it on the ground. There's a little bit of a discrepancy, and I think we'll talk a little bit about that today as well. But kind of the bottom line is I'm here to make developers lives easier, take friction out, make sure that leaders are equipped with the right data and the right storytelling around
Starting point is 00:04:46 why it's important to invest in things like CISD, performance monitoring, all of the tools that make life back. better and lead to the right business outcomes. Cool. And an interesting fact is, if I remember correctly, you're originally from Wisconsin, but now you live in Austria? I am.
Starting point is 00:05:03 Yeah, we're like Austria two-on-one here. Yeah. To Austria, two, to one in America. But, yeah, I was born in Wisconsin, strangely, or maybe not strangely, but it's very common to have German ancestry. So my great-grandparents and great-great were born in Germany. They came over to Wisconsin. Then I went, turned around right back to move to Germany about 10 years ago, and I live in Austria.
Starting point is 00:05:30 Well, I hope you enjoy your new home. We definitely enjoy it. As I said, your keynote at Cloud Native Day's Austria. Laura, in preparation for this podcast, I went through your talk again. And folks, by the way, if you're listening, we will also add links to the recording of the talk because Cloud Native Austria already published all of the videos. really check out the talk and Laura I also know you've done obviously this is not your first podcast I just listened to a podcast that you recorded in September and also took a lot of notes so we will add as many reference links that we discuss today to the podcast summary one of the
Starting point is 00:06:12 things that I want to start with when I give presentations around platform engineering which I do quite frequently I typically start with two metrics that I took from the state of DevOps report, which I think two years ago, a year ago now, the Platform Engineering Edition and they talked about that only 40% of engineering time is spent in productive activities and that I think 36% of developers are leaving an organization because of bad developer experience. And I always bring this up as two numbers, which is for me, first of all, the productivity is already shocking enough, but also as a leader, in the leadership role, where I want to retain the talent that I have in my organization
Starting point is 00:06:56 and I see 36% of engineers are they want to leave because of bad developer experience. This is very alarming, yeah? And I wanted to ask you because in the podcast and also I think in your presentation that I've opened here, you made a very interesting and nice statement. When we talk about developer experience, developer experience is not just ping pong tables and beer taps.
Starting point is 00:07:20 So I would like to get, maybe a definition from you to get started. What is developer experience? What are the things that organizations need to look for to know whether they're doing a good or not a good job? How can they maybe measure it? I know you have a framework. But let's start with what is developer experience from your perspective
Starting point is 00:07:43 and what are the things that are indicators that things need to change. Yeah. I think the question of like what is developer experience is an interesting one, because sort of at its heart is, you know, how do developers think and feel and perceive the work that they do day to day? What is their experience doing work in day today? I think the tricky thing about developer experience is that oftentimes we're not searching for the existence of something. We're looking for evidence of absence, which is really
Starting point is 00:08:13 difficult. Developer experience is the absence of friction in the system. And if we think about, I mean, we can like, let's think very kind of big picture and tie this. You know, you have green tests that verifies that the cases that you thought to test about were, you know, they're green. It doesn't tell you that your software works. It's very, very difficult to prove the absence of something. It's very difficult to prove that the quality isn't there. It's easier to prove that, you know, tests are passing or we have an incident or something. The measurement around this, therefore is just really difficult. We want to prove the absence of something.
Starting point is 00:08:55 So we can look for ways to detect then where is the friction in the system. And so when I'm talking about developer experience, I talk about, you know, a developer getting work done, whatever part of the SDLC they're touching, how much of their time is being wasted due to inefficient processes, lack of tooling, how much time do they spend looking for information they can't seem to find and that they need that? How much cognitive load do they have? Are they getting feedback quickly enough? Do they have enough time to focus on things?
Starting point is 00:09:25 So it's a kind of a mix of all of those things together. I think very importantly, two things I would say developer experience, it's the absence of friction in the system, and it must be measured with multiple dimensions. And there is no one metric that can tell you if you have developer experience or not. I really like the, thank you for that definition. I think this will become maybe our,
Starting point is 00:09:49 our snippet on social media to promote this session well it's hard to say because we still have at least half an hour to go so maybe it got a half hour still exactly but you came out swinging strong
Starting point is 00:10:01 so yeah yeah so I do know that you know if looking at your website and looking what DX is doing you've obviously
Starting point is 00:10:11 you've defined your own framework of metrics and as you just say there's more than one metric multiple dimensions Many of our listeners are probably familiar with the Dora metrics because they've been around for a while. They might be familiar with the space framework
Starting point is 00:10:26 because we've also had sessions about it. What does DX now do? What do you define? What's the next evolution of these frameworks? It's a great way to put it as the next evolution because we work with the Dora team. We work with the researchers that did space. DevX framework was the next evolution after space,
Starting point is 00:10:46 also with Nicole Forsgrin and Dr. Margaret N. Story, who worked on space and worked on Dora. If you've read Accelerate, those names might kind of stick out to you. The DECS framework is all about cognitive load, feedback loops, and flow state, so having enough time to get into the workflow. Those three things together, Dora metrics, which talk about software delivery capabilities. Then we have the space framework, which isn't a set of metrics. It's not a list of metrics, but it's a framework for how to think about developer productivity. all the way from satisfaction to performance, activity, communication, and then efficiency.
Starting point is 00:11:22 Then we have DebX, which is about cognitive low feedback loops and focus. Those things together, it's a lot of ground to cover. And organizations, even though we had all this research from over a decade, organizations were like, what? Okay, but what do I put on my dashboard? You know, what's the number or what's the metric that I should bring to my board or my exec team when they want to talk about how is my organization performing. So we worked with all of the researchers from DORA space and devX. We put together that body of research and we made
Starting point is 00:11:53 a recommendation, which is the DX core four. This framework takes all of the research from the last decade that we have on developer productivity, but turns it into an actionable list, short list of metrics that's easy to deploy and then also benchmarkable. So if you're trying to get started with metrics, there's the DX core four framework, is absolutely the thing I would recommend. I co-authored it. So, of course, I'm going to say that, but I really believe it's the right thing. I co-authored it. It's just, it's there to help organizations get started faster. Cool. Can you enlighten us quickly? What are these four core metrics? Yes. So we are talking about speed, which we're measuring in terms of merge frequency, PR, throughput.
Starting point is 00:12:40 aggregated, never at the individual level, but at an organization team level. Developer experience, and we're measuring that with, you know, when you unpack that, there's like quite a few other dimensions to measure it. Quality, which is our good friend, change failure rate coming from Dora, and then impacts, which is the innovation ratio, how much time developers are actually spent doing new work, innovative work versus maintenance, quality issues, all those other things. So looking across speed, effectiveness, quality, and business impact. This gives a common language to speak about developer productivity
Starting point is 00:13:14 and engineering work performance across the organization. But then also we have just those, you know, those four numbers they need to be looked at together. And they tell a story about quality, innovation, developer experience. And we see that organizations that focus on that developer experience piece, they see the other three numbers rise because developer experience is the core of everything. And that is the place to focus. that's the lever to pull on for organizations that want to improve. That's cool. I was just going to say one thing I noticed,
Starting point is 00:13:48 and I guess this is where the humanity meets the numbers, right, is those are all metrics that are going to be numerically based. There's going to be quantitative, yes, quantitative metrics to see what these numbers are. And I suppose the idea then is when you see, efficiency down or any of these metrics being down, the idea then is not just like,
Starting point is 00:14:18 all right, let me go yell at developers to get more efficient or something. It's let's look at why. That's when maybe the anecdotes come in. That's when you're talking about discussing with the developer. So it's not just a matter of, you know, if you think of a production line, right?
Starting point is 00:14:36 We didn't produce enough cans this time. It's not about let's make people work faster and it's that's fine where the inefficiencies are maybe those inefficiencies or the developers aren't happy with something right whatever it is but that's where like the human then steps in to investigate what's creating the friction there to help figure out a way to remediate that yeah you've you've got that absolutely right dx core four is the only kind of developer productivity framework that has the developer experience as a core pillar those are self-reported qualitative metrics coming from engineers, we have to look at that. We have to look at that part of
Starting point is 00:15:12 the story because what I see so often is, let's say, for example, change failure rate, we see that rate go up, meaning there's more escape defects, quality is going down. We can look in the developer experience measurements and see that developers are already, they've been saying, I don't feel confident when I release changes. I think it's going to break production or, you know, our build and test process has way too many flaky tests, just not reliable. Those are leading indicators for these lagging indicators that are, you know, that the business tends to pay a little bit more attention to. And so, you know, when we want to focus on improvement, it's going back to the, what's the leading indicator? Where's the bottleneck? Where's the friction? Can we remediate that problem
Starting point is 00:15:50 and then see the results downstream, just like in your can factory example? Do you provide, I guess people can obviously go to your website and go through all of the material, but do you provide guidance on how? how people can start measuring these things themselves? Is there anything where people can get started? Does it always need a professional service engagement or tooling? Or can you quickly give us an update on that? Yeah, there's a couple articles, and we'll link them in the show notes.
Starting point is 00:16:22 But all of the research that we do at DX is open source, for lack of a better word. We're not taking contributions because it's like a team of researchers that we have our methodology. But it's open for anyone to use, and you can do it by yourself. that was very important to us, especially with the core four, that any team with access to Google forums or Microsoft forms or whatever can collect all of these metrics self-reported without needing very sophisticated ETL pipelines, and they can get a baseline very quickly, compare it to benchmarks, benchmark against themselves, and then start going. So I have a template for a survey to use to collect all of the core four metrics, including the breakdown of
Starting point is 00:17:05 the developer experience drivers. And that's just in a Google sheet that we can link here. And then I did a pretty long form article on exactly how to use it, when to use it, if you want to collect metadata, all these really descriptive things in Lenny's newsletter, which I'll also send you. We can link. It is behind a paywall if you're not a subscriber to Lenny's newsletter, but it doesn't start until like 80% down the article.
Starting point is 00:17:27 So you still can get pretty much a lot of it without having to be a subscriber, although it's a great newsletter. Yeah. Well, folks, definitely give you email, subscribe to the newsletter and get continuous updates on that topic. Laura, there's obviously a lot of information already about this topic out there. And as we said in the beginning, we want to make sure that we give people, our listeners, a quick overview. But then I also want to dive into a topic and kind of putting the AI elephant in the room. the AI is a big topic and also your talk at the cloud native days it was developer experience in the era of AI and for me the biggest aha moment and I think I also posted this on
Starting point is 00:18:18 LinkedIn later on you had one slide up and I have it here now where you said you did another it did research right and when we think about AI for developers at least the first thing that comes to mind is cogeneration right makes a lot of sense But then the question is, if code generation can make the developer 20, 30, 40 percent more efficient in writing code, what does this mean for the whole developer experience and developer efficiency in productivity? Because you then had a slide up that says, where is most of the development time actually spent or where the developer spent most of the time? And there's only very little in actually coding. There's so many other tasks that are, and excuse my language, time suckers, that are taking away the time, that are making developers improductive.
Starting point is 00:19:09 And I thought this was an aha moment for me. We're always talking about using the latest model to generate more and better code, yet we would only impact 10, maybe 15 percent of the developer time because that's how much time on average they're spending in coding, but they're spending much more time somewhere else. So my question to you now, where do organizations, where do developers waste their time? And what do organizations need to look at? And how do you measure this? What I can say, let's start with a little bit of data about how much time savings developers in general are getting back from AI,
Starting point is 00:19:45 which is only about three hours, 45 minutes, a little bit less, actually. It's strangely gone down, not significantly, but it's gone down a little bit since last quarter. So we're looking at like three to four hours per week. of time savings. So this is not 50%, this is not a huge amount of time. This is maybe 10% of time. Because, Andes, as you said,
Starting point is 00:20:10 we are, as a developer, not spending 100% of our time coding. And so when we look at the surface area of optimization, if we're only using AI coding assistance for code generation, the surface area is already only 25% of the time
Starting point is 00:20:24 that developers are actually coding. And then when we think about how much time within that coding time is appropriate for AI to augment, we're down to even a smaller chunk. And so what happens is that we might get some local optimization in this task acceleration, but those gains are lost in the system because code review is still a bottleneck, information seeking is still a bottleneck. And in fact, we have had AI tools now in organizations for about three years. If I look back to data three years ago,
Starting point is 00:20:56 the two biggest points of friction and two biggest opportunities for time savings were reducing the number of meetings and reducing the number of interruptions. Those two things have remained steady. We're three years into this AI transformation. We just pulled the data actually
Starting point is 00:21:16 for our next Q4 AI report. Sure enough, those two things are still, they are still the highest points of friction for developer teams overall, you know, for a very large sample size. This is due to, you know, it's information seeking. I can't get the information that I need. I'm waiting on a decision from someone. My feedback loops are too long.
Starting point is 00:21:42 Documentation is poor. I mean, there's a million reasons. There's a million reasons where that coordination cost is just high. I'm not suggesting that every organization should stop having meetings. But what if we used AI to solve those problems instead of code generation, when organizations start to do that, that's when they're seeing much more rapid acceleration toward whatever goal they've set with AI. Yeah, so maybe use AI to make the meetings more efficient, right? Maybe using AI to figure out who should really be in that particular meeting and don't invite many people because you think they can contribute. maybe use AI to write better, you know, agendas for meetings and like prepare for meetings.
Starting point is 00:22:26 So the meetings just keep a better outcome. I would even take it five, you know, five steps forward and think about why do we have so many meetings anyway? Why do we have so much? There's so much back and forth. Can we structure our teams and work in a different way now that we have AI? And perhaps there was a lot of handoff between design and PM and engineering so that we could design things in back. validate them in the early stages and then move into execution. But what if that doesn't have to be the case anymore
Starting point is 00:22:55 because your designer and your PM with Claude and some very good quality checks and things can make some prototypes and validate things easier. And now instead of a project needing that four weeks of planning and back and forth in validation, we can get that same thing done in a week. I've seen companies do that before. Takes a lot of effort, but that's quite a, you know,
Starting point is 00:23:19 it's not a, I think what organizations are getting trapped in right now is, you know, faster horses. If you know that's saying about, you know, if we would have good people what they wanted, they would have asked for faster horses when we really need to reimagine altogether. And maybe that's an automobile, maybe it's a rocket ship. And the companies that are just rethinking things, the whole value stream, are seeing a lot of acceleration because the system is the problem. It's not the individual task. That's the problem. Because the typing speed has never been the bottleneck in development. It's always something else.
Starting point is 00:23:55 Yeah, I would take it one step further and say, just send all AI chatbots to the meetings and then see if there's any difference. You know, you hear horror stories. We can laugh about it, but like people do that. People do that. Or it's like, you know, you make a chatbot to apply for a job and then it gets trapped in a conversation with the chatbot of the recruiting tool of the company. and it's just like, can we just, this is a waste of time,
Starting point is 00:24:21 can we just have two humans talk to each other? Yeah. Well, thank you so much. I really like the, I think it was Henry Ford, who has been attributed to that quote with the faster horses versus a car, right? I think we need to rethink engineering in the era of AI, maybe as AI native companies,
Starting point is 00:24:44 it means we need to rethink on how we engineer and what the process is to engineer a product from design phase to first implementation, to test validation. So there's a lot of opportunity to change the way we build software. And with the goal, obviously, to reduce these tests that are currently consuming a lot of time because we're doing back and forth and interruptions and all that. There was another interesting metric. There was, I think it was number four, the CI wait time.
Starting point is 00:25:18 So developers waiting for feedback from their continuous integration. So does my code actually build, do my tests, all run? It strikes me because we've been talking about how important it is to optimize CI, to provide faster feedback loops to engineers. And yet it still seems, based on your research, that this is still, it's in the top four. How is it still possible? I think one thing that a lot of engineering leaders lack the skill of is building business case because we it's like historically not really been a skill that we've had to exercise a lot.
Starting point is 00:26:02 You know, I even think back to my time as hands-on keyboard engineer and then my early days in leadership. It was sort of like, well, we need this because we say it's important and trust us. And it was always sort of like a, you don't know what we're doing because you don't understand code. And so you have to trust us on this one. And this is the thing that we need, which sometimes it worked, but sometimes it didn't. And organizations, you know, the budget holders sometimes don't, they need a little bit more to go off of. And I think it's been difficult as we have moved toward more accountability for how engineering is allocated in terms of like total spend. But also as we move toward, you know, if we have the option to optimize our CICD pipeline versus build a new customer facing thing, how can I, as an engineering leader, articulate the business value of this investment in plumbing against a concrete new feature that we know has X revenue potential?
Starting point is 00:26:58 And it's a lot easier to draw it. It's not apples to apples, you know, but we can quantify the impact. We can think about how many developers is this impacting. How much weight time is being lost? We can talk about it in terms of salary recovery. We can talk about it in terms of capacity creation. We can link it back to research like the DevX framework that shows us that feedback loops is one of the three major components that are research-backed tied to better organizational outcomes. There's a lot of things that we can do.
Starting point is 00:27:30 But because engineering leaders, we haven't been trained in it, it's really a new skill for a lot of engineering leaders to practice to try to talk. about investments in very kind of peer engineering things in that way. And so a lot of the time it's just been neglected or passed over because, oh, we're going to go chase the new customer facing thing. One of the things that I care a lot about is giving engineering leaders the phrases and the words and the mental models to articulate the value of why things like CI wait time are so harmful to your system and actually hurt organizational performance. and how we can change the conversation to not being engineering versus product,
Starting point is 00:28:13 but actually thinking about improving the CICD pipeline is a product investment because it's going to enable us to capitalize on all future customer-facing features much faster, and they're going to be higher quality. And what's that worth to our business? Quite a lot. When I go to a senior leader asking for budget and I say, hey, I have a way that we can execute 10% faster on every single project going forward, and we're going to reduce maintenance costs and quality overhead by 25%.
Starting point is 00:28:42 Any engineering leader or budget holders can be like, tell me, we can we start yesterday? But when I go and say, hey, you know what? Our developers are complaining about CI wait time, I need three weeks to finish it, and we're going to have to delay this project. That doesn't really fly. And so we need to change the way that we talk about things in order to get the results that we want. Engineering is there to support the business. You know, I know earlier, if I think back to earlier in my career, I'm like, oh, this is cool.
Starting point is 00:29:10 I get to build cool things that I'm interested in all day. And I, you know, as I grew up and kind of went up in leadership, I realized that importance of tying what I'm doing back to how it benefits the business. Because at the end of the day, this should be enjoyable, but it's also not a club. It's not a coding club. Yeah. I just had a I think we had a similar conversation a couple of months ago
Starting point is 00:29:38 I was in Prague at the engineering leadership conference they have a great community so Laura Prague is just around the corner from you where you live right now I want to give a quick shout out to Marian Kaminista. He is actually
Starting point is 00:29:54 leading this community and I was invited to speak there about platform engineering and then we had him on a podcast and I think he made the same claim. He said, technical leaders need to understand how to speak the language of business in order to really have an impact right. And so definitely spot on and also what I hear. And maybe this would also be good for Laura for you.
Starting point is 00:30:18 If you want to make a trip to Prague, I'm pretty sure. There will be a great community for you also to chat with. And I assume you have been because I know we also had Tucson. on one of our podcasts a while ago, who was also a technical leader. Dusan Katona is his name. He works at Atta Kama in Prague, or in Bruno, actually, but I met him in Bragg. And he mentioned that is also implementing the DXQA44 metrics right now. Yeah.
Starting point is 00:30:53 Yeah, it's a great, lovely city. We had a team offsite there one year. That's very memorable. I only have good memories of Prague. I would like to, so we talked about, if I do a little bit of a recap, we talked about the metrics, the four-core metrics, speed devX,
Starting point is 00:31:14 quality and impact. We talked about that just using and thinking about AI as a code generator and helping developers to code is not really leveraging as its full potential because there's most time spent in other tasks. But you brought up a good point instead of just thinking, how can we use AI to reduce the amount of meetings we have for time?
Starting point is 00:31:40 Why not rethink the way we engineer, rethink the way we do things? And if you want to make these drastic changes, then tech leaders that need to drive these changes need to understand how to communicate the benefits to the business because it doesn't help you if you say the CI needs to be 10% faster. You know what's interesting with that, Andy? It triggered a thought in me
Starting point is 00:32:04 where we're trying to retrofit AI into the developer experience, right? But if you think about the transformations recently, a few years ago we were talking about cloud native. Recently, we've been talking about AI native. And I think instead of trying to back, you know, retrofit AI into the current process, it might be beneficial.
Starting point is 00:32:29 And Laura, I don't know if you've seen any of this where people have taken small teams and started experimenting to say, okay, we have all these different uses of AI. Let's imagine fresh what a new developer workflow would look like with these AI tools and try building it out
Starting point is 00:32:46 instead of just where can we plug AI into the existing system. Right? So developer native, if you will, maybe. Yeah. There's some orgs that have talked about what they're doing. So Travolopia at the Enterprise Tech Leadership Summit was a great presentation. And they talked about taking 10-person teams and turning them into four-person teams.
Starting point is 00:33:10 What they're doing is having designers literally sketch out things like wireframes on paper. They take a picture of it and then have that generated as a prototype in the browser, going from screenshot to browser. We can do that now. Once it's there, we can validate much. more quickly. We can put it in front of a customer and have conversations about it. We can think about load testing. We can think about all these things much earlier in the process. We can bring engineering in. And then once the design and the spec is really settled on, the engineer can
Starting point is 00:33:46 coordinate agents to do some of the work while also staying in the loop and making sure the quality checks pass. And they've had a tremendous success with this. And I've seen lots of other organizations have tremendous success with moving, just really fundamentally changing the team composition and the workflows. When are things handed off? Who's participating in which section? Which personas or job roles get access to the coding tools and which personas are now coding because you know, designers usually didn't, you have to have a GitHub account. Like, now they do. Now they might be doing prototypes. And just that prototyping and validation is a huge unlock because we can talk about, you know, saving incremental time three hours, four hours a week of speeding
Starting point is 00:34:35 up coding. But where most businesses are losing the most time is that they build the wrong thing. A whole team spends two weeks on something that they decide, oh, actually, that was the wrong strategy. We're building the wrong thing. We need to go do it. Or they build a feature that doesn't have any business impact. That's where the real inefficiencies are as well. And so I'm very excited about using AI to empower that kind of classical maybe product trio. I think that probably needs to change. But the designer, product manager, someone who's on the business side, an engineering lead to validate things and just keep those feedback loops way tighter and much closer to each other. And so that really that team owns everything from, you know, from police to thank you
Starting point is 00:35:21 throughout the whole process. It's exciting. We'll see what happens. Yeah. Because as you said, right, I think the, was it, who was it? I think it was Guaranka, who we had on in the very early days of our podcast. She worked with Facebook and she was giving us some insights on how they're doing feature testing and feature evaluation at Facebook, right? And they basically said, you know, we have an idea of a new feature. It doesn't have to be perfectly implemented, but get it out quickly, get feedback on how many people are actually adopting it ahead, then to specify a success criteria. So if not, within the first month, 20% of the Facebook user base is using that feature at
Starting point is 00:36:05 least once a week. If you don't meet that bar, it doesn't make sense to actually invest and finalize that feature and optimize it. And for me, what you're saying is we, AI now is able to revolutionize the way we would have built this because it would have needed many different people to build this feature and get it out there. But now with AI, you can quickly create a prototype, give it to your test customer base, get feedback, figure out, is this, you know, a feature that we should keep on up?
Starting point is 00:36:34 Because I think we know with statistics, at least I'm sure there's many reports around, that only maybe, you know, 20% or 30% of features actually really take off. And so how can we get faster to understand which 30% we need to keep and invest and the 70% we will need to throw away? And if AI helps us there, that's a huge improvement. Yeah. Yeah, that validation step, you know, most companies are of a certain size or more and plus one and bigger are using something like product board.
Starting point is 00:37:06 they have a ton of user feedback that they have in a centralized system. And what I've seen some companies do is take all of that user feedback. They have longitudinally going back in the past, however many years. And when they create a prototype, they use that data to create certain personas. And then they actually use AI to test their feature against these like AI generated personas and use AI to coordinate the validation or even user acceptance testing of the features before it even goes out to like an actual real human being to point out flaws or point out under-solved problems and things.
Starting point is 00:37:42 I think there's a huge, you know, there's a huge amount of potential for AI in this part of the SDLC. Also on the architecture side as well doing, you know, even like generating huge data sets or generating, it's, AI can be such a powerful tool in early validation. before we go chase it
Starting point is 00:38:08 and build the wrong thing for two weeks before realizing like, oh, we actually miss this thing we need to go back. If I recap the last couple of minutes, basically, it's hopefully an eye-opener for those that haven't seen this already. AI in developer experience
Starting point is 00:38:26 is not limited to just generating code and being an assistance to code generation. It really impacts the whole SDLC. and stuff that you just brought up is interesting to hear like these use cases on taking a screen, taking a photo from a sketch on a piece of paper converting that into a working prototype or using AI to even validate if that design actually makes sense, if there's no flaws in there. Then he mentioned data generation. One of the talks at Cloud Native Austria was also around chaos engineering with AI, where they also talked about using AI for a good
Starting point is 00:39:06 data set generation, right, for your tests, and then also for executing some, some chaos experiments. There's so many use cases out there. I want to now a little bit play devil's advocate in the last couple of minutes that we have. Please. Yeah. So now we have this brand new world where with only half of the people we create 50% more features because everything is optimized and automated and now
Starting point is 00:39:35 things run in production. How much of that code is really generated that runs in production generated by AI versus how much is really a human engineer still involved in creating new code? And
Starting point is 00:39:51 if something happens in production who is fixing it and who can fix it? Brilliant questions. So there's a lot of marketing claims out there from vendors. I understand why they do it, but right, we have, oh, up to 90% of code is going to be generated by AI. There's lots of bold claims about it.
Starting point is 00:40:14 The truth is there's a big up in telemetry when it comes to understanding how much AI authored code gets to production without large human intervention, I would say. DX is working on a tool. We've had a tool in Alpha that's doing this. It's not using AI to determine whether the code was AI generated, which I've seen a couple tools that do that that aren't very accurate.
Starting point is 00:40:39 We're looking on the actual file system level. It's agnostic across any AI tool, across any IDE, and then we can track it on the commit level and then track that commit to production really understand line by line, what was generated by AI, what was copy-pasted, and what was actually typed in by a human being, which is, I think, we just don't have that information right now. What we do have is self-reported data right now, which, you know, we don't have an alternative. So self-reported data is what we're going to go with right now.
Starting point is 00:41:09 I was just actually looking at, because I'm putting together a research report. On this, very things I can tell you the exact number. AI authored code. So looking across a sample of over 43,000 developers at over 265 companies, the average number of, like the average amount of self-reported code that gets, that is AI generated without large human intervention that gets merged upstream and put in a customer-facing environment is 22% looking at the last trailing three months. This is very different than 30%.
Starting point is 00:41:48 It's very different than 50% and it's definitely a lot different than 90%. So that's kind of the, you know, that's what we're seeing on the ground right now. And this is across all sorts of software projects, I guess. Yeah, we're looking at companies all the way from the Fortune 50, very large enterprises and all the way to startups. I mean, any sector is probably represented, and I made sure the sample is pretty well distributed across organization size as well, because that has a lot to do with adoption of AI tools, usage of AI tools.
Starting point is 00:42:27 Interestingly enough, though, that when I broke down, so this was like 40, what did I say just now, 43, 43,000 plus developers, when I broke them down into usage categories and looked at, you know, the people who are actually, and this is system data, this isn't their, like, self-reported data, this is system data. evidence from their AI tools, a usage event. And I'm looking at, okay, those that are using it daily or more, what is their percentage compared to someone who's using it just like monthly? And there's not a huge amount of difference. It's only a couple percent. It's like 21 versus 24 percent. We seem to like be hovering around the 20 percent range, 22 percent range pretty steadily for the last couple months. I'm really curious about how this changes as companies adopt more like agentic workflows, as people get more confident as we see some investments in specifically like SRE tooling, quality testing, become a bit more robust because
Starting point is 00:43:30 people have realized, oh, what's holding us back from using AI more effectively is not the latest model or whether we're using cursor or clod, it's that our testing is not sufficient. And like, we really need to rethink how we're going to get code that was not authored by a human safely to production. Companies are making investments in making that more robust. And, like, you know, I think the next six to 12 months are going to be very telling in terms of what happens. So that tells me back to basics.
Starting point is 00:44:01 Absolutely back to basics. Yeah. Mm-hmm. Because Laura, for you to understand Brian and I, we both have a big background and how we started in quality engineering. And that's why if we are trying, if we're getting into a world where AI generates more code, obviously we need to make sure that the safety nets are there and the best safety net is a good test. Hoping that the test was not generated by a body of that AI,
Starting point is 00:44:29 meaning another AI body that kind of created flaky tests. maybe as a parting thought I think back to basics is a great headline for what makes company successful with AI when we look at the things that make agentic workflows we'll just say agentic workflows really well it's like clearly defined services really robust testing the ability to detect anomalies in the system and fix them quickly production debugging observability all of these things are core components to develop our experience they're all the basics we've known them for decades now. And the interesting moment that we're at is that Agent X, which is what I'll call this, it's the same as developer experience, but companies are more willing to open their wallet
Starting point is 00:45:14 for Agent X. That's just where we are in time and space right now is like people are willing to invest in AI. So actually, we have a great opportunity to capitalize on this and make a lot of the investments that we've been wanting and advocating for, but couldn't, you know, maybe couldn't secure budget for it. Now we can frame it as an AI. investment and have access to bigger piles of money. But it is very much back to basics. AI is not a magic silver bullet. We need to approach it just like we approach every other tool and engineering fundamentals.
Starting point is 00:45:46 The physics of great software haven't gotten away just because AI is here. So does that mean I can look for a new job and paste some chips and stuff on me and say, look, I'm AI, hire me? You could try. Or maybe the people that really know the basics well, right, that write solid test frameworks that really build a foundation that invest in speeding up their CICD
Starting point is 00:46:15 where more tests can be done, faster feedback. This is a trade that will definitely not go away but is more important than ever. It'll be interesting to see if AI gets chipped up on the N-plus-1 problem that's when that just keeps
Starting point is 00:46:34 following every innovation around is N plus one. Hey Laura, thank you so much time flies and there are so many additional things that I would have liked to ask you I wanted to talk about
Starting point is 00:46:49 Egentic versus assisted versus workflow. I wanted to talk about the EI measurement framework that you've also mentioned in previous talks but I think we will leave this to hopefully another episode if you agree to come come back to us.
Starting point is 00:47:05 Yeah, I would be happy to and I am posting very frequently on LinkedIn research snippets, whatever I'm thinking about that day observations that I have so if folks want more information they should find me on LinkedIn
Starting point is 00:47:16 I'm pretty sure I'm the only Laura Tacho out in the world and we will obviously have the link to your LinkedIn and your talks and everything in the description of this podcast. It would be very interesting to see where this
Starting point is 00:47:29 all leads in the next five years or so. I think we're at an inflection point for sure, but all remains to be seen where we need that out. But thank you again, Laura, for taking the time and be with us. We really, really appreciate it. I was more quiet today because I was just listening and learning, which sometimes is my thing. So fantastic stuff. Thank you very much for being on. We look forward to having you on the next time. Bye, everybody. My pleasure. Thank you. You know what I'm going to be.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.