Daybreak - Indian IT’s AI coding boom has arrived. The cleanup engineers haven’t

Episode Date: April 22, 2026

A study gave 16 experienced developers the best AI coding tools available.They predicted they'd be 24% faster. They felt 20% faster. They were actually 19% slower — and still didn't believe... it when told.That gap between belief and reality is now being deployed at enterprise scale.TCS, Infosys, Wipro, and Cognizant have committed to over 50,000 AI coding licences each. Bugs per developer are up 50%. Code is reaching production without any human review.And the senior engineers who could catch the mistakes are buried too deep in the flood to look up.Is India's IT sector selling a productivity story it hasn't actually earned yet?Tune in. With inputs from Mrunmayee Kulkarni. Daybreak is produced from the newsroom of The Ken, India’s first subscriber-only business news platform. Subscribe for more exclusive, deeply-reported, and analytical business stories.

Transcript
Discussion (0)
Starting point is 00:00:00 In early 2025, 16 experienced developers sat together in a room to take part in a study. The study was being conducted by Meador, a research non-profit that proclaims itself to scientifically measure whether and when AI systems might threaten catastrophic harm to society. These 16 coders were given access to their own code bases and projects that they had been working on for years. They were given real tasks to complete in the same project, like fixing bugs, building, features, restructuring the code, you know, the usual. Half of them were given access to the best AI coding tools available at the time, like Cursor Pro and Claude's Sonnet 3.5 and 3.7. The other half worked without the tools. Before they began, the developers were asked to estimate how much faster they would be with
Starting point is 00:00:51 the tools than without. They said 24%. Once they finished, they felt that they'd been about 20% faster. But the study showed that they were actually 19% slower. This gap between the feeling and blind belief in productivity gains and the reality of it didn't happen in a vacuum. You see, by the time the study was published in July 2025, City, one of the largest banks in the world, had already rolled out GitHub co-pilot to its 40,000 strong developer workforce. SAP, the enterprise software had deployed an AI coding tool company-wide and was citing 20% faster code production. Also, Google CEO had told investors that over a quarter of new code deployed at Google was AI generated. Basically, the bet had already been placed on the belief, while the actual results
Starting point is 00:01:49 were only just coming to light. Now, you might say that AI coding tools have come a long way since then. In fact, this was before Claude Code, which is one of the most popular AI coding tools right now, even came into the picture. And you would be right. Even Meeter agrees that the results of the study are subject to change over time, considering how quickly AI develops. And another larger study that was published just this month showed that things have changed, just not necessarily for the better. Farrow's AI studied 22,000 developers across 4,000 teams and found that while developers were faster, companies themselves weren't seeing meaningful productivity gains. Developers were writing more code and closing more tasks, but the
Starting point is 00:02:40 AI-generated code was just getting bigger and buggyer, shifting the bottleneck to the review stage and canceling out any of the speed gained at the development stage. The industry calls is tech debt, the trade-off between short-term speed and long-term sustainability. And it's a debt that is only going to compound. Because think about it, the more junior developers are pushed to rely on AI before they've built the skills, the fewer senior developers there will be in the future who can actually review and debug what it produces. Meanwhile, Indian IT companies like InfoS, WIPRO, TCS and Cognizant are all placing their own bets on the deployment of AI coding tools in the workforce. InfoS claimed 20 to 25%
Starting point is 00:03:25 productivity gains on its own financial product and up to 40% on specific client-facing work. TCS, WIPRO, Info, Enfossus and Cognizant all signed a partnership with Microsoft to deploy over 50,000 co-pilot licenses across each of their operations. Clearly, the productivity story is just too compelling to resist. Even as the evidence for it is only, getting thinner. So what does this shift risk losing? Welcome to Daybreak, a business podcast from the Ken. I'm your host, Richard Wurgis, and every day of the week, my co-host, Nika Sharma and I will bring you one new story that is worth understanding and worth your time. Today is Wednesday, the 22nd of April.
Starting point is 00:04:09 In April last year, Toby Lutker, the CEO of Shopify, posted an internal memo on X after it had already leaked. The message said, reflexive AI usage is now a baseline expectation at Shopify. In practice, it meant that managers at the company who were hiring would have to first demonstrate why it was a job that AI could not do. The memo had come out after Lutker had first invited his engineers to tinker with AI tools. After it was clear that the invitation had been understood as too much of a suggestion, it became a director. It was even tied to the engineer's performance reviews. Other companies soon started following a similar model.
Starting point is 00:05:12 The point was no longer organic adoption. It was to push AI and the productivity promise whether the developers wanted it or not. Amazon took it a step further. In November last year, an internal memo established Key Row, which is Amazon's own AI coding agent, as the standard for all software engineers in the company. There was even a weekly usage target of 80% established for management to track. That's just a usage tracker, by the way.
Starting point is 00:05:44 Not an evaluation of the quality or even quantity of code. It's only a measure of how often employees used AI tools. My colleague, the Kenrapportem Ron Michael Carney, spoke to a supervisor who worked at Amazon. He told her about an incident where a senior developer went it to him. The developer said, and I quote, Kiro hallucinates API so much, and it tries to convince the coder that this is what they asked for. Why should we learn a new coding technique for something that needs so much verification? This developer was not the only one frustrated with this mandate.
Starting point is 00:06:22 Earlier this year, 1,500 engineers at Amazon's US headquarters signed a petition asking management to allow the use of Claude Code instead. Now, to be fair, there's no bad. ban per se on ClaudeCode at Amazon. But when the company announced that Kiro was its recommended tool, it also said that it would not be supporting the use of third-party tools. Remember how I said that the tech debt that's coming out of this shift is just compounding? Well, for Amazon, the recommendation to use Kiro came back to bite it soon enough. Twice, in fact. The first time, just a month after the memo in December 2025. AWS suffered a third. 13-hour interruption to its cost explorer service in a region of mainland China.
Starting point is 00:07:09 Financial Times reported that it was because some engineers had allowed Keiro to make some system changes. Now, as an agentic tool, it's supposed to take autonomous action based on human instruction. This time, the tool determined that the best way to solve a specific problem was to simply delete and recreate the environment. It could do this because it had inherited the engineers high-level permissions. Amazon denied these reports and placed the blame on the fact that the user's instructions were broader than they were supposed to be. The second incident was just last month. This time, it was a six-hour outage on Amazon's main e-commerce platform. Shoppers got drawn delivery estimates, approximately 1 million-plus website errors, and the company lost 1,000
Starting point is 00:07:59 orders. While the media attributed this outage to AI-Gen. bad code, Amazon insisted that that was not the case. It said that the reason was an engineer who followed inaccurate advice that an AI tool inferred from an outdated internal wiki. The internal response to these system failures is where things get interesting. Regardless of whether the mistakes were AI generated or a factor of engineers making mistakes through or because of AI, Amazon realized that changes needed to be made in the pipeline. It mandated two-person approval for major changes to critical systems,
Starting point is 00:08:35 and that senior engineers would need to sign off on all AI-assisted code written by junior staff. And this development in the pipeline is exactly what the Farrow study observed what's happening and why it was a problem. More on this in the next segment. The Farrow study has a name for what's happening. Acceleration Whiplash. The two-year-long study showed that, AI is entering a system that was built for human ability. And the output now is basically a flood
Starting point is 00:09:12 which the existing system just does not have a way to absorb. Now, I am going to be reciting a bunch of numbers, but bear with me because they really do illustrate how and why the existing system was not ready for this flood. As I mentioned earlier, the output gains are real. The number of tasks completed by individual developers is genuinely up. But the quality of the quality of the code generated is another story. Bugs per developer are up by more than 50%. To put that in context, it was just 9% in Farrow's own report last year. Plus, for every code submitted for review, the probability of a production incident,
Starting point is 00:09:55 which is basically things like outages, security failures or user-facing system failures, has more than tripled. The median time for review is up by more than 4.4.4. 100% and more than 30% of code is reaching production or going live without any human review at all. That's because reviewers simply cannot keep up with the sheer volume of the code being generated. Now, Faro's itself describes the production incident as an outage of security event that hits users after code goes live. The study data set covered major companies operating in finance, healthcare and infrastructure, and infrastructure. On the user end, outages in those companies and those sectors mean your bill payment
Starting point is 00:10:42 failing, your transaction processing twice, or a medical record system going down entirely. In my last episode on Anthropic, I had mentioned a recent article from the New York Times. It called the current situation a code overload. Considering that the median review time is more than 400%, it only stands to reason that companies need more senior engineer to reduce that time. NYT wrote that now companies are finding it difficult to hire enough application security engineers. These are the people who can monitor AI code for risks. The problem?
Starting point is 00:11:19 That there aren't that many engineers qualified for that role. In the world. Joe Sullivan, an advisor to Kustanoa Ventures, a Silicon Valley venture firm, told NYT that the reality is that there are not enough application security engineers on the planet. to satisfy what just American companies need. So with Indian companies going all in on adopting coding tools for themselves, what are they signing up for? Stay tuned. Since the AI Impact Summit this year,
Starting point is 00:11:57 the push for AI adoption has only gotten stronger in India. My colleague, Mirren Michael Carney, covered the change in her piece in February. Here's what she covered. For one, OpenAI expanded its partnership with the Tata group. TCS or Tata Consultancy Services, which is Tata's IT services arm, is building AI infrastructure and rolling out enterprise chat GPT internally. Meanwhile, even as Anthropic opened its first India office in Bangalore, it also tied up with Infisys. The idea is to deploy cloud models and AI agents into Infoises enterprise workflows. India has also emerged as one of the fastest growing markets for Microsoft's AI programmer called GitHub Go.
Starting point is 00:12:41 In December, the company had announced that more than 200,000 copilot licenses would be deployed across TCS, Infoys, WIPRO and Cognizant. As you can see, it's all a very top-down approach. Leadership teams decide which license to deploy and where based on what makes the most business sense to them. It's all about preserving existing partnerships, cutting costs and getting the best deal. The cost of that is the real productivity of the developers. Here's what Rn Mn Mn Mew observed. The problem is, developers prefer different tools from the ones they've been assigned. In fact, some are even ready to go back to no tools given the amount of time they have to spend
Starting point is 00:13:24 verifying AI generated code instead of simply writing it. Forget about tracing output levels. These tools have made their existing job more cumbersome. It doesn't end up mattering for companies because at the end of the day, adoption is much easier to track than real productivity. But it's likely to matter soon enough. With a shortage of developers who are actually qualified to comb through these large amounts of code and with newer coders never having built the skill in the first place,
Starting point is 00:13:53 soon enough there will be far too much code and far too few experts to fix it. A data engineer at a US-based startup told Runmay that new kids are learning to code with AI from day one. So, while they're good at generating average code, they're not as good at debugging. He explained that older engineers over time build an ability to suss out when something feels wrong. But for the junior staff, they'll always have AI generate the first draft.
Starting point is 00:14:23 They're unlikely to ever build the muscle. So they won't learn how to trace errors from first principles. And here's an example of what happens when someone's not trained to spot mistakes. Last year, engineers at e-commerce company Flipkart, who had been using GitHub co-pilot ended up with a wrong line of code. A senior programmer told Runmay that the bad code would automatically
Starting point is 00:14:46 empty customers' cards after they had added just five items. Lucky for Flipcard though, the same programmer caught the glitch just minutes before deployment. Considering the results from the Faro study, these aren't even anomalies. In a lot of ways, it's baked into how
Starting point is 00:15:03 LLMs work. Because there's also the fact that LLMs also work on the statistical average of what already exists, which means that the code you get from AI is well average, which is fine for recreating patterns or making one-size-fits-all code, but when you have multiple companies in the same sector fighting to stand out, differentiation is indispensable. There's a tangible gap between something that just works and something that's outstanding.
Starting point is 00:15:33 And that differentiation cannot come around if a strong, talented and proficient engineer isn't leading the charge assisted by AI or not. But if that same strong, talented and proficient engineer who makes the difference between average and outstanding systems is buried under millions of lines of code. If they are expected to catch and fix the mistakes made by others who never even learn to spot it themselves, it's unlikely that that creative edge in Spark is going to remain for very long. Because what also doesn't get talked about enough is that even these developers, those who are highly experienced, are increasingly relying on AI tools, which means that the skill set that is increasingly becoming
Starting point is 00:16:17 more important for the deployment of AI code is also slowly eroding. Luciano Nuijan, a lead developer, wrote a blog post wanting other developers to be cautious about making AI a key part of the workflow. He recounted how he had started relying on the AI tools provided by the software, his company because he had started to feel left behind. About a year later, he removed those tools from his workflow entirely. Why? Well, because while he was working on a personal project, when he didn't have access to his
Starting point is 00:16:51 company's provided tools, he realized that he had lost the instinct for tasks that used to be second nature for him. So this is what's at stake. The structural integrity of some of the most popular platforms we all use. and also the skill sets of the architects who prop it all up. It's an entirely solvable problem because you see, it's not like these developers are anti-AI. Even the 1500 Amazon petitioners were petitioning for Claude Code as an alternative. The problem lies in the fact that companies are prioritizing adoption to present nice looking numbers to shareholders and partners
Starting point is 00:17:30 and not the needs of their own engineers. So what is needed is a balance between quantity and quality and between mediocrity delivered quickly and real skill developed over time. Daybreak is produced from the newsroom of the Ken India's first subscriber-focused business news platform. What you're listening to is just a small sample of our subscriber-only offerings. A full subscription offers daily long-form feature stories, newsletters and a whole bunch of premium podcasts. To subscribe, head to the podcast. ken.com and click on the red subscribe button on the top of the Ken website.
Starting point is 00:18:12 Today's episode was hosted and produced by my colleague Rachel Vargis and edited by Rajiv Sien.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.