Microsoft Research Podcast - AI Frontiers: AI in India and beyond with Sriram Rajamani

Episode Date: August 31, 2023

Powerful large-scale AI models like GPT-4 are showing dramatic improvements in reasoning, problem-solving, and language capabilities. This marks a phase change for artificial intelligence—and a sign...al of accelerating progress to come. In this Microsoft Research Podcast series, AI scientist and engineer Ashley Llorens hosts conversations with his collaborators and colleagues about what these models—and the models that will come next—mean for our approach to creating, understanding, and deploying AI, its applications in areas such as healthcare and education, and its potential to benefit humanity.This episode features Sriram Rajamani, Distinguished Scientist and Managing Director of Microsoft Research India. Rajamani talks about how the lab’s work is being influenced by today’s rapidly advancing AI. One example? The development of a conversational agent in India capable of providing information about governmental agricultural programs in farmers’ natural language, particularly significant in a country with more than 30 languages, including 22 government-recognized languages. It’s an application Microsoft CEO Satya Nadella described as the “mic drop moment” of his trip to the lab early this year.Learn moreAI4Bhārat | Organization homepageMEGA: Multilingual Evaluation of Generative AI | Publication, May 2023AI and Microsoft Research | Learn more about the breadth of AI research at Microsoft

Transcript
Discussion (0)
Starting point is 00:00:00 . I'm Ashley Lorenz with Microsoft Research. I've spent the last 20 years working in AI and machine learning, but I've never felt more fortunate to work in the field than at this moment. The development of increasingly powerful large-scale AI models like GPT-4 is accelerating the advancement of AI. These models and the systems they power are exhibiting surprising new abilities like reasoning, problem solving, and translation across languages and domains.
Starting point is 00:00:33 In this podcast series, I'm sharing conversations with fellow researchers about the latest developments in large AI models, the work we're doing to understand their capabilities and limitations, and ultimately, how innovations like these can have the greatest benefit for humanity. Welcome to AI Frontiers. Today, I'll speak with Shiram Rajamani, Managing Director of Microsoft Research India. For nearly 20 years, this lab has focused on interdisciplinary research, blending theory and practice, and computer science with social science. Our researchers in India have made many contributions to advanced AI in areas like causal reasoning.
Starting point is 00:01:17 But the latest wave of powerful AI models has made a profound impact on all the lab's work, including their approach to creating technologies for underserved communities. All right, so Sriram, let's dive right in. I think it's fairly obvious for me to say at this point that chat GPT and generative AI more broadly is a worldwide phenomenon. But what's so striking to me about this is the way that so many people around the world can pick up the technology and use it in their context in their own way. I was on a panel discussion a few weeks ago where I saw a comedian discover in real time that GPT-4 could write jokes that are actually funny. And shortly after that, I spoke to a student who was using ChatGPT to write an application to obtain a grazing permit for cattle.
Starting point is 00:02:09 You know, the work of your lab is situated in its own unique societal context. So what I really want to know and start with here today is like, what's the buzz been like for you and your part of the world around this new wave of AI? Yeah, first of all, Ashley, you know, thank you for having this conversation with me. You're absolutely right that our lab is situated in a very unique context on how this technology is going to play out in this part of the world, certainly. And you might remember, Ashley, a sort of a mic drop moment that happened for Satya when he visited India earlier this year in January. So one of our researchers, Pratish Kumar, he is also co-founder of our partner organization
Starting point is 00:02:54 called AI for Bharat. He works also with the government on a project called Barshini, which the government endeavors to bring conversational AI to the many Indian languages that are spoken in India. And what Pratyush did was he connected some of the AI for Bharat translation models, language translation models, together with one of the GPT models, to build a bot for a farmer to engage and ask questions about the government's agricultural programs. So the farmer could speak in their own language, you know, it could be Hindi. And what the AI for Bharat models would do
Starting point is 00:03:41 is to convert the Hindi speech into text and then translate it into English. And then he taught, you know, either fine-tuned or did retributable augmented generation. I don't quite remember which one, one of those, where he made a GPT model customized to understand the agricultural program of the government. And he chained it together with this speech recognition and translation model. And the farmer could just now talk to the system, the AI system in Hindi, and ask, you know, are they eligible for their benefits? And many details. And the model had a sensible conversation with him. And Satya was just really amazed by that. And he calls that,
Starting point is 00:04:23 he called that as the mic drop moment of his trip in India, which I think is indicative of the speed at which this disruption is impacting very positively in various parts of the world, including the Indian subcontinent. You referenced the many Indian languages written and spoken. Can you just bring that to life for us? How many languages are we talking about? So I think there are at least, you know, 30 or 40, you know, mainstream languages. I mean, the government recognizes 22.
Starting point is 00:04:58 We call them as IN22. But I would think that there are about 30plus languages that are spoken very, very broadly, each of them with several tens of millions, hundreds of millions of speakers. And then there's a long tail of maybe 100 more languages, which are spoken by people in smaller population counts. There are also very low resource languages like Gondi and Yiddu Mishmi, which are just spoken by maybe just only a million speakers or even under a million speakers. Those languages probably don't have enough data resources. So India is an amazing testbed because of this huge diversity and
Starting point is 00:05:47 distribution of languages in terms of the number of speakers, the amount of available data, and many of these tail languages have unique, you know, socio-cultural nuances. So I think in that sense, there's a really good testbed for, you know for how conversational AI can inclusively impact the entire world. And you mentioned TAIL languages. And so maybe we mean they're low-resource languages like you also mentioned. What's the gap like between what languages AI is accessible in today versus the full extent of all those languages that you just described, even just for the Indian subcontinent?
Starting point is 00:06:30 So what we are seeing is that with IN22, the top languages, if you look at successive versions of the GPT models, for example, the performance is definitely improving. So if you just go from GPT-2 to GPT-3 to 3.5 to 4, you can sort of see that these models are increasingly getting capable. But still there is a gap between what these models are able to do and what custom models are able to do, particularly if you go towards languages in which there is not enough training data.
Starting point is 00:07:09 So people in our lab are doing very systematic work in this area. There is a benchmarking work that my colleagues are doing called MEGA, where there is systematic benchmark being done on various tasks, on a matrix that consists of tasks on one axis and languages on another axis, to just systematically and empirically study what these models are able to do. And also, we are able to build models to predict how much more data is needed in each of these languages in order for the performance to be comparable to, say, languages like English, right? What is the gap and how much data is needed? The other thing is that it turns out that these models, they learn also from related languages. So if you want to improve the performance of a language,
Starting point is 00:08:07 it turns out there are other languages in the world and in India that have similar characteristic, you know, syntactic and semantic characteristics to the language that we are thinking about. So we can also sort of recommend, you know, what distribution of data we should collect so that all the languages improve. So that's the kind of work that we're doing. That's one of the most fascinating parts of all of this, how diversity in the training data set improves across the board, like even the addition of code, for example, in addition to language.
Starting point is 00:08:37 And now we're even seeing even other modalities. And the wave of AI and the unprecedented capabilities we're seeing has significant implications for just about all of computing research. In fact, those of us in and around the field are undergoing now a process that I call, you know, reimagining computing research. And, you know, that's a somewhat artful way to put it. But beyond the technical journey, there's an emotional journey happening across the research community and many other communities as well. So what has that journey been like for you and the folks at the India Lab? You know, our work in the lab spans four areas. You know, we do work in theory and algorithms. We do work in AI and machine learning. We do systems work. And we also have an area called technology and empowerment. It's about making sure that technology benefits people.
Starting point is 00:09:45 And so far, our conversation has been about the last area. But all these four areas have been affected in a big way using this disruption. Maybe I'll just say a few more things about the empowerment area first and then move on to the other ones. If you look at our work in the empowerment area, Ashley, right? This lab has had a track record of doing work that makes technology inclusive, not just from an academic perspective, but by also deploying the work. We have spun off startups, many startups that have taken projects in the lab and scale them to the community. Examples are Digital Green, which is an agricultural extension, 99 Dots, which is a tuberculosis medication adherence system. Karya is a platform for dignified digital labor to enable underprivileged users, rural users, to contribute data and get paid for it.
Starting point is 00:10:50 HAMS is a system that we have built to improve road safety. We have built a system called BlendNet that enables rural connectivity. And almost all of these, we have spun them off into startups that have been funded by venture capitalists, impact investors. And we have a vibrant community of these partners that are taking the work from the lab and deploying them in the community. So the second thing that is actually happening in this area is that, as you may have heard, India is playing a pivotal role in digital public infrastructure. Advances like the Aadhaar biometric authentication system, UPI, which is a payment system, they are pervasively deployed in India, and they reach several hundreds of millions of people, and in the case of Aadhaar, more than a billion people, and so on.
Starting point is 00:11:44 And the world is taking note. India is now head of the G20 and many countries now want to be inspired by India and build such digital public infrastructure in their own countries.
Starting point is 00:12:00 And so what you saw is the mic drop moment, right? It actually has been coming for a long time. There has been a lot of groundwork that has been laid by our lab, by our partners, you know, such as AFR Bharat, the people that work on digital public goods, to get the technical infrastructure and our know-how to a stage where we can really build technology that benefits people, right? So going forward, in addition to these two major advancements, which is the building of the partner and alumni ecosystem, the digital public good infrastructure,
Starting point is 00:12:39 I think AI is going to be a third and extremely important pillar that is going to enable citizen-scale digital services to reach people who may only have spoken literacy and who might speak in their own native languages. And public services can be accessible to them. So you mentioned AI for Bharat, and I'd love for you to say a bit more about that organization and how researchers are coming together with collaborators across sectors to make some of these technology ideas real. Yeah, so AI for Bharat is a center in IIT Madras,
Starting point is 00:13:23 which is an academic institution. It has multiple stakeholders, not just Microsoft Research, but our Search Technology Center in India also collaborates with them. Nandan Nilakani is a prominent technologist and philanthropist. He is behind a lot of India's digital public infrastructure. He also funds that center significantly through his philanthropic efforts. And there are a lot of academics that have come together. And what the center does is data collection.
Starting point is 00:14:00 I talked about the diversity of Indian languages. They collect various kinds of data. They also look at various applications, like in the judicial system. In the Indian judicial system, they are thinking about how to transcribe judgments, enabling various kinds of technological applications in that context, and really actually thinking about how these kinds of AI advances can help ride on top of digital public goods. So that's actually the context in which they are working on. Digital public goods. Can you describe that? What do we mean in this context by digital
Starting point is 00:14:40 public good? So what we mean is, if you look at Indian digital public infrastructure, right, there is, as I mentioned, there is Aadhaar, which is the identity system that is now enrolled in more than 1.3 billion Indians. There is actually a payment infrastructure called UPI. There are new things that are coming up, like something that's called Beckham, there's something called ONDC, that is poised to revolutionize how e-commerce is done. So these are all, you know, sort of protocols that through private-public partnership, right, government together with think tanks have developed, that are now deployed in a big way in India. And they are now pervasively
Starting point is 00:15:26 impacting education, health, and agriculture. And every area of public life is now being impacted by these digital public infrastructures. And there is a huge potential for AI and AI-enabled systems to ride on top of this digital public infrastructure to really reach people. You talked about some of the infrastructure considerations. And so what are the challenges in bringing digital technologies to the Indian context? And you mentioned the G20 and other countries that are following the patterns. What are some of the common challenges there? So, I mean, there are many, many challenges.
Starting point is 00:16:12 One of them is lack of access. You know, though India has made huge strides in lifting people out of poverty, people out there don't have the same access to technology that you and I have. Another challenge is awareness. People just don't know, you know, how technology can help them, right? You know, people here in this podcast know about, you know, LinkedIn to get jobs, they know about, you know, Netflix or other streaming services to get entertainment. But there are many people out there that don't even know that these things exist.
Starting point is 00:16:46 So awareness is another issue. Affordability is another issue. So many of the projects that I mentioned, what they do is actually they start not with the technology. They start with the users
Starting point is 00:17:02 and their context and their situation and what they're trying to do and then map back. And technology is just only one of the pieces that these systems, all of these systems that I mentioned, right? Technology is just only one component. There's a socio-technical piece that deals with exactly these kinds of access and awareness and these kinds of issues. And we're kind of taking a walk right now through the work of the lab and there are some other areas that you want to get into. But I want to come back to this. Maybe this is a good segue into the emotional journey part of the question I asked a few minutes
Starting point is 00:17:36 ago. As you get into some of the deep technical work of the lab, what were some of the first impressions of the new technologies and what were some of the first things that you and your colleagues there and our colleagues felt in observing these new capabilities? So I think Peter mentioned this very eloquently as stages of grief. And me and my colleagues, I think, went through the same thing. I mean, we went from disbelief saying, oh, wow, this is just amazing. I can't believe this is happening to sort of understanding what this technology can do and over time understanding what its limitations are and what the opportunities are as a scientist and technologist and engineering organization to really push this forward and make use of it. So that's, I think, the stages that we went through.
Starting point is 00:18:31 Maybe I can be a little bit more specific. As I mentioned, the three other areas we work on are theory and algorithms in machine learning and in systems. And I can sort of say how my colleagues are evolving their own technical and research agendas in the right of this disruption. If you take our work in theory, this lab has had a track record of, you know, cracking long-standing open problems. For example, problems like the Caddison-Singer
Starting point is 00:18:57 conjecture that was open for many years, many decades, was actually solved by people from the lab. Our lab has incredible experts in arithmetic and circuit complexity. They came so close to resolving the VP versus VNP conjecture, which is the arithmetic analog of the P versus NP problem. So we have incredible people that working on theoretical computer science, and a lot of them are now shifting their attention to understanding these large language models, right? Instead of understanding just arithmetic circuits, you know, people like, you know, Neeraj Kayal and Ankit Garg are now thinking about mathematically what does it take to understand transformers?
Starting point is 00:19:35 How do we understand, how might we evolve these models or training data so that these models improve even further in performance, in their capabilities, and so on. So that's actually a journey that the theory people are going through, bringing their brainpower to bear on understanding these models foundationally. Because as you know, currently our understanding of these foundation models is largely empirical. We don't have a deep scientific understanding of them. So that's the opportunity that the theoreticians see in this space.
Starting point is 00:20:12 If you look at our machine learning work, you know, that actually is going through a huge disruption. I remember now, one of the things that we do in this lab is work on causal ML. Amit Sharma, together with Emery Ketchum and other colleagues working on causal machine learning. And I heard a very wonderful podcast that you hosted them some time ago. Maybe you can say a little bit about what you heard from them, and then I can pick up back and then connect that with the rest of the lab. Sure. Well, it's, you know, I think
Starting point is 00:20:46 the common knowledge, there's so many things about machine learning over the last few decades that have become kind of common knowledge and conventional wisdom. And one of those things is that, you know, correlation is not causation and that, you know, learned models don't, you know, generally don't do causal reasoning. And so we, you know, we've had very specialized tools created to do the kind of causal reasoning that Amit and Emery do. And it was interesting. I asked them some of the same questions I'm asking you now, you know, about the journey and the initial skepticism. But it was been really interesting to see how they're moving forward. They recently published a position paper on archive where they conducted some pretty compelling experiments, in some cases showing something like, you know,
Starting point is 00:21:42 causal reasoning, you know, being exhibited, or at least, I'll say, convincing performance on causal reasoning tasks. Yeah, absolutely. Yeah, go ahead. Yeah, absolutely. So, I would say that their journey was that initially, they realized that, I mean, of course, they built specialized causal reasoning tools like Do-Why, which they've been building for many years. And one of the things they realized was that, oh, some of the things that Do-Why can do with sophisticated causal reasoning, these large language models were just able to do out of the box. And that was sort of stunning for them, right?
Starting point is 00:22:21 And so the question then becomes, you know, does specific vertical research in causal reasoning is even needed, right? And so the question then becomes, you know, does specific vertical research and causal reasoning is even needed, right? So that's actually the shock and the awe and the emotional journey that these people went through. But actually, after the initial, you know, shock faded, they realized that there is actually better together story that is emerging in the sense that they you know once you understand the details what they realized was that natural language contains a lot of causal information right if you just look at the literature the literature has many things like you know a causes b if there is if there is you know hot weather then ice cream sales go up. You know, this information is present in the literature. So if you look at tools like DoY, what they do is that in order to provide causal machine learning,
Starting point is 00:23:14 they need assumptions from the user on what the causal model is. They need assumptions about what the causal graph is, what is the user's assumptions about which variables depend on which variables, right? And then, and what they realize is that models like GPT-4 can now provide this information. Previously, only humans were able to provide this information. But in addition to that, right, tools like DUI are still needed to confirm or refute these assumptions statistically using data.
Starting point is 00:23:41 So this division of labor between getting assumptions from either a human or from a large language model, and then using the mathematics of do-why to confirm or refute the assumptions now is emerging as a real advance in the way we do causal reasoning, right? So I think that's actually what I heard in your podcast. And that's indicative of actually what the rest of my colleagues are going through, you know, moving from first thinking about, oh, the GPT-4 is like a threat, you know, in the sense that it really obviates my research area to understand, oh, no, it's really a friend. It really helps me do, you know, some of the things that required primarily humor intervention. And if I combine GPT or these large language models together with domain-specific research,
Starting point is 00:24:27 we can actually go after bigger problems that we didn't even dare going after before. Let me ask you, I'm going to pivot here in a moment, but have you covered the areas of research in the lab that you wanted to walk through? Yeah, there's more. Thank you for reminding me. Even in the machine learning area, right, there is another work direction that we have
Starting point is 00:24:50 called extreme classification, which is about building very, very classifiers with large number of labels, you know, hundreds of millions and billions of labels. And, you know, these people are also benefiting from large language encoders. You know, they have come up with clever ways of taking these language encoders that are built using self-supervised learning, together with supervised signals from things like clicks and logs from
Starting point is 00:25:14 search engines and so on, to improve performance of classifiers. Another work that we've been doing is called DISC-ANN or Approximate Nearest Neighbor Search. You know, as you know, actually in this era of deep learning, retrieval works by converting everything in the world, you know, be it a document, be it an image, you know, be it audio or video file, everything into an embedding. And relevant retrieval is done by nearest neighbor search in a geometric space. And our lab has been doing, I mean, we have probably the most scalable vector index
Starting point is 00:25:54 that has been built. And these people are positively impacted by these large-language models because, as you know, retrieval augmented generation is one of the most common design patterns in making these large language models work for applications.
Starting point is 00:26:14 And so their work is becoming increasingly relevant and they are being placed huge demands on pushing the scale and the functionality of the nearest neighbor retrieval API to do things like, oh, can I actually add predicates? Can I add streaming queries and so on? So they are just getting stretched with more demand for their work. If you look at our systems work, which is the last area that I want to cover,
Starting point is 00:26:40 we have been doing work on using GPUs and managing GPU resources for training as well as inference. And this area is also going through a lot of disruption. You know, prior to these large language models, these people were looking at relatively smaller models. You know, maybe not, you know, hundreds of billions to trillions of parameters, but maybe hundreds of millions and so on. And they invented several techniques to share a GPU cluster among training jobs. The disruption that they had was, oh, these models are so large that nobody is actually sharing clusters for them. But it turned out that some of the techniques that they invented to deal with migration of jobs and so on are now used for failure recovery in very, very large models. So it turns out that, you know, at the beginning, it seems like, oh, my work is not relevant anymore.
Starting point is 00:27:38 But once you get into the details, you find that there are actually still many important problems. And the insights you have from solving problems for smaller models can now carry over to the larger ones. And one other area I would say is the area of programming. I myself work in this area. We have been combining machine learning together with program analysis to build a new generation of programming tools. And the disruption that I personally faced was that the custom models that I was building
Starting point is 00:28:07 were no longer relevant. They aren't even needed. So that was a disruption. But actually what me and my colleagues went through was that, okay, that is true. But we can now go after problems that we didn't dare to go before. Like, for example, we can now see that Copilot and so on let you give recommendations in the context of the
Starting point is 00:28:30 particular file that you're editing. But can we now edit an entire repository which might contain millions of files with hundreds of millions of code? Can I just say, let's take, for example, the whole of the Xbox code base or the Windows code base and in the whole code, can I just say, let's take, for example, the whole of the Xbox code base or the Windows code base, and in the whole code base, I want to do this refactoring,
Starting point is 00:28:49 or I want to, you know, migrate this package from, migrate this code base from using, you know, this serialization package to that serialization package. Can we just do that, right? I think we wouldn't even dare going after such a problem two years ago. But now with large language models, we are thinking, can we do that? And large language models cannot do this right now because whatever context size you have, you can't have a hundred million line code as a context to a large language model.
Starting point is 00:29:13 And so this requires combining program analysis with these techniques. That's as an example. And actually, furthermore, there are many things that we are doing that are not quite affected by large language models. For example, actually, youmore, there are many things that we are doing that are not quite affected by large-language models. For example, actually, you know about the highway project where we are thinking about technology to make hybrid work better. And we are doing work on using GPUs and accelerators for database systems and so on.
Starting point is 00:29:41 And we do networking work. We do low-Earth-orbit satellite work for connectivity and so on. And we do networking work, we do low-Earth-orbit satellite work for connectivity and so on. And those we are doubling down though they have nothing to do with large-language models because those are problems that are important. So I think to summarize, I would say that most of us
Starting point is 00:29:57 have gone through a journey from shock and awe to some amount of insecurity saying, is my work even relevant, to sort of understanding, oh, these things are really aids for us. These are not threats for us. These are really aids. And we can use them to solve problems that we didn't even dream of before. That's the journey I think my colleagues have gone through. I want to step into two of the concepts that you just laid out. Maybe just to get into some of the intuitions as to what
Starting point is 00:30:27 problem is being solved and how generative AI is sort of changing the way that those problems are solved. So the first one is extreme classification. I think a flagship use of generative AI and foundation models is Bing chat. And so I think this idea of internet search as a, you know, as a home for these new technologies is in the popular imagination now. And I know that extreme classification seeks to solve some challenges related to search and information retrieval. But what is the challenge problem there? How is extreme classification addressing that? And how is that being done differently now? So as I mentioned, where my colleagues have already made a lot of progress is in combining language encoders with extreme classifiers to do retrieval. So there are these models called NLR. Like, for example, there's a tooling NLR model, which is a large language model, which does
Starting point is 00:31:36 representation, right? It actually represents, you know, keywords, keyword phrases, documents, and so on into encodings, you know, based on, you know, self, keyword phrases, documents, and so on, into encodings, you know, based on, you know, self-supervised learning. But it is a very important problem to combine the knowledge that these large language models have, you know, from understanding text. We have to combine that with supervised signals that we have from click logs, because, you know, we have search engine click logs. We know, you know, for example, when somebody searches for this information
Starting point is 00:32:07 and we show these results, what users click on, that's supervised signals. And we have that in huge amounts. And what our researchers have done is they have figured out how to combine these encoders together with the supervised signals from click logs in order to improve both the quality
Starting point is 00:32:24 and cost of retrieval, right? And actually, as you said, retrieval is an extremely important part of experiences like Bing chat, and retrieval augmented generation is what prevents hallucination and grounds these large language models with appropriate information retrieved and presented so that the relevant results are grounded without hallucination, right? Now, the new challenge that this team is now facing is, okay, that's so far so good as far as retrieval is concerned, right? But can we do similar things with generation, right? Can we now combine these NLG models, which are these generative models, together with supervised signals, so that even generation can actually be guided in this manner, improved in both performance as well as accuracy. And that is an example of a challenging problem that the team is going after.
Starting point is 00:33:17 Now let's do the same thing with programming. And maybe I'm going to engage you on a slightly higher level of abstraction than the deep work you're doing. And then we can get back down into the work. But one of the popular ideas about these new foundation models is that you can effectively, through interacting with them, you're sort of programming them in natural language. How does that concept sit with you as someone who, you know, is an expert in programming languages? What do you think when someone says, you know,
Starting point is 00:33:56 you're sort of programming the, you know, the system in natural language? Yeah, so I find it fascinating. And, you know, for one, you know, can be an important, the programming language community has been able to solve it only in narrow domains. For example, Excel has Flash Fill where through examples people can program Excel macros and so on. But those are not as general as these kinds of
Starting point is 00:34:37 LLM based models. And it is for the whole community not just me, right? It was stunning when users can just describe in natural language what program they want to write and these models emit Python or Java or C-sharp code. But there is a gap
Starting point is 00:34:57 between that capability and having programmers just program in natural language, right? Like, you know, the obvious one is, I can sort of Like, you know, the obvious one is, I can sort of say, you know, write me Python code to do this or that, and it can generate Python code, and I can run it. And if that works, then that's a happy path. But if it doesn't work, what am I supposed to do if I don't know Python?
Starting point is 00:35:16 What am I supposed to do, right? I still have to now break that abstraction boundary of natural language and go down into Python and debug Python. So one of the opportunities that I see is that can we build representations that are also in natural language, but that sort of describe, you know, what the application the user is trying to build and enable non-programmers, could be lawyers, could be accountants, could be doctors, to engage with the system purely in natural language. And the system purely in natural language,
Starting point is 00:35:45 and the system should talk back to you saying, oh, so far, this is what I have understood. This is the kind of program that I am writing without the user having to break that natural language abstraction boundary and having to go and understand Python, right? I think this is a huge opportunity in programming languages to see whether,
Starting point is 00:36:02 can we build, like, for example, right, actually, right, I'm a programmer, and one of the things i love about programming is that i can write code i can run it see what it produces and if i don't like the results i can go change the code and rerun it and that's sort of the you know coding evaluating uh we call it the ripple loop right so that's that's what a programmer faces right can we now provide that to natural language programmers in the sense that i want to say here's's the program I want to write. And now I want to say, oh, I want to run this program with this input. And if it doesn't work, I want to say, oh, this is something I don't like. I want to change this code this way, right? So can I now provide that kind of experience to
Starting point is 00:36:34 natural language programming? I think that's a huge opportunity if you manage to pull that off. Mm-hmm. And now let's maybe return to some of the more societally oriented topics that you were talking about at the top of the episode. In the context of programming, because being able to program in natural language, I think, really changes who can use the technologies, who can develop technologies, what a program, what a software development team can actually be and who that kind of a team can consist of. So can you paint a picture, you know, what kind of opportunities for, you know, software development does this open up when you can sort of program in natural languages, assuming we can make the AI compatible with your language, whatever that happens to be? Yeah, I think there are a lot of opportunities. Maybe I'll describe a few things that we're already doing. My colleagues are working on a project called Vellum, which is now a co-pilot assistant for societal scale applications. And one application they are going after is education.
Starting point is 00:37:49 So, you know, India, like many other countries, has made a lot of educational resources available to teachers in government schools and so on. So that if a teacher wants to make a lesson plan, you know, there is enough information available for them to search, find out many videos that their colleagues have created from different parts of the country and put them together to create a lesson plan for their class, right? But that is a very laborious process. I mean, you have an information overload
Starting point is 00:38:20 when you deal with it. So my colleagues are thinking about, can we now think about, in some sense, the teacher as a programmer and have the teacher talk to the Vellum system, saying, hey, and here is my lesson plan. Here is what I'm trying to put together in terms of what I want to teach. And I now want the AI system to collect the relevant resources that are relevant to my lesson plan and get them in my language, in the language that my students speak. How do I do that?
Starting point is 00:38:54 And all of the things that I mentioned, you have to now index all of the existing information using vector indices. You have to now retrieve augmented generation to get the correct thing. You have to now deal with the trunk and tail languages because this teacher might be speaking in a language that is not English, right? And the teacher might get a response that they don't like, right?
Starting point is 00:39:17 But they're not a programmer. How are they going to deal with it, right? So that's actually an example. If we pull this off, right, and a teacher in rural India is able to access this information in their own language and create a lesson plan which contains the best resources
Starting point is 00:39:30 throughout the country, right, we would have really achieved something. Yeah, you know, it's a hugely compelling vision. And I'm really looking forward to seeing where you and our colleagues in Microsoft Research India Lab
Starting point is 00:39:42 and MSR more broadly, you know, take all these different directions. Lab and MSR more broadly, you know, take all these different directions. So I really appreciate you spending this time with me today. Thank you, Ashley. And I was very happy that I could share the work that my colleagues are doing here and bringing this to your audience. Thank you so much.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.