Microsoft Research Podcast - AI Frontiers: AI in India and beyond with Sriram Rajamani
Episode Date: August 31, 2023Powerful large-scale AI models like GPT-4 are showing dramatic improvements in reasoning, problem-solving, and language capabilities. This marks a phase change for artificial intelligence—and a sign...al of accelerating progress to come. In this Microsoft Research Podcast series, AI scientist and engineer Ashley Llorens hosts conversations with his collaborators and colleagues about what these models—and the models that will come next—mean for our approach to creating, understanding, and deploying AI, its applications in areas such as healthcare and education, and its potential to benefit humanity.This episode features Sriram Rajamani, Distinguished Scientist and Managing Director of Microsoft Research India. Rajamani talks about how the lab’s work is being influenced by today’s rapidly advancing AI. One example? The development of a conversational agent in India capable of providing information about governmental agricultural programs in farmers’ natural language, particularly significant in a country with more than 30 languages, including 22 government-recognized languages. It’s an application Microsoft CEO Satya Nadella described as the “mic drop moment” of his trip to the lab early this year.Learn moreAI4Bhārat | Organization homepageMEGA: Multilingual Evaluation of Generative AI | Publication, May 2023AI and Microsoft Research | Learn more about the breadth of AI research at Microsoft
Transcript
Discussion (0)
.
I'm Ashley Lorenz with Microsoft Research.
I've spent the last 20 years working in AI and machine learning,
but I've never felt more fortunate to work in the field than at this moment.
The development of increasingly powerful large-scale AI models like
GPT-4 is accelerating the advancement of AI.
These models and the systems they power are exhibiting surprising new abilities
like reasoning, problem solving, and translation across languages and domains.
In this podcast series, I'm sharing conversations with fellow researchers
about the latest developments in large AI models,
the work we're doing to understand their capabilities and limitations,
and ultimately, how innovations like these can have the greatest benefit for humanity.
Welcome to AI Frontiers. Today, I'll speak with Shiram Rajamani, Managing Director of Microsoft
Research India. For nearly 20 years, this lab has focused on interdisciplinary research,
blending theory and practice, and computer science with social science.
Our researchers in India have made many contributions to advanced AI in areas like causal reasoning.
But the latest wave of powerful AI models has made a profound impact on all the lab's work, including their approach to creating technologies for underserved communities.
All right, so Sriram, let's dive right in. I think it's fairly obvious
for me to say at this point that chat GPT and generative AI more broadly is a worldwide
phenomenon. But what's so striking to me about this is the way that so many people around the
world can pick up the technology and use it in their context in their own way.
I was on a panel discussion a few weeks ago where I saw a comedian discover in real time
that GPT-4 could write jokes that are actually funny. And shortly after that, I spoke to a
student who was using ChatGPT to write an application to obtain a grazing permit for cattle.
You know, the work of your lab is situated in its own unique societal context.
So what I really want to know and start with here today is like,
what's the buzz been like for you and your part of the world around this new wave of AI?
Yeah, first of all, Ashley, you know, thank you for having this conversation with me.
You're absolutely right that our lab is situated in a very unique context on how this technology is going to play out in this part of the world, certainly.
And you might remember, Ashley, a sort of a mic drop moment that happened for Satya
when he visited India earlier this year in January.
So one of our researchers, Pratish Kumar, he is also co-founder of our partner organization
called AI for Bharat. He works also with the government on a project called Barshini,
which the government endeavors to bring conversational AI to the many Indian
languages that are spoken in India. And what Pratyush did was he connected some of the AI
for Bharat translation models, language translation models, together with one of the GPT models, to build a bot for a farmer to engage
and ask questions about the government's agricultural programs.
So the farmer could speak in their own language,
you know, it could be Hindi.
And what the AI for Bharat models would do
is to convert the Hindi speech into text
and then translate it into
English. And then he taught, you know, either fine-tuned or did retributable augmented generation.
I don't quite remember which one, one of those, where he made a GPT model customized to understand
the agricultural program of the government. And he chained it together with this speech recognition and
translation model. And the farmer could just now talk to the system, the AI system in Hindi,
and ask, you know, are they eligible for their benefits? And many details. And the model had
a sensible conversation with him. And Satya was just really amazed by that. And he calls that,
he called that as the mic drop moment of his trip in India, which I think is indicative of the speed at which this
disruption is impacting very positively in various parts of the world, including the Indian
subcontinent. You referenced the many Indian languages written and spoken. Can you just
bring that to life for us?
How many languages are we talking about?
So I think there are at least, you know,
30 or 40, you know, mainstream languages.
I mean, the government recognizes 22.
We call them as IN22.
But I would think that there are about 30plus languages that are spoken very, very broadly,
each of them with several tens of millions, hundreds of millions of speakers.
And then there's a long tail of maybe 100 more languages,
which are spoken by people in smaller population counts.
There are also very low resource languages like Gondi and Yiddu Mishmi,
which are just spoken by maybe just only a million speakers or even under a million speakers.
Those languages probably don't have enough data resources. So India is an amazing testbed because of this huge diversity and
distribution of languages in terms of the number of speakers, the amount of available data,
and many of these tail languages have unique, you know, socio-cultural nuances. So I think in that
sense, there's a really good testbed for, you know for how conversational AI can inclusively impact the entire world.
And you mentioned TAIL languages.
And so maybe we mean they're low-resource languages like you also mentioned.
What's the gap like between what languages AI is accessible in today
versus the full extent of all those languages that you just described,
even just for the Indian subcontinent?
So what we are seeing is that with IN22, the top languages,
if you look at successive versions of the GPT models, for example,
the performance is definitely improving.
So if you just go from GPT-2 to GPT-3 to 3.5 to 4,
you can sort of see that these models are increasingly getting capable.
But still there is a gap between what these models are able to do
and what custom models are able to do,
particularly if you go towards languages in which there is not enough training data.
So people in our lab are doing very systematic work in this area.
There is a benchmarking work that my colleagues are doing called MEGA,
where there is systematic benchmark being done on various tasks, on a matrix that consists of tasks on one axis and languages on another
axis, to just systematically and empirically study what these models are able to do.
And also, we are able to build models to predict how much more data is needed in each of these languages in order for the performance to
be comparable to, say, languages like English, right? What is the gap and how much data is needed?
The other thing is that it turns out that these models, they learn also from related languages.
So if you want to improve the performance of a language,
it turns out there are other languages in the world and in India
that have similar characteristic, you know,
syntactic and semantic characteristics to the language that we are thinking about.
So we can also sort of recommend, you know,
what distribution of data we should collect so that all the languages improve.
So that's the kind of work that we're doing.
That's one of the most fascinating parts of all of this, how diversity in the training data set
improves across the board, like even the addition of code, for example, in addition to language.
And now we're even seeing even other modalities. And the wave of AI and the unprecedented capabilities we're seeing has significant implications for just about all of computing research.
In fact, those of us in and around the field are undergoing now a process that I call, you know, reimagining computing research.
And, you know, that's a somewhat artful way to put it. But beyond the technical journey, there's an emotional journey happening across the research community and many other communities as well. So what has that journey been like for you and the folks at the India Lab? You know, our work in the lab spans four areas.
You know, we do work in theory and algorithms.
We do work in AI and machine learning.
We do systems work.
And we also have an area called technology and empowerment.
It's about making sure that technology benefits people.
And so far, our conversation has been about the last area.
But all these four areas have been affected in a big way using this disruption.
Maybe I'll just say a few more things about the empowerment area first and then move on to the other ones.
If you look at our work in the empowerment area, Ashley, right? This lab has had a track record of doing work that makes technology
inclusive, not just from an academic perspective, but by also deploying the work. We have spun off
startups, many startups that have taken projects in the lab and scale them to the community. Examples are Digital Green, which
is an agricultural extension, 99 Dots, which is a tuberculosis medication adherence system.
Karya is a platform for dignified digital labor to enable underprivileged users, rural users, to contribute data and get paid for it.
HAMS is a system that we have built to improve road safety.
We have built a system called BlendNet that enables rural connectivity.
And almost all of these, we have spun them off into startups that have been funded by venture capitalists, impact investors.
And we have a vibrant community of these partners that are taking the work from the lab and deploying them in the community.
So the second thing that is actually happening in this area is that, as you may have heard, India is playing a pivotal role in digital public infrastructure.
Advances like the Aadhaar biometric authentication system, UPI, which is a payment system, they
are pervasively deployed in India, and they reach several hundreds of millions of people,
and in the case of Aadhaar, more than a billion people, and so on.
And the world
is taking note.
India is now head of the G20
and many countries now
want to be
inspired by India and
build such digital public infrastructure
in their own countries.
And so
what you saw is the mic drop
moment, right?
It actually has been coming for a long time.
There has been a lot of groundwork that has been laid by our lab, by our partners, you know, such as AFR Bharat, the people that work on digital public goods, to get the technical infrastructure and our know-how to a stage where we can really build technology that benefits people, right?
So going forward, in addition to these two major advancements,
which is the building of the partner and alumni ecosystem,
the digital public good infrastructure,
I think AI is going to be a third and extremely important pillar
that is going to enable citizen-scale digital services to reach people who may only have spoken literacy and who might speak in their own native languages.
And public services can be accessible to them. So you mentioned AI for Bharat,
and I'd love for you to say a bit more about that organization
and how researchers are coming together
with collaborators across sectors
to make some of these technology ideas real.
Yeah, so AI for Bharat is a center in IIT Madras,
which is an academic institution.
It has multiple stakeholders, not just Microsoft Research, but our Search Technology Center in India also collaborates with them.
Nandan Nilakani is a prominent technologist and philanthropist.
He is behind a lot of India's digital public infrastructure.
He also funds that center significantly
through his philanthropic efforts.
And there are a lot of academics that have come together.
And what the center does is data collection.
I talked about the diversity of Indian languages.
They collect various kinds of data.
They also look at various applications, like in the judicial system.
In the Indian judicial system, they are thinking about how to transcribe judgments,
enabling various kinds of technological applications in that context, and really actually thinking about how these kinds of AI advances
can help ride on top of digital public goods. So that's actually the context in which they
are working on.
Digital public goods. Can you describe that? What do we mean in this context by digital
public good?
So what we mean is, if you look at Indian digital public infrastructure,
right, there is, as I mentioned, there is Aadhaar, which is the identity system that is now enrolled
in more than 1.3 billion Indians. There is actually a payment infrastructure called UPI.
There are new things that are coming up, like something that's called Beckham,
there's something called ONDC, that is poised to revolutionize how e-commerce is done. So these are
all, you know, sort of protocols that through private-public partnership, right, government
together with think tanks have developed, that are now deployed in a big way in India. And they are now pervasively
impacting education, health, and agriculture. And every area of public life is now being impacted
by these digital public infrastructures. And there is a huge potential for AI and AI-enabled systems
to ride on top of this digital public infrastructure to really reach people.
You talked about some of the infrastructure considerations.
And so what are the challenges in bringing digital technologies to the Indian context?
And you mentioned the G20 and other countries that are following the patterns.
What are some of the common challenges there?
So, I mean, there are many, many challenges.
One of them is lack of access.
You know, though India has made huge strides
in lifting people out of poverty,
people out there don't have the same access
to technology that you and I have. Another challenge is awareness. People just don't know, you know, how technology can help
them, right? You know, people here in this podcast know about, you know, LinkedIn to get jobs,
they know about, you know, Netflix or other streaming services to get entertainment.
But there are many people out there that don't even know that these things exist.
So awareness is another
issue.
Affordability
is another issue.
So many
of the projects that I mentioned,
what they do is actually they start not with the technology.
They start with the users
and their context and their
situation and what they're trying to do and then map back.
And technology is just only one of the pieces that these systems, all of these systems that I mentioned, right?
Technology is just only one component.
There's a socio-technical piece that deals with exactly these kinds of access and awareness and these kinds of issues.
And we're kind of taking a walk right now through the work
of the lab and there are some other areas that you want to get into. But I want to come back to
this. Maybe this is a good segue into the emotional journey part of the question I asked a few minutes
ago. As you get into some of the deep technical work of the lab, what were some of the first
impressions of the new technologies and what were some of the first things that you and your colleagues there and our colleagues
felt in observing these new capabilities?
So I think Peter mentioned this very eloquently as stages of grief.
And me and my colleagues, I think, went through the same thing. I mean, we went from disbelief saying, oh, wow, this is just amazing.
I can't believe this is happening to sort of understanding what this technology can do and over time understanding what its limitations are and what the opportunities are as a scientist and technologist and engineering organization to really push this forward
and make use of it.
So that's, I think, the stages that we went through.
Maybe I can be a little bit more specific.
As I mentioned, the three other areas we work on
are theory and algorithms
in machine learning and in systems.
And I can sort of say how my colleagues
are evolving their own technical and research agendas in the
right of this disruption. If you take our work in theory, this lab has had a track record of,
you know, cracking long-standing open problems. For example, problems like the Caddison-Singer
conjecture that was open for many years, many decades, was actually solved by people from the lab. Our lab has incredible experts in arithmetic
and circuit complexity. They came so close to resolving the VP versus VNP conjecture,
which is the arithmetic analog of the P versus NP problem. So we have incredible people that
working on theoretical computer science, and a lot of them are now shifting their attention
to understanding these large language models,
right?
Instead of understanding just arithmetic circuits, you know, people like, you know,
Neeraj Kayal and Ankit Garg are now thinking about mathematically what does it take to understand transformers?
How do we understand, how might we evolve these models or training data so that these models improve even further in performance,
in their capabilities, and so on. So that's actually a journey that the theory people are
going through, bringing their brainpower to bear on understanding these models foundationally.
Because as you know, currently our understanding of these foundation models
is largely empirical.
We don't have a deep scientific understanding of them.
So that's the opportunity that the theoreticians see
in this space.
If you look at our machine learning work,
you know, that actually is going through a huge disruption.
I remember now, one of the things that we do in this lab
is work on causal ML.
Amit Sharma, together with Emery Ketchum and other colleagues working on causal machine learning.
And I heard a very wonderful podcast that you hosted them some time ago.
Maybe you can say a little bit about what you heard from them, and then I can pick up back and then connect that with the rest of the lab.
Sure. Well, it's, you know, I think
the common knowledge, there's so many things about machine learning over the last few decades
that have become kind of common knowledge and conventional wisdom. And one of those things is that, you know, correlation is not causation and that, you know, learned models don't, you know,
generally don't do causal reasoning. And so we, you know, we've had very specialized tools
created to do the kind of causal reasoning that Amit and Emery do. And it was interesting. I
asked them some of the same questions I'm asking you now,
you know, about the journey and the initial skepticism. But it was been really interesting
to see how they're moving forward. They recently published a position paper on archive where they
conducted some pretty compelling experiments, in some cases showing something like, you know,
causal reasoning, you know, being exhibited, or at least, I'll say, convincing performance on causal reasoning tasks.
Yeah, absolutely.
Yeah, go ahead.
Yeah, absolutely. So, I would say that their journey was that initially, they realized that, I mean, of course, they built specialized causal reasoning tools like Do-Why,
which they've been building for many years.
And one of the things they realized was that, oh, some of the things that Do-Why can do
with sophisticated causal reasoning, these large language models were just able to do out of the box.
And that was sort of stunning for them, right?
And so the question then becomes, you know, does specific vertical research in causal reasoning is even needed, right? And so the question then becomes, you know, does specific vertical research and
causal reasoning is even needed, right? So that's actually the shock and the awe and the emotional
journey that these people went through. But actually, after the initial, you know, shock faded,
they realized that there is actually better together story that is emerging in the sense that they you know once you understand the details what they realized was that natural language contains a lot of causal information
right if you just look at the literature the literature has many things like you know
a causes b if there is if there is you know hot weather then ice cream sales go up. You know, this information is present in the literature.
So if you look at tools like DoY,
what they do is that in order to provide causal machine learning,
they need assumptions from the user on what the causal model is.
They need assumptions about what the causal graph is,
what is the user's assumptions about which variables depend on which variables, right?
And then, and what they realize is that models like GPT-4 can now provide this information.
Previously, only humans were able to provide this information.
But in addition to that, right,
tools like DUI are still needed to confirm
or refute these assumptions statistically using data.
So this division of labor between getting assumptions from either a human or from
a large language model, and then using the mathematics of do-why to confirm or refute
the assumptions now is emerging as a real advance in the way we do causal reasoning, right? So I
think that's actually what I heard in your podcast. And that's indicative of actually what the rest of
my colleagues are going through, you know, moving from first thinking about, oh, the GPT-4 is like a threat, you know, in the sense that
it really obviates my research area to understand, oh, no, it's really a friend.
It really helps me do, you know, some of the things that required primarily humor intervention.
And if I combine GPT or these large language models together with domain-specific research,
we can actually go after bigger problems
that we didn't even dare going after before.
Let me ask you, I'm going to pivot here in a moment,
but have you covered the areas of research in the lab
that you wanted to walk through?
Yeah, there's more. Thank you for reminding me.
Even in the machine learning area, right,
there is another work direction that we have
called extreme classification,
which is about building very, very classifiers
with large number of labels, you know,
hundreds of millions and billions of labels.
And, you know, these people are also benefiting
from large language encoders.
You know, they have come up with clever ways of taking these language encoders that are built using
self-supervised learning, together with supervised signals from things like clicks and logs from
search engines and so on, to improve performance of classifiers.
Another work that we've been doing is called DISC-ANN or Approximate Nearest Neighbor
Search. You know, as you know, actually in this era of deep learning, retrieval works by converting
everything in the world, you know, be it a document, be it an image, you know, be it audio or video
file, everything into an embedding. And relevant retrieval is done by nearest neighbor search
in a geometric space.
And our lab has been doing,
I mean, we have probably the most scalable vector index
that has been built.
And these people are positively impacted
by these large-language models
because, as you know,
retrieval augmented generation
is one of the most common design patterns
in making these large language models
work for applications.
And so their work is becoming increasingly relevant
and they are being placed huge demands
on pushing the scale and the functionality
of the nearest neighbor retrieval API
to do things like, oh, can I actually add predicates?
Can I add streaming queries and so on?
So they are just getting stretched with more demand for their work.
If you look at our systems work, which is the last area that I want to cover,
we have been doing work on using GPUs and managing GPU resources for training as well as inference.
And this area is also going through a lot of disruption.
You know, prior to these large language models, these people were looking at relatively smaller models.
You know, maybe not, you know, hundreds of billions to trillions of parameters, but maybe hundreds
of millions and so on. And they invented several techniques to share a GPU cluster among training
jobs. The disruption that they had was, oh, these models are so large that nobody is actually
sharing clusters for them. But it turned out that some of the techniques that they invented to deal with migration of jobs and so on are now used for failure recovery in very, very large models.
So it turns out that, you know, at the beginning, it seems like, oh, my work is not relevant anymore.
But once you get into the details, you find that there are actually still many important problems.
And the insights you have from solving problems for smaller models
can now carry over to the larger ones.
And one other area I would say is the area of programming.
I myself work in this area.
We have been combining machine learning together with program analysis
to build a new generation of programming tools.
And the disruption that I personally faced was that the custom models that I was building
were no longer relevant.
They aren't even needed.
So that was a disruption.
But actually what me and my colleagues went through was that, okay, that is true.
But we can now go after problems that we didn't dare to go before.
Like, for example, we can now see that Copilot
and so on let you
give recommendations in the context of the
particular file that you're editing.
But can we now edit
an entire repository
which might contain
millions of files with hundreds
of millions of code? Can I just say,
let's take, for example, the whole of the
Xbox code base or the Windows code base and in the whole code, can I just say, let's take, for example, the whole of the Xbox code base or the Windows code base, and in the whole code base, I want to do this refactoring,
or I want to, you know, migrate this package from, migrate this code base from using,
you know, this serialization package to that serialization package. Can we just do that,
right? I think we wouldn't even dare going after such a problem two years ago. But now with large
language models, we are thinking, can we do that?
And large language models cannot do this right now
because whatever context size you have,
you can't have a hundred million line code
as a context to a large language model.
And so this requires combining program analysis
with these techniques.
That's as an example.
And actually, furthermore,
there are many things that we are doing
that are not quite affected by large language models. For example, actually, youmore, there are many things that we are doing that are not quite affected by large-language models.
For example, actually, you know about the highway project where we are thinking about technology to make hybrid work better.
And we are doing work on using GPUs and accelerators for database systems and so on.
And we do networking work.
We do low-Earth-orbit satellite work for connectivity and so on. And we do networking work, we do low-Earth-orbit satellite work for connectivity
and so on. And those we are doubling down
though they have nothing
to do with large-language models because those are problems that are
important. So I think
to summarize, I would say that
most of us
have gone through a journey
from shock and awe
to some amount of insecurity
saying, is my work even relevant,
to sort of understanding, oh, these things are really aids for us. These are not threats for us.
These are really aids. And we can use them to solve problems that we didn't even dream of before.
That's the journey I think my colleagues have gone through.
I want to step into two of the concepts that you just laid out. Maybe just to get into some of the intuitions as to what
problem is being solved and how generative AI is sort of changing the way that those problems are
solved. So the first one is extreme classification. I think a flagship use of generative AI and foundation models is Bing chat. And so I think this idea of internet search as a, you know, as a home for these new technologies is in the popular imagination now. And I know that extreme classification seeks to solve some challenges related to search and information retrieval. But what is the challenge problem there?
How is extreme classification addressing that?
And how is that being done differently now?
So as I mentioned, where my colleagues have already made a lot of progress
is in combining language encoders with extreme classifiers to do retrieval.
So there are these models called NLR.
Like, for example, there's a tooling NLR model, which is a large language model, which does
representation, right?
It actually represents, you know, keywords, keyword phrases, documents, and so on into
encodings, you know, based on, you know, self, keyword phrases, documents, and so on, into encodings, you know, based on,
you know, self-supervised learning. But it is a very important problem to combine
the knowledge that these large language models have, you know, from understanding text.
We have to combine that with supervised signals that we have from click logs,
because, you know, we have search engine click logs. We know, you know, for example,
when somebody searches for this information
and we show these results,
what users click on, that's supervised signals.
And we have that in huge amounts.
And what our researchers have done
is they have figured out how to combine
these encoders together with
the supervised signals from click logs
in order to improve both the quality
and cost of retrieval,
right? And actually, as you said, retrieval is an extremely important part of experiences like
Bing chat, and retrieval augmented generation is what prevents hallucination and grounds these
large language models with appropriate information retrieved and presented so that the relevant results are grounded without hallucination, right?
Now, the new challenge that this team is now facing is, okay, that's so far so good as far as retrieval is concerned, right?
But can we do similar things with generation, right?
Can we now combine these NLG models, which are these generative models, together with supervised signals, so that even generation can actually be guided in this manner, improved in both performance as well as accuracy.
And that is an example of a challenging problem that the team is going after.
Now let's do the same thing with programming.
And maybe I'm going to engage you on a slightly higher level of abstraction than the
deep work you're doing. And then we can get back down into the work. But one of the popular ideas
about these new foundation models is that you can effectively, through interacting with them,
you're sort of programming them in natural language.
How does that concept sit with you as someone who, you know,
is an expert in programming languages?
What do you think when someone says, you know,
you're sort of programming the, you know, the system in natural language?
Yeah, so I find it fascinating. And, you know, for one, you know, can be an important, the programming language community has been able to solve it only
in narrow domains.
For example, Excel has Flash Fill
where through examples
people can program Excel macros
and so on. But those are not as general
as these kinds of
LLM
based models.
And it is for the whole community
not just me, right? It was stunning
when users can just
describe in natural language what program they want to write
and these models emit Python or Java or C-sharp
code. But there is a gap
between that capability and having
programmers just program in natural language, right? Like, you know,
the obvious one is, I can sort of Like, you know, the obvious one is,
I can sort of say, you know,
write me Python code to do this or that,
and it can generate Python code, and I can run it.
And if that works, then that's a happy path.
But if it doesn't work, what am I supposed to do if I don't know Python?
What am I supposed to do, right?
I still have to now break that abstraction boundary
of natural language and go down into Python
and debug Python.
So one of the opportunities that I see is that can we build representations that are also in natural language, but that sort
of describe, you know, what the application the user is trying to build and enable non-programmers,
could be lawyers, could be accountants, could be doctors, to engage with the system purely in
natural language. And the system purely in natural language,
and the system should talk back to you saying,
oh, so far, this is what I have understood.
This is the kind of program that I am writing
without the user having to break
that natural language abstraction boundary
and having to go and understand Python, right?
I think this is a huge opportunity
in programming languages to see whether,
can we build, like, for example, right,
actually, right, I'm a programmer, and one of the things i love about programming is that i can
write code i can run it see what it produces and if i don't like the results i can go change the
code and rerun it and that's sort of the you know coding evaluating uh we call it the ripple loop
right so that's that's what a programmer faces right can we now provide that to natural language
programmers in the sense that i want to say here's's the program I want to write. And now I want to say, oh, I want to run
this program with this input. And if it doesn't work, I want to say, oh, this is something I don't
like. I want to change this code this way, right? So can I now provide that kind of experience to
natural language programming? I think that's a huge opportunity if you manage to pull that off.
Mm-hmm. And now let's maybe return to some of the more societally oriented topics that you were talking about at the top of the episode.
In the context of programming, because being able to program in natural language, I think, really changes who can use the technologies, who can develop technologies, what a program, what a software development team can actually be and who that kind of a team can consist of.
So can you paint a picture, you know, what kind of opportunities for, you know, software development does this open up when you can sort of program in natural languages, assuming we can make the AI compatible with your language, whatever that happens to be?
Yeah, I think there are a lot of opportunities.
Maybe I'll describe a few things that we're already doing.
My colleagues are working on a project called Vellum, which is now a co-pilot assistant for societal scale applications.
And one application they are going after is education.
So, you know, India, like many other countries, has made a lot of educational resources available to teachers in government schools and so on.
So that if a teacher wants to make a lesson plan, you know, there is enough information available for them to search, find out many videos
that their colleagues have created
from different parts of the country
and put them together to create a lesson plan
for their class, right?
But that is a very laborious process.
I mean, you have an information overload
when you deal with it.
So my colleagues are thinking about,
can we now think about,
in some sense, the teacher as a programmer and have the teacher talk to the Vellum system,
saying, hey, and here is my lesson plan. Here is what I'm trying to put together in terms of
what I want to teach. And I now want the AI system to collect the relevant resources that are relevant to my lesson plan
and get them in my language, in the language that my students speak.
How do I do that?
And all of the things that I mentioned, you have to now index all of the existing information using vector indices.
You have to now retrieve augmented generation to get the correct thing. You have to now deal with
the trunk and tail
languages because this teacher might be speaking
in a language that is not
English, right?
And the teacher might get a response
that they don't like, right?
But they're not a programmer. How are
they going to deal with it, right? So that's actually an example.
If we pull this off, right,
and a teacher in rural India
is able to access this information
in their own language
and create a lesson plan
which contains the best resources
throughout the country, right,
we would have really achieved something.
Yeah, you know,
it's a hugely compelling vision.
And I'm really looking forward
to seeing where you
and our colleagues
in Microsoft Research India Lab
and MSR more broadly,
you know, take all these different directions. Lab and MSR more broadly, you know,
take all these different directions. So I really appreciate you spending this time with me today.
Thank you, Ashley. And I was very happy that I could share the work that my colleagues are doing here and bringing this to your audience. Thank you so much.