ACM ByteCast - Peter Stone - Episode 84

Episode Date: April 16, 2026

In this episode of ACM ByteCast, Rashmi Mohan hosts 2024 ACM/AAAI Allen Newell Award recipient Peter Stone, Professor at the University of Texas at Austin and Chief Scientist at Sony AI. He received t...he award for significant contributions to the theory and practice of AI, especially in reinforcement learning (RL), multiagent systems, transfer learning, and intelligent robotics. As a leading figure in AI research, Stone has fundamentally advanced how autonomous agents learn, plan, and collaborate. His groundbreaking work on RL algorithms has enabled robots to acquire skills through experience. He is an ACM, AAAI, AAAS, and IEEE Fellow, an Alfred P. Sloan Research Fellow, and a Fulbright Scholar. At UT Austin, he is the founder and director of the Learning Agents Research Group (LARG) within the Artificial Intelligence Laboratory, as well as Founding Director of Texas Robotics. In the past, he also worked at AT&T Labs - Research and co-founded Cogitai, Inc. (acquired by Sony). Peter explores the intersection of professional research and personal passion, detailing how his lifelong love for soccer fueled his involvement in RoboCup, where he aims to develop humanoid robots capable of competing at a World Cup level by 2050. The conversation highlights his leadership as the Chief Scientist of Sony AI, focusing on landmark projects like GT Sophy, an AI that mastered the complexities of Gran Turismo, and the development of FHIBE, an ethically sourced dataset designed to mitigate bias in machine learning. Throughout the interview, Stone emphasizes the importance of ad hoc teamwork—the ability of autonomous agents to collaborate on the fly with unfamiliar partners. He also shares his passion for undergraduate research and advocacy for AI education at all levels.

Transcript
Discussion (0)
Starting point is 00:00:01 This is ACM Bythcast, a podcast series from the Association for Computing Machinery, the world's largest educational and scientific computing society. We talk to researchers, practitioners, and innovators who are at the intersection of computing research and practice. They share their experiences, the lessons they've learned, and their own visions for the future of computing. I am your host, Rashmi Mohan. We've all heard the warnings about AI taking up. over the world. But today's guest is more interested in whether AI can win a championship. He is a mastermind behind the digital athlete, creating machines that can navigate a 180 mile
Starting point is 00:00:46 per hour pair pin turn with more finesse than a pro racer and lead a squad of robotic soccer players to a world title. But his work isn't just about games. He has been the architect behind many complete intelligent agents for many decades. Dr. Peter Stone is the Trushard Foundation Chair in Computer Science at the University of Texas at Austin, the founder and director of the Learning Agents Research Group within the AI Lab, director of Texas Robotics, and the chief scientist of Sony AI. From leading the Global Robo Cup Federation to developing G.T. Sophie, the AI that outraced world champion drivers in Grand Tourismo, his work focuses on the ultimate challenge,
Starting point is 00:01:35 creating autonomous agents that can navigate the messy, unpredictable reality of the physical world. Peter is also an ACM fellow and has most recently won the ACM Triple AI Alan Newell Award, amongst many other honors. We are so excited to speak with you, Peter. Welcome to ACM Bikecast. I'd love to lead with a simple question that I ask all my guests,
Starting point is 00:01:59 which is if you could please introduce yourself Tell us about the work that you do and give us some insight into what drew you into this field. Yeah. Thank you for inviting me. It's a pleasure to be here. My name is Peter Stone. I'm a professor of computer science at the University of Texas at Austin and also chief scientist at Sony AI. And my research has always been in artificial intelligence. So from far before, it was as cool as it is now. In fact, when I got into the field in the early 90s, many people thought that its time had come and gone. But what drew me, to the study of artificial intelligence was really the mystery of how the brain works and what is the nature of intelligence. I'm not the first to say in the AAAI presidential speech where Kambapati laid out what he thought was the three big challenges of our time, scientific challenges. One is how did the universe begin? How did life get started on Earth? And then the third is,
Starting point is 00:02:54 what's the nature of intelligence. And that's what piqued my interests, starting from when I was in high school. And so when I went to University of Chicago as an undergrad, I explored different ways to try to get at that question. And I went to some neuroscience classes, some psychology classes. But neither of those really hit the mark for me. In the neuroscience classes, we were learning how, you know, about axons and dendrites and neurotransmitters and sort of the molecular questions of how the brain works at that time. And I sort of thought, but yeah, but that's not telling me anything about what happens in my brain when somebody tells me their telephone number and I remember it or something like that, which was a thing back in the day.
Starting point is 00:03:32 Now we don't have to remember anybody's telephone number. We just put him into our phone. And so I went to some psychology classes. And there it was very, very interesting to learn about what people knew about human behavior, what caused people to, you know, under what circumstances people would behave in different ways. I learned all about the work of Kahneman and Tversky. And it was fascinating. But again, that was too high level as opposed to the neuroscience, which seemed too level for me.
Starting point is 00:03:56 It still wasn't telling me what's going on in the brain, when I remember something. And so then I started going to the computer science classes and was introduced to artificial intelligence. At the time, there was actually no computer science major at University of Chicago. I was a math major, but took all the computer science courses I could, including some graduate courses in artificial intelligence.
Starting point is 00:04:15 And that's where it sort of, to me, felt like there was a way to try to get it, what is the nature of intelligence by trying to recreate intelligence. They have theories about how people may be doing intelligent things and try to recreate those in computers. and that fascinated me. In some sense, the rest is history, but it's been a long path since then.
Starting point is 00:04:33 I did go on to graduate school at one of the top places for artificial intelligence time still, Carnegie Mellon University. I was fortunate to have a fantastic PhD advisor who gave me the freedom to explore the ideas that most interested me. That's Professor Manuel Velosso.
Starting point is 00:04:50 And then when I graduated, I went to AT&T Labs research, a fantastic team, the best people in the world working on reinforcement learning and other forms of artificial intelligence, at the time, was there for about three years, and then in 2002 came to University of Texas at Austin as a new assistant professor. Wonderful. Thank you for sharing that. And I mean, I think you
Starting point is 00:05:08 already called this out for, you got into this field so much earlier than the excitement of today. I know that your work is significant. You've made a lot of investments as well as progress in multi-agent systems and interactions. That has been sort of the cornerstone of your research. Why did this area in particular fascinate you, Peter? And like sort of what has been the progression of thoughts and ideas that you have seen in this field? Yeah, that's a great question. Thank you for that. Yeah, I usually say my research is sort of at the confluence of three different sub-areas of artificial intelligence. Multi-agent systems is one of them, for sure, and then also machine learning, specifically reinforcement learning, and robotics and putting those three together.
Starting point is 00:05:47 But multi-agent system is the way I got into it. It was actually, just after my first year as a PhD student, during the first year, I was immersing myself in the area of artificial intelligence planning, There was a system at Carnegie Mellon called Prodigy, that my advisor was key participant in. So I was working on sort of ways to do better and more robust planning. And then I went to the main AI conference in the summer of 1994. I was in Seattle. So this was AAAI, 1994. And there there was a demonstration by Alan McQuartz Lab from University of British Columbia and his student at the time, Michael Sohota.
Starting point is 00:06:27 And they were showing a one-on-one soccer match. And it was two robots on a ping pong table playing soccer, trying to knock the ball into goals on either side. They had, Alan at the time, I think in 1993, had published a paper, putting forth robot soccer is a great challenge task for artificial intelligence. And so they were showing this demonstration in connection with that. And I looked at that. I was on my soccer team as an undergrad. I've played my whole life. I was on the varsity team at the University of Chicago.
Starting point is 00:06:55 and I got there and I was immediately captivated. I was like, oh, robots playing soccer, that's fantastic. It could merge my personal interest, outside interest, with my professional interest in artificial intelligence, but it felt like something was missing. One-on-one to me did not make a soccer match, did not make a soccer game. You need to have multiple teammates on each side.
Starting point is 00:07:16 And so I sort of made it my goal to expand what they were doing into at least three-on-three, eventually up to 11-on-11, which is what people do. And that got me deeply into the sort of burgeoning field at the time of multi-agent systems where people were thinking about, you know, if you created a single autonomous agent, that wouldn't be, you know, complete enough. They would always need to interact either with people or with each other. There were people thinking about how could autonomous agents communicate. They were thinking, people thinking about what would happen if they were competing. And soccer sort of had this dual aspect of multi-agent systems, both teammates that you collaborate with and adversaries. that you compete against. So it's both collaborative and adversarial at the same time. And so as I dove into that and learned about the literature in multi-agent systems in the mid-1990s, I learned that people were making a strong case that really resonated with me, which is that a lot of intelligence, a lot of human intelligence, is focused on our ability to communicate
Starting point is 00:08:14 and interact with other agents. We can look at other people and predict what they're going to do. we make assessments about whether they're good to cooperate with or not based on their body language or what their facial expressions are. People were starting to make the case that multi-agent systems and interaction is one of the main seats of intelligence and might be one of the keys to unlocking artificial intelligence. So I became sort of involved in all aspects of artificial intelligence, not just the ones that were motivated by robot soccer. I've been involved in other types of autonomous agents like autonomous bidding agents,
Starting point is 00:08:49 autonomous traffic agents, you know, autonomous cars are also in a multi-agent system. And they're neither collaborative nor adversarial, by the way. They're just in a system with other self-interested agents. And so I've sort of, through my career, looked at many different aspects of multi-agent systems, and it continues to fascinate me. I could talk later about ad hoc teamwork, which is something that's area of multi-agent systems that I helped introduce and which has fascinated me over the last 15 years or so. Oh, absolutely.
Starting point is 00:09:15 I mean, this is fascinating. I was thinking through, I was trying to do some sort of research on your work, and I saw your deep involvement in Robo Cup, and I was wondering how that engagement began. You know, you're one of the leading figures that participates and develops strategies for this. I'm wondering, when you have so many agents that are all sort of interacting in this system, you know, as we do as humans, right? We're optimizing, in the case of a soccer match, trying to optimize for the same goal, what are the sort of key considerations that you need to keep in mind as you are trying to design this system?
Starting point is 00:09:45 It's a very complex question. There are many things. There's no, and one of the fascinating things about Robococor and Robot Soccer is that there's challenges at so many different levels. So first of all, I should say, yeah, I was very fortunate. I told you about how I became interested in multi-agent systems based on this demonstration I saw at AAA 924 from Alan McQuart. You know, I became captivated by it. My advisor gave me the freedom to explore in that direction. And it just so happened.
Starting point is 00:10:11 You know, I was in the right place and the right time. There were people from around the world that were thinking about trying to turn robot. soccer into really a global challenge problem and connected with some colleagues in Japan, Hiroaki Kitano, Izki Notam, Minora Asada, and others who were thinking of trying to create a robot soccer community, the competition, which they did. And I was one of the, you know, sort of, I was, I think the first participant in the United States, along with my advisor, along with Manuela. And it sort of turned into something that is now a real global community. There's thousands of people around the world who have become captivated by the challenges that we look at in Robocop.
Starting point is 00:10:49 One of the sort of main challenges, the founding challenge that Hiroaki Kitano put forth to the community was to create a team of humanoid robots that can beat the World Cup champion soccer team on a real soccer field by the year 2050. We're more than halfway there from when he first stated that, I think, in the late 1990s. Robocup has expanded since then to include things like disaster rescue and general purpose service robots at homes, there's a RoboCup at Home, RoboCup at Rescue. But to answer your question, what are the key challenges in robot soccer? There's, you know, they sort of span the gamut from starting at the low-level hardware issues, just how do you get a robot to move quickly and
Starting point is 00:11:27 agilely to control a soccer ball? There have been, you know, a lot of Robocuff historically has been with wheeled robots, but now there's a lot more focus on humanoid robots. But even, you know, with wheeled robots, how do they intercept the ball and move it in the direction that they want to? just an individual, how can it sense perceive and a sense decide and act on its own? That's on top of the hardware once you have the low-level capabilities. How do you intercept the ball? How do you kick the ball? How you walk fast if you've got legs? And then you get into the strategic and multi-agent issues. How do you coordinate with teammates? How are you flexibly play different positions?
Starting point is 00:12:03 So not all the robots are just chasing the ball like with five-year-olds playing soccer. So how do you organize the team strategically? then how do you play differently based on the opponent that you're playing against or the situation in the game or there's challenges at many different levels. Part of the reason that I've stayed involved and stayed fascinated by it throughout my career is I've had the experience that it has spun off many interesting research questions. I'd say probably I've had more than 25 students graduate with PhDs from my lab now and I think at least half, if not more, have been inspired by some challenge in robot soccer. None of them wrote a thesis that said,
Starting point is 00:12:40 goal is to solve robot soccer or that they did. It's still an unsolved problem. But they found some problem, some sub-problem that motivated them. For instance, one of my first students, Mohan Sriturran, worked on the computer vision aspects of robot soccer. And just at the time, you had to calibrate the cameras of the robots very carefully based on the lighting in the room. And then if the light turned off or a bulb burnt out, all of a sudden the robots wouldn't be able to see the ball anymore. He wanted to make robots more robust to that. So a thesis that went very deeply on the problem of robot perception, robot vision, that, you know, ended up not being at all really about robot soccer, but was motivated by a problem that he came across there. And that's just one
Starting point is 00:13:20 example. There's been, I'd say, you know, like I say, more than 10, probably 15 pieces from my lab that have been motivated by those problems. And I think there's still many open problems in robot soccer that they're going to, that are PhD thesis worthy. And so hopefully as a community, I'm certainly not the only person working on it. I have a great set of colleagues internationally that we're working together in some sense, but using a competition every year as our meeting point where we sort of benchmark what our progress is every year. And I hope that collectively we'll be able to reach that goal of 2050. I still don't know if it'll be possible. I look forward to that. But I think a lot of things that you say, you know, make a lot of sense
Starting point is 00:13:58 in the sense that the principles of interaction that you see within a robo soccer game can be applied across so many different applications, even if not all sort of working towards the same goal, but interacting with each other, multiple sort of systems or agents interacting with each other. One question I had around that, Peter, was what happens if there is an error, right? A robot makes an error. How do you recover from that? Is there a coach? Is there some learning that is happening or coaching that is happening on the fly? Yeah. So, first of all, the robots make errors all the time. But one of the challenges with robotics is it's a lot harder to figure out to trace back an error that a robot makes than it is a traditional laptop or desktop. It's hard to debunk things because it's hard to understand
Starting point is 00:14:42 what's the internal state of the robot without compromising its computations and its performance. But I think you're asking about, look, yeah, is there that next level of sort of oversight of coaching that humans have, which I didn't, I left out when I was talking about all the challenges of robot soccer, but you're absolutely right. In fact, in Robocop, there was for several years a coach competition where all the teams were given sort of fixed players and what they're job was to come up with a coach that could analyze the current team that they were playing against and impart high-level strategic advice in a formal language. The question was who could sort of guide these teams of agent to perform the best. And yes, that requires them to be adaptive,
Starting point is 00:15:24 to be able to change their behavior based on the commands that, or the advice that the coach would give. And then, you know, sort of the other dimension of your question is just can the robot soccer players learn? And there's opportunities at two different granularities. Early in my career, I worked a lot on the robots learning before the games, some sense, learning skills and learning strategies well before, when there's no restriction on it has to happen quickly during the course of a game. You can take lots and lots of data. An early influential paper I wrote was a collaboration with Rich Sutton, who recently won the Turing Award along with Andy Bartow for their pioneering work in reinforcement learning. And he and I created a multi-agent
Starting point is 00:16:04 reinforcement learning agent to play the game of robot soccer keepaways. So you can imagine like three, and this was in the RoboCup simulator, not with real robots, but you could imagine three simulated agents trying to just keep the ball away from one in a small space in a square, just figuring out when to pass and where to pass. This is a drill that human soccer players do all the time. And so we sort of took that and used reinforcement learning to have the agents learn to cooperate, to form a team, to be able to the keepers, the ones that are trying to keep the ball away from the other, to have them figure out where to move, to when to pass, where to pass.
Starting point is 00:16:38 And we started them with making those decisions randomly, but over the course of 24 hours of just, you know, working on their own and experiencing what works and what doesn't work, we were able to find that they could keep the ball away from the take of her, I think, two and a half times longer than random. They improved dramatically. This all happened, like I say, sort of at a single game time scale. This was at the scale of, you know, at least a day with a lot of computation
Starting point is 00:17:04 power. And there's been other examples with Josiah Hanna. We used a human that we used machine learning to create a humanoid robot, the now robot, to have it learn how to walk faster than any other group in the world had allowed that, had taught that robot to walk. And again, that was using machine learning, a sim to real kind of approach where we learned partly in simulation and partly in the real world, a method that we introduced called grounded simulation learning. But again, that all took place not during the course of a single game. Then there's, as I've been hinting, also the granularity of what can you do during the course of an individual game. There's a lot less opportunity for adaptation. People do that. I don't think there's been lots of examples yet of
Starting point is 00:17:45 autonomous robots being able to quickly figure out what the opponents are doing. I've had a little bit of work where in my PhD thesis, I created some extreme scenarios where the opponent, for some reason, only plays on one side of the field or the other side of the field and tested to see whether the agents, my team could on the fly learn to adapt its strategy accordingly. And also this learning on the fly very quickly is one of the main challenges of an area that I mentioned earlier called ad hoc teamwork. And ad hoc teamwork is motivated by people's ability to do things like playing pickup games of soccer or basketball and to cooperate on the fly in other ways. So for example, if I, you know, I'm a soccer player myself, if I go see other people playing soccer in a foreign
Starting point is 00:18:30 country. I've never seen the people before. I may not even speak the language. I can jump onto the field and start playing with them immediately. And I need to adapt my strategy based on quickly assessing, you know, am I the worst player on the field? Am I the best player on the field? There's nobody who knows how to play defense, so I should do that. This sort of collaboration with teammates on the fly, I introduced the term ad hoc teamwork to describe that. And that requires very quick adaptation, very quick modeling of what your teammates can do and then very quick assessment of how you can make the team shine. People can do this all the time, right?
Starting point is 00:19:05 If there's a disaster on the street, if there's an accident and a bunch of strangers see two cars crash into each other, even if they're all strangers, they'll immediately self-organize into a team and have one try to rescue the victims, one call 911, direct traffic around. People can just, you know, can do this. Traditionally, robots have not been able to do that
Starting point is 00:19:22 because they've needed to train together to become a team or they've had to have hand-coded paradigms or protocols given to them to become a team. And so I've been very fascinated by this second dimension of multi-agent learning, of doing it quickly on the fly with new TVs. Thank you for sharing that. So many interesting concepts in there. You also have a very unique take on just reinforcement learning, but with this new age of LLMs and the vast knowledge that is now available,
Starting point is 00:19:51 how do you see those two coming together? Yeah, that's a great question. And, you know, I think the crux of reinforcement learning is learning from experience and being able to change behavior based on trying in action and seeing what happens and seeing if your expectations for what would happen are met or whether they're going to change. And at least the initial LLMs and, you know, sort of the ones that have gone viral don't do that, right? They're trained offline. They first through pattern recognition, traditionally with transformers. There is a fine-tuning phase that people call reinforcement learning from human feedback, but it's not after deployment.
Starting point is 00:20:32 It's still in the lab back before it gets deployed to the real world. This concept actually of reinforcement learning from human feedback in some ways was introduced by one of my early students, Brad Knox. He created a system called Tamer that also learned from people saying, you know, yes and no, or good job, bad job to an agent, found that it was very powerful. So it's really nice to see that's playing a role in, the training of large language models. But once they're deployed into the real world, they're not learning anymore, right? They're fixed entities. Their weights are there. Of course,
Starting point is 00:21:04 people are working on changing that, but I think that's the missing piece for the sort of initial large language models is getting them to learn from experience in the way that reinforcement learning agents do so that they can improve their performance based on trying things and seeing what works it does it in the world. So I'm excited to see what the confluence of these areas will bring, the power of large language models for pattern recognition, in language and vision, and now in some sense in actions for robots,
Starting point is 00:21:32 with the ability to try things through trial and error, see what works and see what doesn't, and be able to change behavior, change weights, change the internals after deployment. ACM Bytecast is available on Apple Podcasts, Google Podcasts, Google Podcasts, Podbean, Spotify, Stitcher, and Tune In. If you're enjoying this episode, please subscribe
Starting point is 00:21:57 and leave us a review on your favorite platform. I want to talk a little bit about your role as chief scientist at Sony AI. You obviously had a very long and successful career in academia, but you've also had a lot of interactions with industry, and now you have a role very much in a corporation. What led you there? Tell us a little bit more about. about the work that's coming out of that rule?
Starting point is 00:22:22 I never really considered myself entrepreneurial at all. I saw myself as an academic through and through. I really love my role at the university and working with students and doing basic research. I think it was AAAI in Austin, around 2014 or 2015 sometime. I was approached my longtime longtime colleague of mine, somebody who I met in graduate school, Mark Ring,
Starting point is 00:22:44 who pioneered the concept of continual learning, which is very related to reinforcement, learning, but is in some sense, in some ways goes the next step of not just learning a single task from experience, but learning an ongoing stream of skills and knowledge and tasks over the quote lifetime of an agent. So learning continually, not just, you know, learning to achieve something and then say, now I know how to do it. You know, he approached me with the idea of founding a startup company to take that kind of idea to industry to see where we could go with it. We were joined also by a longtime colleague of mine, Satinder Singh.
Starting point is 00:23:21 Well, we met early on. We were colleagues at AT&T Labs research back in the late 1990s and early 2000s. So the three of us teamed up to found this code we called a Kojitai, a startup company, and continual learning and hired up a fantastic team. And then ended up in 2019 being acquired by Sony. And this is what led to the founding of Sony AI. So Tinder and Mark both went different ways by either before or right at that time. But I continued on with many of the employees that we had brought together at Kogitai, some of my star former students.
Starting point is 00:23:55 We sort of seated a new organization within Sony called Sony AI. And that is in collaboration. The leaders of that have been Pete Werman, who was a part of Kodgatai and is one of the founders of Kiva systems, the fantastic robots that transformed Amazon warehouses. So he has been a part of the team since the beginning. And also Mika Spranger, been within Sony for many years. and Hiroaki Kitano, who I've already mentioned as one of the founders of RoboCoh. And so, you know, we all came together to form Sony AI with our initial challenge of taking
Starting point is 00:24:26 the technology of continual learning and reinforcement learning to Grand Turismo, the really high-fidelity racing simulator that is a very successful PlayStation game of Sony. You mentioned this in the introduction. We eventually, three years later, so we started, I guess, we officially launched in 2020, And then by a couple years later, we had an article on the cover of Nature featured G.T. Sophie, our agent, that is able to outrace the very best human drivers in Grand Tresmo. People who've put in their 10,000 hours to become experts. People take this very seriously. There's international competitions.
Starting point is 00:25:02 People use it as a training ground to become professional race car drivers. And it wasn't known if an AI agent could ever do better than the best human participants at this real-time control. task. AI agents that could beat people at chess and checkers and poker and jeopardy and other games, but it hadn't really be done in such a physically realistic real-time control task. And so it was, I think, a real landmark result. That's why it was featured on the cover of nature. And it was also very important to one of the products of Sony. So it was a perfect project in some sense that it was a landmark scientifically and also useful to the company. And so that's is one example of the kinds of things we try to do at Sony AI to do really ambitious projects
Starting point is 00:25:45 that require some of the top researchers around the world. We've been fortunate to hire a fantastic team, now a couple of hundred people working on several different problems, each of which either is it succeeds is ambitious enough to be a real scientific landmark, the kind of thing that could lead to an article in science or nature, or is directly useful to the company, but ideally both. As chief scientist, I serve as the scientific advisor to all of these projects. They span the gamut, though the GT Sophie one was our first one that we were able to go public with. The second one was featured on the cover of nature just in December of 2025, so just a month or two ago from this conversation. And it was from our AI ethics team that's led by Alice Jean.
Starting point is 00:26:32 It was the description deployment of a system called Phoebe, which is a data set of, of images that are collected in a completely transparent and ethical way in the sense that all participants are compensated for appearing in the data set. The labels that describe them are all self-described. So it's not an external person saying what the race or nationality of a person is, but the person who's shown in the picture self-identifying. It's very diverse in the sense of people from, I think, 81 different countries around the world, different ages, genders, everything balanced in the way, and it's meant to give a blueprint for how can we create ethical data sets for artificial intelligence, as opposed to just scraping the web and using whatever
Starting point is 00:27:18 happens to be out there on Reddit and Instagram and things like that. And then also to sort of identify biases in existing methods. It's not a large enough data set to train a new vision model, but it is large enough to evaluate existing models, and that's how it's used in the paper. So that's a second project from Sony AI that was able to land on the cover of nature. We have another project that has actually been accepted as a paper in nature. It'll be, it should appear within, probably around March, March or April in this year of 2026. Watch for that. Another, I think, really big landmark result. And then we have about, you know, five or six other projects within Sony AI that are all, you know, very ambitious scientifically and also, you know, are tied to
Starting point is 00:28:00 the interest of Sony, which is an entertainment company that, you know, in many ways. Sony does a lot of different things. There's hardware cameras, but there's also Sony music and movies and PlayStation. A theme is trying to bring fulfillment and happiness to people who interact with technology. I really resonate with that message. The business isn't about getting people to click on ads and buy things,
Starting point is 00:28:22 but really about sort of having a fulfilling and enjoyable time in life, looking for ways to bring cutting edge artificial intelligence to that aspect of humanity. Yeah, that's amazing. It's fascinating. I think the work that you just spoke about in terms of Phoebe and building ethical data sets, I mean, such wide impact, right? I mean, it serves as a great example. It serves as a way for people to sort of model and mimic that and, you know, continue to build on that. It's amazing that it is also in line with what the organization, that's, you know, that's an actual business that has to sort of meet its goals. Thank you for sharing that, Peter. Yeah, and Sony really prides itself on being it as a, has a, a rating as one of the most ethical companies in the world for several years in a row. And I think, you know, this is this aspect of wrestling with how can we use artificial intelligence to advance technology, but also well keeping people in control of, you know, the creative process and doing it in a way that keeps humanity centered. That is really a core principle personian.
Starting point is 00:29:24 You know, this isn't just an add-on for the company. It's really important to be able to think about how to use artificial intelligence in a way that magnifies humanity. I think also the hallmark of a great research lab is one. You're exploring brand new ideas that can actually set off a series of other research questions and explorations while also finding a way to actually apply this to real world situations and products and services that people want to use. One other area that I wanted to talk about a little bit is you strongly believe in undergraduate research. I recently saw a video of yours. You're fueling a lot of programs within the university for that target audience.
Starting point is 00:30:02 Would you like to talk a little bit about that and also areas that you think, you know, students should be focused on? I also read a little bit about your recent paper on AI literacy. So just curious more about your engagement on those aspects of your work. Yeah, I'd be delighted to. Thanks for the opportunity. In today's world, AI education is critical for everybody at different levels. There's many different populations for AI education. When you talk about undergraduate research, I've been, had the good fortune to be involved in a really innovative program here at UT Austin called the freshman research initiative that introduces
Starting point is 00:30:37 first year students in the spring semester of their freshman year to research in sort of a class setting but introducing them to the skills they would need to do research in a particular lab. And then the students who are successful go on to do research projects in the fall of their sophomore year and then to mentor the next cohort in the spring of the sophomore year. And the ones who are successful end up then having another couple of years to be engaged, in research for their undergraduate career, which is very different than, you know, in many places where undergraduate research happens in the last semester or last year of the career. So this is a way of nurturing and making undergraduate research really a large part of the undergraduate experience.
Starting point is 00:31:16 This happens across the College of Natural Sciences in UT Austin, so not just in computer science, but there's streams in physics and chemistry and biology. I've led one in computer science on autonomous robots since 2007. It's now being run by Justin Hart, but, you know, it's been very successful at identifying some really amazing students. That's within the computer science majors. Connected to that, we also launched recently a honors program in robotics that students can apply to from high school that allows them to get into this freshman research initiative with some new streams in robotics and also to take a minor in robotics.
Starting point is 00:31:50 And we had an immense success. It launched for the first time this year and for the 50 spots. So I think we had more than 3,000 applications. So we're hoping to be able to expand that. We have also an online master's program in artificial. intelligence that was launched recently. We actually have three online master's programs that service right now almost 5,000 students, one in computer science, one at data science, and now one in artificial intelligence. This is allowing us to a very affordable price of $10,000 for the whole
Starting point is 00:32:16 master's, so $1,000 a course, and it's allowing us to really at-scale, educate professionals who are trying to go back and upskill in the technical aspects of artificial intelligence. I teach the reinforcement learning course in the program. We have all kinds of courses there in that. been very successful. And I should say, you know, one of the unique properties of UT Austin is our size, our scale. So we're a top 10 department, very high quality, but we're the largest top 10 department, I believe. So quality at scale, it's the biggest combination of quality and quantity. And so we do that both at the undergraduate level. We have more than, I think, 2,400 undergrads currently, and then this online master's where we're scaling up to thousands of students. So those are for the computer
Starting point is 00:32:56 science majors and master's students. But nowadays, I think it's, you know, as an educator, it's imperative that all students, regardless of discipline, whether they be in the humanities, in fine arts, social science, that they all have at least minimal literacy in artificial intelligence to be able to understand what's realistic, look at the headlines, and think what's height and what's grounded. And so together with colleagues, especially Jouillejit Biswas and Don Fasel, we launched a course on the essentials of AI for life and society, which is designed as an undergraduate course for anybody. It scales up to hundreds of students. We hope it'll actually go to thousands in the coming years. All of the materials are online. And I think they're appropriate, by the way, for, you know,
Starting point is 00:33:38 anyone from upper high school to adult to lifelong learners. I think, you know, there's no programming background needed. You don't need to have come in with any prerequisites. It's an introduction to what is this term artificial intelligence, what's the history of it, what have been the different paradigms, what are the societal impacts. There are some homework assignments that are all online, but they're not, you know, they don't require programming, as I said. I've been really excited about this as, I think, a responsibility. I hope that we get to a point where every university student for sure, and maybe even every high school student, has at least one course in artificial intelligence.
Starting point is 00:34:12 And we've designed this course to be that one course. If it's the only course someone ever takes in artificial intelligence, this should be the one, I think, because it'll give them that grounding and perspective about what the field is all about. And it can also serve as a launching point for deeper interest if somebody does want to go on then take more in-depth courses. So all, like I say, all the materials are online from my webpage. It's easy to find them. I would be delighted if anybody can take them and repurpose them for their own courses. I'd love to hear about that directly from people doing that. We've written
Starting point is 00:34:40 an article about the course. A couple articles now in EAAI, the Educational Advances of Artificial Intelligence Symposium presented it just last week, the second paper on it. And so we're trying to, you know, not only did we create the course, but we're taking big efforts to make it so that it can be impactful and useful for everybody, and I hope people will take it and run with it. Oh, fantastic. I mean, I think, you know, one, reducing that barrier for entry into artificial intelligence, because there's a term that's bandied about an espit for somebody who's not even in technology, it can be intimidating, and for actually providing, you know, material and ways in which people can take this and sort of propagate it even further. This has been an absolutely
Starting point is 00:35:20 riveting conversation, Peter. Thank you so much for taking the time to speak with us at ACM Bycast. My pleasure. Thanks for. inviting me. ACM Bipecast is a production of the Association for Computing Machineries Practitioners Board. To learn more about ACM and its activities, visit acm.org. For more information about this and other episodes, please visit our website at learning.acm.org slash bikecast.
Starting point is 00:35:50 That's learning.acm.org slash B-Y-T-E-C-A-S-A-A-S-E-A-S-E. Thank you.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.