ACM ByteCast - John Hennessy & Dave Patterson - Episode 1

Episode Date: April 29, 2020

In this inaugural episode of ACM ByteCast, Rashmi Mohan is joined by 2017 ACM A.M. Turing Laureates John Hennessy and David Patterson. Their conversation touches on the paths that led these two lum...inaries to pursue computing careers and the "aha moment" that inspired their breakthrough work on RISC microprocessor architecture. They also discuss how they see the future of computing architecture unfolding in the coming years, the need for new memory technologies and better security, the importance of collaboration in innovation, and the promise of the open source community to develop both better software and hardware.

Transcript
Discussion (0)
Starting point is 00:00:00 This is ACM ByteCast, a podcast series from the Association for Computing Machinery, the world's largest educational and scientific computing society. We talk to researchers, practitioners, and innovators who are at the intersection of computing research and practice. They share their experiences, the lessons they've learned, and their own visions for the future of computing. Today's guests need very little introduction. Dave Patterson and John Hennessey are legends in the field of computer architecture. They recently won the ACM Turing Award for their
Starting point is 00:00:36 invaluable work impacting the microprocessor industry. Dave and John, welcome to ACM ByteCast. Thank you. Glad to be here. Thanks for inviting me. Great. So I'd like to lead with a simple question that I ask all my guests. If you could please introduce yourself and talk about what you currently do, and also give us some insight into what drew you into this field of work to start with. Okay, this is John. I'm back teaching at the university, running a new scholarship program. And in my spare time, I chair the board of directors of a small startup called Alphabet.
Starting point is 00:01:12 This is Dave Patterson. I'm in the subsidiary of that small organization of Alphabet. I'm working right now most of my time at Google. I was a professor at Berkeley for more than 40 years. I still work there part-time. And I also worked for the RISC-V Foundation. In terms of how I got into this, it goes back to, I think we'll have to go back to college. I was a math major and a math class was canceled. So I took a computer course. And like what Fred Brooks says in The Mythical Man once, I was just captured by the idea that the thoughts
Starting point is 00:01:45 in your mind could come alive, and I was completely hooked. And so that's how I got into it. I started, I did a little bit of computing in a computer club when I was in high school, and then got more and more interested as an undergraduate. And then microprocessors were just coming of age when I began my graduate work, and it looked like a good opportunity to begin to think about how software systems would be developed for this new generation of computing technology. You know, from that simple starting point, the way you describe it, to all that you've achieved, your work has been a result of years of research and experimentation. I'm sure it's been a very iterative process. Do you have or recall any breakthrough or aha moments in your career? What was your product market fit journey like? I don't know what that last phrase is. John is the person who's had a company. So I think the early aha moment was when we were doing
Starting point is 00:02:43 the RISC technologies at Berkeley and Stanford. And RISC stands for reduced instruction set computers. It was back in the 80s, instead of building really complicated instruction sets, which are the vocabulary that software speaks to hardware, we built really simple ones. We knew that at the time the clock rate would be relatively high, but the question is how many more of these simple instructions would we execute? And we actually took this toy program, the puzzle benchmark, and we actually calculated how many more instructions we had to execute than the more complicated vocabularies at the time. And the aha moment back then, which must have been right around 1980, well, it wasn't that bad. Maybe we had to execute 30 or 40% more even with our crummy compilers, but this could work. A high clock
Starting point is 00:03:31 rate, not that many more instructions. This might be a good idea. And so what I did is right after that, I wrote this paper that's kind of an op-ed piece called A Case for Reduced Instruction Set Computers. Because I thought, because I'd seen this result, this might really work. I think Dave and I both also had exposure to the primary way in which many computers and even mainframes were designed then using lots of microcode. And I think we both looked at it and said,
Starting point is 00:04:00 well, this machine is doing a lot of things at runtime that could be done at compile time with less overhead and more efficiency. So why not just do it then, simplify the instruction set? Why not make the micro instructions the instruction set rather than add an extra level of interpretation? And I think that was a great insight. Super. I mean, you speak about efficiency. At that time, when you saw the improvements that you did, was the concerns we seem to have today, as we've grown in terms of the usage that we have, seem to be around energy efficiency in the world of parallel computing.
Starting point is 00:04:38 I've heard the term performance per watt being bandied about quite a bit as a unit of measurement. What do you think of this unit? How have you seen the changes happen over a period of time? And is this what you try to optimize for when you were designing new systems today? I think a question you can ask with Morse law is how come you could keep doubling the number of transistors and the chip didn't burn up? And the answer was this thing called Dennard scaling, where he made this observation that every time you got more transistors, you could lower the threshold voltage. So actually the power stayed constant,
Starting point is 00:05:10 which was kind of amazing. So back when we were doing this, microprocessors were, I think there were five Watts and they stayed five Watts for a long time. And then they started going up fast. And so suddenly they started using up so much power that it ended up that Intel built a processor intended for laptops that was so hot you couldn't put in a laptop. So they had to stop what they were doing. They had to cancel that project. So since the end of Denard scaling, which has been at least a decade ago, this has become a primary limit to what you could pull in a microprocessor. It's the power you use, how efficiently is your design using the power that
Starting point is 00:05:51 you have available. So it's your power budget that limits you more than the transistors. Yeah, Rashmi, I think you actually said it right. It's about efficiency. When Dave and I started, the number of transistors you had available on a microprocessor was in the tens of thousands rather than the tens of millions or billions. And so you had to efficiently use those transistors. Today, power is what you have to use efficiently. But also with the slowdown in Moore's law, making sure you use transistors efficiently is going to become more important in the future. Got it. And it's interesting that you say that, you know, we're talking about laptops, but in the last 10 to 12 years, as the form factor of our devices have shrunk significantly, you know, you've had to think about smaller devices as well. So what
Starting point is 00:06:35 you're saying is very applicable in that space too. But in the similar vein for smaller devices, we've also had to think of memory architecture differently. We're suddenly storing so much data, much more than we ever have before. That puts pressure on our storage needs too. Could you elaborate a little bit on that and tell us where that field of research is headed? Are there new viable options that we should be considering? I think over John's and my career, there've been a bunch of claims of new exciting memory technologies, and they usually don't end up being commercially viable. An exception to that over my career, when John and I started, I think, memory was not made out of semiconductors. It was a magnetic memory. And this was what's called core memory. And you can still hear people refer to memory as core. That's like baby boomers or
Starting point is 00:07:20 something. But it's been semiconductor for a long time. But the backing storage that would store all the information in the world would be magnetic disks. They would still use magnetism to store information that they would spin. Well, a new technology that really made it was flash memory, which is a semiconductor technology. And so virtually, these laptops today don't have disks. They used to have disks. They used to make noise and spin. Now it's a semiconductor memory. So that technology has made tremendous advances. There are people who think there are even better technologies out there based on more exotic memory technologies
Starting point is 00:07:57 that be much better than Flash or are kind of the standard bearer that we execute programs at called DRAM. It's hard to know, based on our career, whether these technologies are going to be commercially viable a lot, but there's a lot of enthusiasm around them in the research community. Yeah. And just as Flash has changed what we can do and the amount we can store and really was the one technology that beat magnetic disks out. The question of whether these other technologies will be successful, I think a key insight here is density becomes the key limit. And as DRAMs begin to see their end of life and their slowdown, whether or not we'll have an alternative memory technology is going to be a critical question going forward.
Starting point is 00:08:44 I mean, are there popular names that get thrown around? I know I did a little bit of research, alternative memory technology is going to be a critical question going forward. I mean, are there popular names that get thrown around? I know I did a little bit of research, and I read a little bit about 3D NAND. Any thought about that? Yeah, well, Flash, one of the reasons Flash has been so successful is it's been able to go 3D, and I think that certainly helps. There's been discussions about various kinds of magnetoresistive or other approaches that could allow you to go to 3D. The problem is conventional DRAM technology can't go there. So whether or not one of these technologies makes it, I think the thing to understand from a computing viewpoint is the time between when a person who's a technology person builds one in the lab and the time that technology
Starting point is 00:09:27 is competitive in the market as a replacement technology is probably close to a decade. It just takes a long time to go from building one to building millions and billions of them at prices that are competitive. The place to look right now is in the laboratories and see what people are doing. You know, but along that same vein, John, you know, the way we're talking about it, just the process of doing this research and, you know, investing in something like this for 10 years, to me, sounds fairly expensive. So where is the innovation in this field coming from? Is it restricted to large organizations that have deep pockets? Where would you look for that innovation? Yeah, it is very expensive. And I think traditionally semiconductor companies have
Starting point is 00:10:08 devoted a lot of money to it. I think the big change that's occurred in the industry over the last five to 10 years is that you now have a number of big vertically oriented companies, Google, Facebook, Amazon, Microsoft, who are also playing in the basic technology arena, right, with quantum and other fundamental breakthrough technologies. And they're doing that because they see that as critical to their long-term success and their ability to innovate. Do startups have a chance at all in this? And how could they make this a viable option for them without the large amount of funding that they might need? I think they have a chance at it if you can find an
Starting point is 00:10:49 interesting technology, which might well be a university spin out initially. Universities are still doing lots of very fundamental technology work. You'll need several hundred million dollars even to get to the point where you can actually demonstrate a first product so it's challenging but moore's law and the importance of the integrated circuits industry till the entire information technology industry is so critical that if you really had a viable technology to either replace dram to get back to something like Dennard scaling, as Dave mentioned, you could find the money for it. Yeah.
Starting point is 00:11:28 One of the changes, we used to bemoan the fact that software people can get startups, but hardware people can't. That's no longer the case. This excitement around machine learning accelerators, I think there's 100 startup companies that are building their own version of that. So I think conventional wisdom has changed that there are opportunities for new companies around hardware. So far, it's been in this accelerator space. But I think if these are successful, I can imagine continued investment around more innovative memory technologies, too. Got it.
Starting point is 00:12:03 You know, I want to continue along that path, Dave. I mean, as I was preparing for this session, I listened to one of your lectures where you spoke about the proliferation of ML papers being the only count that follows Moore's law today. I thought that was funny. But, you know, indeed, the world is all about ML and AI from software job opportunities to startup ideas. You know, you had spoken about hardware-enabled AI. Would you care to elaborate a little bit on that? Well, I'm recording this right now at Google in the Brain Group, which is one of the leading groups in the world in this area. And these are the type of colleagues you want if you're a
Starting point is 00:12:36 computer designer. These are thirsty people. They want as many computer cycles as we can build. And the ironic thing about it is it's right as Moore's law is slowing down, where you can't just get the next processor from Intel and get this doubling or tripling performance by just giving them cash. Exactly when we need an explosion in computing to explore this exciting field of machine learning, computers are slowing down. So this presents a tremendous opportunity for people who are the age when
Starting point is 00:13:05 John and I did our risk work to invent brand new fields of how to build hardware that can execute machine learning well. One of the exciting things about it is that not only do we need new hardware ideas and compiler ideas, the field of machine learning is still changing really fast. So it's like you're trying to, you know, like a skeet shooting, except the skeets have rockets associated with it, right? It's really difficult to know exactly what the right thing to do. But like you said, if you read the newspapers, it's a really important thing to do. So these are really fun times.
Starting point is 00:13:43 Yeah, and I think it's somewhat reminiscent of an earlier time with the sort of explosion of microcomputer companies, companies that were taking microprocessors and building computers for a whole range of uses before the desktop became the primary motion. Or even earlier than that, the mini computer revolution. There were 16 mini computer companies in the US at one point. Now there are none. But for a long time, there were two or three big dominant players. And I think we'll see the same thing here.
Starting point is 00:14:13 Lots of creativity and a few of these companies in the machine learning space will become successful companies going forward. Definitely, you know, sort of looking forward to that. You know, but as we're talking about this, one of the areas that we haven't spoken about is really about security, right? When we have so much computing that's happening at the edge, we have so much data that we're collecting, security is a huge consideration, you know, especially in the case of like, when I think of security, I think about somebody hacking into my bank account to steal money. But now we're talking about, you of automotives, advancement in that technology, which is creating new risks.
Starting point is 00:14:51 You could take over my car and cause catastrophic damage. So could you speak a little bit to that in terms of what's happening in the world of security enabled by hardware? How much time you got? You've got all the time in the world. Go for it. I think I'll speak for John. Let's let him say it. But I would say, you know,
Starting point is 00:15:11 we've been in this field for four decades, right? I think security is one of the embarrassing parts of our field. There's so many things we can be proud of and how it's saved lives
Starting point is 00:15:20 and spread education and enabled people. But security is just embarrassingly how bad it is that we're enabling criminals all over the world to hold stores ransom. It's really embarrassing how bad it is. So I mentioned RISC-V, which is this open architecture, which is different. That means everybody in the world can work on it. And in the past, attempts at security,
Starting point is 00:15:46 because the instruction sets have been proprietary, have been limited kind of to those companies. They'll announce some new feature, kind of slowly roll it out. With RISC-V, everybody can work on it. And with field programmable gate arrays, you can have implementations that run slowly, but fast enough you can connect to the internet.
Starting point is 00:16:04 So I'm actually giving a keynote talk here in Silicon Valley to something called HOST, which is the Hardware for Security and Trust, I think it's called. And my keynote address is going to be, it's time for the security community to start creating things that are supposed to be secure rather than just complaining about other people's hardware as being insecure. So I think that community needs to shift to doing synthesis rather than what has largely been criticism. And I'm hoping that if we can innovate in the hardware and the software, we'll make some progress area where I think we don't even use all the technology that's been developed in the research labs around the university around the world. And I think we're still fighting to find that balance between building a system that is sufficiently secure, but which preserves ease of use, right? How many of us use the same password for many different accounts? And if we change it, if we use different password for every account, we inevitably forget some of them. And then we have this crazy way to get the password back, which isn't particularly secure because somebody
Starting point is 00:17:16 can probably figure out how to fish it in many cases. So there's a lot of work to be done here. I think the big change that's occurred is 20 years ago, you sent out spam or you sent out an unacceptable message. You became a social outcast. Now we get more spam than we get legitimate email. And people are actually out there trying to steal information, trying to steal accounts, trying to steal money. So I think we just have to double down and really find ways to get better security, but still make it easy to do the right
Starting point is 00:17:52 thing to have a secure system. Got it. And I think both of you touched upon one piece of, you know, information that I wanted to sort of dig deeper into, which is really about collaboration, you know, whether this is about collaboration between the, you know, security experts and the hardware experts who are actually building these systems, or between, you know, universities and, you know, academia and industry. How do we foster that? Do you have any ideas? I know we've tried in many ways in the past, in some cases, there's been a lot of success. In some cases, those conversations are still difficult to have. I'm very excited about the open star software is now mainstream you know when 20 years ago it was a kind of a backwater thing that maybe
Starting point is 00:18:31 universities funny people at universities used rather than companies around the world collaborate to build software that we all use every day you know it's a thing i brag about this is a positive of our community it sounds pollyannish but this is the real world. They found, there was a person who used to be at HP, and I think he said to run the HP operating system, he spent $300 million a year, but if he ran the Linux part, it cost him only $3 million a year to maintain. Companies spread that development cost across lots of companies, and then they share the software and work on it together. The RISC-V community is talking about bringing that same ethos to hardware. So I see this next decade,
Starting point is 00:19:14 plenty of opportunity for companies and researchers and people all over the world to collaborate around the open source movement, which is a mainline technology development. It already makes economic sense to collaborate with competitors and with universities. So I'm hopeful this opportunity will lead to tremendous collaboration. But I think the point about collaboration across various boundaries from a hardware design to software is the key one. After all, a system is only as secure as the weakest link. And so you have to think about security as an end-to-end function, which means you're talking not only to hardware designers and architects, but also to operating
Starting point is 00:19:58 system people, to people who are building metaware and through other levels. And we've seen the break-ins occur because they exploit one little difficulty in this large software stack that's become extremely complex. We need to think about how we get that, increase the security of those systems. Yeah, I agree with John. Like I said, I'm hopeful that the people who care about security, which are in all these disciplines, will start trying to create secure systems. It's easy to get depressed to think it's unsolvable. If you go to these meetings, it's like, well, this problem will never get solved, giving the legacy around a piece of hardware, a piece of software. So I'm hopeful that they'll start getting into the ideas of creating things that are secure,
Starting point is 00:20:43 using ideas like formal verification of kernels and formal verification of hardware and more exotic ideas. Like I think there were some people at Michigan that changed the instruction set encoding every 20 milliseconds and stuff like that. But just starting to experiment, put something out there, offer rewards for attacks and see if it survives. Otherwise, it's like global warming, right? It'll be like, well, this is a problem we ought to solve. It doesn't seem like we solve it. And it'll just keep getting worse. Oh, absolutely. You know, and we're talking about complex problems. We're talking about technology. I'd like to switch gears a little bit and also talk about technical careers related to what we were just talking about. You know,
Starting point is 00:21:20 both of you come from academia and are now deeply entrenched in industry, and you make that transition. I don't think that's right about John. I have a foot in both camps. All right. But we are also talking about complex systems and development that's going to get more specialized. I'm wondering if you have any advice for practitioners who are our, who are trying to stay abreast of the latest changes in their industry, you know, and we're also delivering against a business goal. How does one sort of future-proof their career and job? I'll let John take that. Well, I think you hit
Starting point is 00:21:57 the nail on the head. I mean, this is a field that continues to move quickly. None of any of us, if we went to school and didn't keep up with what the changes in the field that were occurring would fall behind. The good news is increasingly there's lots of ways to learn online now, lots of course material and lots of opportunities to learn different things. If I look at the Google course on machine learning, boy, I went through that course and it had great things to help you teach how to crash course in machine learning, boy, I went through that course and it had great things to help you teach how to crash course in machine learning. So I think there are things like that, that we ought to be doing. And it's hard. I mean, people work hard and then you tell them, well,
Starting point is 00:22:35 in addition to that, you've got to keep your skills up. But I think that's the reality of what it's like in this field right now. I do remember when I was a graduate student, my research contract had ended. So I started working at an aerospace company to support myself and my family while I finished my PhD. And I was struck how the people at that aerospace company just coasted on their knowledge. It was very different from the university. And I ended up working almost full-time there for a year or so. And then I went back to the university. It was like my mind was open to him. And these days, it's not like the IBM of the world where you can stay there for your whole career and they'll assure you a job.
Starting point is 00:23:14 That's long gone. So it's up to you to be on top of the latest technology to be marketable. And so you have to factor that in. The good news is, like John said, there's tremendous resources online. There are these massive open online courses where great universities all over the world make this stuff available pretty inexpensively, intended to be able to pick up in small doses at small time. It's on you to stay on top of technologies. If you're at a university, there's less excuse not to stay on top of
Starting point is 00:23:45 technology because everybody's trying to teach. But if you're going to have a long career, you need to learn about the latest things that are going on. Yeah, absolutely. But do you feel like some of the online courses that are available are more sort of geared towards software professionals? Or is it just my lens that I've only been looking for those? Probably true. But I think the massive, there's probably the MOOCs, so-called massive online courses. MOOCs are in just about every topic, I would think. I know there are architecture ones.
Starting point is 00:24:15 There are a lot of beginning courses, but there are also more advanced courses as well that are in new fields, whether it be security or machine learning. Machine learning, there's a lot of them in machine learning. And some of them have quite good material and lots of open exercises. Because of course, in this field, you just look at the lecture, the online lecture, you're not going to, you will not learn the material. You have to actually have some practice with it, right? Nobody learns programming by just reading a book about how to program or watching a lecture on how to program.
Starting point is 00:24:48 And it's the same thing, I think, in these other areas. Got it, yeah. I wanted to talk about one more thing, you know, as we're coming to the close of our conversation. You know, your contributions to the world of computing has been immense. And I'm sure some of that is attributed to the power of your partnership.
Starting point is 00:25:04 I mean, you guys finish each other's sentences, you know, you know, sort of how you mean this, this interview has been so simple because of how well you work together. How does one go about seeking a partner to work with? You know, what should you be looking for? How much of it is serendipity? Well, I can say that I was telling my wife how fortunate I was to find John and be working with him all these years. I think we both grew up on the opposite coast. I was kind of a surfer. He grew up in New York. Both families are large families of religious. We both went to public schools, inter-PTs, not the most prestigious schools in the United States. So we're kind of similar. I think we both have pretty good common sense. I think we make
Starting point is 00:25:45 good decisions. So even though we're, there would be no particular reason why we'd be pals. We just had a, we had the same worldview or something. And I was just really fortunate to find him to work with. So I would, I guess, look for people who you can get along with, who have, you know, seem to be smart people with interesting thoughts and, you know, are nice people. I certainly discourage people, you know, if somebody thinks they're the smartest person in the room, my advice is to run away because that's not going to be somebody you want to work with. I think we've had a great partnership. I think some of it was cemented early on when we were both working on the early ideas. They did not gain an audience readily in industry. In fact, quite the opposite. Many people in industry were quite antagonistic. So we found ourselves on the same side of the line. And then when we decided to begin writing the book, I think we,
Starting point is 00:26:45 you know, in our initial meetings, we found that we agreed on so many things. And I think one of the really magic things besides our personal relationship, we're able to help each other do better. That really means that we can work as a team and get something done that we couldn't do as individuals. And I think that's a remarkable opportunity and it's been a great experience. Yeah, it's made my career. And I would say, I think both of us are naturally collaborative. I don't think I've written very many solo authored papers. I don't think you have either, right? I mean, most of the stuff we do collaboratively. this is a team sport as far as I can tell.
Starting point is 00:27:25 I think early in the computing industry, one person can take on a problem and do it all by themselves. I think that's less true today. It helps to be able to work well with other people to be able to pull this off. We couldn't have written this book by ourselves. No, no. There's no way I could have possibly finished it. And it wouldn't be as good. it and it wouldn't be as good. Yeah. It wouldn't be as good. Yeah. Cause we would, the process was we took time off to work
Starting point is 00:27:50 on it, but, and we'd meet three times, twice, three times a week. And, uh, you know, we'd say, Hey, I got this idea. And what do you think of that? And, you know, we'd read through it and give feedback to each other on it and make it a lot better. And, you know, the ideas evolved quickly and, uh, without somebody to, you know, you would feel guilty if you didn't get the work done, because you're going to see the next day and then get really great suggestions how to make it better. I just read an article about this somewhere. They're talking about Jeff Dean and Sanjay Gimwa. And it was in the New Yorker. And in that article, there was a kind of a theory of how two people can be much more powerful than, you know, two times one. And that if you find
Starting point is 00:28:33 partnerships at that, they can do extraordinary things. The two people seemed like the knee of the curve of the most effectiveness renovation. I can't remember the, I think that's the right reference, but I don't remember exactly what their argument was. But they had a lot of examples of that. So if you can find that, that'd be great. Absolutely. I mean, the power of collaboration and the power of feedback, I think very valuable inputs. And I will definitely look for that article and put it into the show notes for our listeners so that they can actually go up and look at that.
Starting point is 00:29:00 It's called The Friendship That Made Google Huge. Awesome. Thank you so much at that. It's called The Friendship That Made Google Huge. Awesome. Thank you so much for that. So one other thing that I wanted to ask you is, you know, would you share with us anything that you may not have shared in another interview? I know you do a lot of interviews, but who are John and Dave when they're not computer scientists? Okay, well, I love sports.
Starting point is 00:29:23 I actually wrestled in high school and college and still, I think what makes me happy in a week versus, you know, brilliant idea or playing well at soccer, it's a tough tall. Really enjoyed that. I'm a big sports fan. I'm sitting here recording this a few days before the Superbowl. I'm sitting here wearing my 49ers jersey. I'm a big sports fan as well. Both of us are very family-oriented. I'm fortunate to live near my kids and grandkids and spend a lot of time with them. I've been married to my wife for 52 years. John's been married for a little while. Yeah. Not a real number like that.
Starting point is 00:30:08 I think Dave's love of sports is one axis. Mine is probably culture and travel and the arts. My wife and I are big arts fans. So, you know, we... But you read books, right? And books. And I'm a voracious reader a I'm a voracious reader. I'm a voracious reader. That's been my other passion.
Starting point is 00:30:30 But as Dave said, we're both fortunate enough to be grandparents. And it's the one redeeming thing about getting older. Well, you should mention John's book, too, about leadership matters. Is that the right leading matters Leading Matters. Leading Matters. Yeah. But in there, you can find about John's history, and he talks about how he's been such a successful leader. And to me, what I picked up was he learned a lot about how to avoid mistakes.
Starting point is 00:30:56 He was president of Stanford for 60 years without making a single misstep. And he learned that from reading the books of great people, which I think is really interesting. Yeah, agreed. It's much easier to learn from other people's mistakes than it is to have to make them yourself. Thank you so much for sharing that insight into who you are as people. It's very interesting to me and I'm sure to all of our listeners. For our final bite, I'd like both of you to answer this question, please, which is what is it that you're most excited about in the field of technology
Starting point is 00:31:29 maybe over the next five or 10 years? I think when people answer this question, if you think this is such a great idea, why aren't you working on it? Would be my final question. So I'm working on the thing I'm very excited about, which is both open instruction set architectures, bringing the innovations of open source software
Starting point is 00:31:46 to open source hardware around the RISC-V effort, and this how do we accelerate machine learning. It feels like machine learning is a revolutionary technology as big as the internet and the microprocessor. You can only tell later or the World Wide Web. I think it's one of those things, and we need to figure out how to build the hardware software systems that accelerate it. So I'm pretty sure for the next
Starting point is 00:32:10 five years, for sure, that's going to be an exciting topic to work on. And I believe it's how much I'm working on it. Yeah, I agree with Dave. And I think it's important to understand that this machine learning revolution certainly required people who did the fundamental work on algorithms and backpropagation and training methods and other inference methods. But it also required a tremendous amount of computing power to be harnessed. And I think the key to going forward is going to be to make sure that we not only have new algorithms and innovation in terms of the methods, but we also have the computing power that's going to be necessary to enable those. And with that, I think we're going to continue to be surprised by the rate of progress in this field.
Starting point is 00:32:53 I think every now and then, and the fascinating thing about computer science as a discipline is every now and then some part of the discipline just explodes in terms of the rate of progress. And I think that's what Dave and I saw in the 80s with the risk revolution and with what happened in microprocessors. And now it's happening in AI and machine learning. And that's really exciting to see as a field. Perfect. Thank you so much. I mean, so much to look forward to. And this has been an absolutely fascinating conversation.
Starting point is 00:33:22 Thank you, John and Dave, for taking the time to speak with us at ACM ByteCast. Thank you. Thanks for inviting us. ACM ByteCast is a production of the Association for Computing Machinery's Practitioners Board. To learn more about ACM and its activities, visit acm.org. For more information about this and other episodes, please visit our website at acm.org slash ByteCast. That's acm.org slash B-Y-T-E-C-A-S-T.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.