Behind The Tech with Kevin Scott - Judy Estrin: Co-creator in developing Internet – engineer, entrepreneur & executive

Episode Date: September 24, 2018

In this episode of Behind the Tech with Kevin Scott, we’ll hear from Judy Estrin, whose distinguished career in the technology industry includes work as an executive, engineer, and serial entreprene...ur. Besides serving on the boards at Disney and FedEx, she is former CTO of Cisco. Judy was one of the key figures in the development of the Internet, working with Vint Cerf on the initial TCP-project at Stanford. Tune in to find out to whom and what she attributes her success – it might surprise you!

Transcript
Discussion (0)
Starting point is 00:00:00 Right now, so much of what is going on in the industry is driven by the values of scale and speed. It is about maximizing growth. It is about maximizing connectivity. What about thinking about what we're losing in those values, there's nothing wrong with growth, but not at the expense of humanity. We're going to see the future. Hi, everyone. Welcome to Behind the Tech. I'm your host, Kevin Scott, Chief Technology Officer for Microsoft. In this podcast, we're going to get behind the tech. We'll talk with some of the people who've made our modern tech world possible and understand what motivated them to create what they did.
Starting point is 00:00:55 So join me to maybe learn a little bit about the history of computing and get a few behind the scenes insights into what's happening today. Stick around. Today, I'm joined by my colleague, Christina Warren. Christina is a Senior Cloud Developer Advocate at Microsoft. Welcome back, Christina. Hey, Kevin. Great to be here again. So, we are going to talk with Judy Estrin today,
Starting point is 00:01:26 and I have been super excited to get this interview on the books because Judy is truly one of the most amazing individuals in tech who I've ever met. Yeah, just seeing how far back her association with the internet goes and all the different roles that she's had as an engineer and as an entrepreneur. Yeah, her story is incredible. Like the fact that it starts with amazing anecdotes like her dad working for John von Neumann at the Institute for Advanced Study and sort of goes all the way through her direct involvement in the creation of the Internet protocols to illustrious career in the technology industry as an executive and board member. I mean, just incredible stuff. I'm really looking forward to hearing what you two discuss. Awesome. Well, thanks for chatting, Christina.
Starting point is 00:02:05 We'll catch up later in the show. Talk soon. Coming up next, Judy Estrin, internet pioneer, entrepreneur, tech executive, author of the book Closing the Innovation Gap, and one of the most incisive thinkers on the intersection of the internet, democracy, and humanity. So, welcome, Judy. It's a pleasure to be here. Awesome, awesome. I've been so looking forward to chatting with you. So I wanted to start with your childhood,
Starting point is 00:02:32 because you had this amazing career in technology, and your father is a very well-known computer scientist, so with you it started especially early, right? Yeah, I had the advantage of being a second-generation technologist when computers didn't really exist. But it wasn't just my father. It was my father and my mother. My father worked with von Neumann on the part of the team at the Institute for Advanced Sciences on the first computer. He and my mom were then invited to go to Israel to help lead the team that built the first computer in the Middle East, the White Sack. They then ultimately came back and came to
Starting point is 00:03:13 California, and my dad started the computer science department at UCLA and was there for years. My mother also has a PhD in electrical engineering, And when she got her PhD from Wisconsin, there was one other woman in the country who got a PhD in electrical engineering that year. Wow. At least that was the story in our family. And so... And when was that?
Starting point is 00:03:35 It was in the late 40s. An interesting story. When they came to UCLA, she could not work in the engineering department where my dad was because of nepotism rules. And so she ended up starting a data processing lab in the Brain Research Institute at UCLA. So she was one of the first biomedical engineers and really excelled and helped build that field. So I had two unbelievable role models and lots of positive aspects to that.
Starting point is 00:04:06 And I'm guessing there's some intimidation, too, as well. Some intimidation. I can still remember in high school when my dad brought home some videotapes to learn Fortran. And my dad ended up leaving me notes in Fortran. So he would leave me if-then-else statements of what I was supposed to do that day. And so it was my way to bond with my father that I ended up probably going into computer science. Well, let's talk about your mother a little bit more. So your dad is Gerald Estrin. Right.
Starting point is 00:04:37 And I think I know less about your mother, but it sounds like her career path in many ways might be the most extraordinary one. Did she share any stories with you about what it was like to get that PhD in electrical engineering at that point in time? And, like, she must have had this beautiful vantage point to sort of see how the field unfolded over the past many decades. Yeah, both of them were pioneers in their own right and in different ways. But one of the interesting things is that I remember when I was younger and people assume that my father probably got my mother into engineering. It actually was somewhat reversed. I just grew up in an environment where my parents were so equal in terms of their balance of career and passion. My mom was a very strong personality, very outspoken.
Starting point is 00:05:34 And so, yes, we heard a lot about how hard it was. Every step of the way, being a pioneer as a woman in the field, she was very involved in the IEEE. She was always running for office in different things. She often talked about being a woman in a man's world. She went on to also, though, be very, very involved in organizations that would help other women. And so whether it was the Society for Women Engineers or Grace Hopper started out as the Anita Borges Institute. She was a very strong advocate for exposing girls and women to engineering in the sciences
Starting point is 00:06:13 and providing opportunities for them. So it sounds like some of your first exposure to programming were these Fortran tapes and your dad's if-then-else instructions about chores and whatnot. Do you remember the first program you wrote and what it did? Well, I don't know if I remember the first program. I vividly remember. So at UCLA, now, let me back up a minute. For people who are exposed to the world today, most people who wrote their first program, you know, maybe it
Starting point is 00:06:47 goes back to having an Apple or some kind of, remember, there were no personal computers. So, in order to write a program, you needed access to a mainframe. So, there really wasn't an opportunity to write a program before being exposed to— Yeah, so like you had to get to a terminal room or like probably not even terminal room. You were doing punch cards. So you were still doing punch cards and batch submitting. One of my favorite stories, and I ended up using this in the book, is that I can still remember that I was working on something. I kept submitting these cards, and it kept coming back,
Starting point is 00:07:28 abend, abnormal ending. So, you know, the equivalent of crashing your computer today. And I'd been up all night doing it. I just came home in tears. I mean, I just was so frustrated. And my father sat me down and said, take a step back when you have a problem that you just are overwhelmed by. Try to figure out how to break it into pieces and how can you solve the individual pieces, but keep in mind how all the pieces fit together. That approach to problem solving stuck in my head through my whole life, and not just when it comes to writing a program, but how do you tackle really
Starting point is 00:08:14 complex problems, not going to the extreme. And frankly, we can talk about this if you want. One of my problems with Agile these days is the notion of break it into little pieces, but not keeping the architectural piece sometimes of knowing how they fit together. That balance of addressing complex issues by being able to figure out pieces that you can solve and the interconnections, that system level of thinking was something that my dad, I think it was so much about his approach to life and how he saw things. So what you were just talking about, this amazing advice that your dad gave you about decomposing a problem is something that has come up somewhat frequently in the conversations I have with other computer scientists recently. And there's this question of kids these days who are
Starting point is 00:09:12 being trained as software engineers are dealing with such higher level abstractions than the ones that we were dealing with when we were up and coming as programmers. And the question that I hear people asking over and over again is whether or not they're missing something by not being exposed to just this wide differential between the top-level understanding of the problem and just these absolute nitty-gritty atomic details of how to get a solution. And I don't know what the answer to the question is.
Starting point is 00:09:43 I know that I learned code on Apple IIe's and a RadioShack color computer too, which hooked up to a little 13-inch television and subsequently got the chance to muck around with manufacturing equipment that were controlled by PDP-8s and systems that had wire-wrapped backplane buses and whatnot. But I've never coded, you know, with punch cards. And, you know, like, I just sort of wonder, like, what did I miss by not having that? So, I don't think it's the punch cards that you miss. Meaning, I do think you're really onto something here. The punch cards are part of it, which has to do with a little bit of friction, a little bit of obstacle being put in. But let me start this off by saying, yes, we're missing something.
Starting point is 00:10:33 But it isn't that everybody needs to go back to doing that. It's that we need to recognize that there's a certain type of training that is not happening and figuring out how to be additive. The underlying thing I would say is we've lost system thinking. In the days when computer science was about building systems, building infrastructure, understanding how compilers work, how you talk to a computer, all of those things taught you a sense of learning of system thinking. And when I say system thinking, I mean two different things. One is how pieces are interconnected into a whole. So there's an architectural sense there, as opposed to looking at things as just individual points. And then the second thing
Starting point is 00:11:27 is thinking about the consequences of things. When you're building a system, you have inputs and then a complex set of interaction that may have different outputs and poking it in different ways that unless you're thinking about systems that are complicated, you don't think about consequences, intended and unintended consequences. And so those two things are really important. And I think that whereas the abstraction in some ways is so beneficial, because, you know, I remember writing assembly code for a problem programmer, you know, that wasn't necessarily a good use of how many hours it took to do it or the punch cards. But it seems to me we have three side effects. One is, do people know how to tackle hard problems?
Starting point is 00:12:20 Or do they only look for things that can be easily solved? And do we turn away or avoid or actually try to simplify the problem to something we can solve and then there are world problems that get delayed? And as I mentioned before, I really worry about the extreme of everybody take a little piece and not enough thinking about the architecture, how they fit together. Do you have a foundation that is not brittle? I worry about reactiveness. But there's a third thing, which is it is so easy to iterate on things. We all talk about iterating, iterating, iterating. So what happens is that you don't take as much time to think about what you're doing. And that's one of the things. The batch cards, it wasn't so easy to do a turnaround and try again.
Starting point is 00:13:25 And so you thought a little bit more about how you were fixing it before you submitted the cards. When you have very rapid iteration, on one hand, it is – we all know this. It's wonderful. You can fix problems quickly. You can do A-B testing. It also makes you sloppy in your thinking at times, or I shouldn't say sloppy, lazy. We don't have to put the same amount of thought into it. And if you're talking about failure that is not just inconvenient but can be harmful, you can't just say, oops, right?
Starting point is 00:14:07 So whether it is a self-driving car that hits somebody or a social media system whose implications is threatening democracy, we're not allowed to just say, oops. We have to be thinking about the consequences of the technologies that we also have not been training ourselves or training new engineers and computer scientists to think about those systems and consequences in the same way. Actually, the culture is in the other direction, which is move faster, get out a minimal viable product, get out your feature, and then we'll learn from our mistakes. I'm a big believer in learning from mistakes and learning from failure and taking risks. But we need to take a step back and say, what are those risks?
Starting point is 00:15:19 What are the consequences of the risks? And, you know, I couldn't agree with you more. I think there were a bunch of really powerful points you made there. Like, sometimes the struggle itself, you know, like having to sort of slow down and like think about something holistically really sharpens your thinking. And some of the most interesting breakthroughs happen that way. senior in high school taking my first real computer science class at my professor who made us write the solutions to our programming assignments down on paper before we typed them into the computer. And he could tell if you type the thing into the computer and done this like iterative debug cycle to try to get it right. And I didn't appreciate at the time like what slowing down was doing to help me be a better thinker about what it was that I was doing. where we're sort of more fragmented in the way that we do software development and we've atomized things up into a bunch of different pieces, which on the one hand, you've got to have some mechanism to deal with complexity,
Starting point is 00:16:32 but just because you're trying to solve a complex problem doesn't absolve you of the responsibility of having to think very carefully about the problem and its second-order effects. So I bring up these issues to say what we need to do is shift our values or remind ourselves of the values that drive that talent. Right now, so much of what is going on in the industry is driven by the values of scale and speed. It is about maximizing growth. It is about, even in some ways, which it sounds like a good value, maximizing connectivity. What about thinking about what we're losing in those values? There's nothing wrong with growth, but not at the expense of humanity, not at the expense of society. It might be nice
Starting point is 00:17:30 to think that maximum flat borderless connectivity should be a goal. But if you actually look at the way humans act and understand a little bit more about people, you might say, you know what? With connectivity needs to be some containment. Look what happens when we build fast in Houston and don't think about how waters move and a hurricane comes and the floods were way worse because we built for scale and we didn't pay attention to containment. When you're fighting a wildfire, you think about containment.
Starting point is 00:18:14 Well, misinformation spreads if you have maximum connectivity without thinking about where you need to contain because bad things can happen. So our industry is filled with so many talented, wonderful people. But I think sometimes we as leaders are pointing those people in a direction and setting values, which are not in the end changing the world for good. Yeah. So I want to dig in more on this whole notion of values, but through the lens of some of your early experience. So you were present at the literal creation of the modern internet.
Starting point is 00:18:54 So tell us a little bit about that story, like how after you graduated UCLA. So it goes back a little bit more in that, again, my dad was chairman of her time at the computer science department at UCLA. And actually, UCLA and Stanford compete for where the internet began because one of the initial ARPANET nodes was at UCLA. Paul Barron, who's the father of packet switching, was one of my dad's PhD students. And so I, you know, again, by osmosis, I'm being exposed to this. The other person who was one of my dad's PhD students was Vint Cerf, who is known as the father of the internet. So I ended up coming to Stanford doing my master's and Vint Cerf was my
Starting point is 00:19:39 advisor. And he was leading the team that was developing the initial TCP protocol spec. At that time, it wasn't TCP IP yet. It was just TCP Carl Sunshine, Yogan Dalal. And so there was no computer science department then. It was EE computer engineering. And I was one of three women in the master's program. They had already mapped out the initial spec for TCP. was one of three women in the master's program. They had already mapped out the initial spec for TCP,
Starting point is 00:20:13 and it was my job to do the testing of the initial implementations. I can still remember the basement lab that I used to go sometimes at 3 o'clock in the morning because the two sites we tested was BBN in Boston and the University of College London. So we don't think of it today because one of the beauties of the Internet is that you have asynchronous communication. But we would send packets, and then we would have a teletype machine in real time to say, did you get it? So we had to be on the same time frame. So it was a pretty exciting time in terms of being part of that. I will say that it was also my first exposure of people not accepting. They didn't really want a girl joining their group.
Starting point is 00:20:54 Not all of them. Not Yogan. But there were three unnamed people in that group that just made my life miserable for that year. And it was, I think, my first time of, oh, this is what my mom's been talking about all the time. It was a wonderful experience. While I was at Stanford, I built a local area network out of these three circuit board suitcases with Z80s. That was the very beginning of Z80 microprocessors. And so I built this local network. I ended up, when I left Stanford, my first job was at Zilog. And you were on the CPU design team for one of their microprocessors, right?
Starting point is 00:21:39 I was on the design team of the Z8 and Z8000. I was part of the software group. And, you know, now it sounds obvious, but in those days, it was very advanced. My boss and the people at Zilog decided it would be good for me to be part of the group to look at the instruction sets from a software developer's perspective. Nobody had ever done that before. And actually, I suggested an instruction in the Z8000. Which one? CompareString that nobody had ever put in a microprocessor. I was very lucky to be in an environment of that notion of it's not so much interdisciplinary, but the different perspectives and the strength of having software and hardware designers working together. This is one of the early examples of the abstraction, of how do you abstract and how you – and one of the benefits of doing it because it just made things like that so much easier.
Starting point is 00:22:44 Yeah, and now it's just incredible to think about where the abstractions are. You can, with a few lines of code, maybe no code at all, you can go to a cloud provider and click a few checkboxes and have a petabyte scale data storage system deployed planet wide that can do tens of millions of transactions per second and build your application on top of this. So on the one hand, I think the abstractions are absolutely beneficial. You don't want everybody all the way back at the dawn of time having to write assembly language for their apps. But being able to punch past those abstraction boundaries when you need to and to be able to think holistically about these systems, I think, is still just as important as ever.
Starting point is 00:23:26 We do realize that there's a huge amount of power in some of these systems that we are making, enabling people to use who don't understand the power of it. abstractions that maybe you shouldn't do. Because if you understand how a system works, you understand how it can go wrong. And then you're a little bit more careful in terms of how you deploy it. And we look at cars and how cars have evolved and where you can fix things where you can't. And let's take self-driving cars out for a minute. When we let somebody get behind the wheel and yield the power of a machine that can do wonders and do harm, we train them. They get a license. We could abstract brakes and putting your foot on the gas even more than we do. And we don't for a reason because that we unleash technology without people having to have an understanding of the consequences. And the stuff that's gone on about facial recognition, we really need to think about how we communicate those implications. to young engineers sometimes where you'll just sort of disregard a bunch of learning that other
Starting point is 00:25:07 people have done, especially if it's in the analog world and just sort of assume that you can do better with software. And, you know, sometimes like, you know, the control systems for cars, these things have evolved over the course of a century where people iterated and, you know, looked at the data and they understand that these things are safer when they're implemented in this particular way. I remember one of my bosses told me this story. So he was a power systems engineer by training, and he was telling me about Three Mile Island and that one of the reasons that they didn't notice the meltdown sooner was the monitoring systems were just too noisy. Like the operator was staring at this wide, wide panel that had dials and gauges and indicators and flashing lights and just sort of missed the one important one.
Starting point is 00:26:00 And I've always thought about that from an operations engineering perspective. You can monitor a large-scale distributed system and literally have millions of instruments out there sort of looking at everything. But if you're not able to surface the most important thing to folks who are going to have to fix things, then it's all sort of pointless. And we're having this conversation, I think, more urgently now around AI and data, you know, the facial recognition stuff and bias and data sets. I like the direction that the conversation is proceeding in, but I think there's still a lot more to be done. Right. So I'm not just saying this to flatter you and because it's your podcast, but my conversations with you on this over the last year that we've been talking about this have given me some hope. Because I think there are forces in the industry, some that are really just so focused on progress at all costs,
Starting point is 00:27:02 and others that are understanding of the technology, but are conscious of those trade-offs. And then there are a lot of people who can throw stones at all of this, but if you're not in the middle of it and you don't understand, look, I understand some, but I've been out of the sphere for enough time to know, I don't know, at a visceral level, the details of everything. And so, again, not just saying this because it's your podcast, but the role you play and being in a position to actually understand, make a difference, and having that conscious is just, it's heartwarming.
Starting point is 00:27:43 I appreciate you saying that. And I know that there are a bunch of other very thoughtful folks pushing against the problem as well at a bunch of companies. So I'm hopeful. We're in your hands, so please be. Yeah. We all need to look at this with the gravity that it deserves. It's just too important to be cavalier with. So you're at Zilog doing microprocessor software co-design, and then you become
Starting point is 00:28:12 an entrepreneur. Well, so I was at Zilog doing the microprocessor stuff. I was the project manager on something called ZNet, which was the first commercial local area network. So while at Zilog, I had the opportunity to be involved in the development and introduction of this local area network. And when is this? I think in 79, 79 or 80. So this is super early, like the state of the art for high performance computing then or like these big vector machines. Yeah, so we built a system.
Starting point is 00:28:44 Part of the reason this was a semiconductor company, a microprocessor company, why were we building a local area network? Zilog decided to go into the systems business around their microprocessors. And so what we did was develop a local area network to interconnect these systems. And our computing system was called an MCZ system,
Starting point is 00:29:08 which was an early PC, but it wasn't a PC. And we had an operating system that was called Rio, Rio I-O, which was a predecessor to the operating systems of PCs. But Zilog was a microprocessor company, not a systems company. And so one of the things I learned was up until then, I was an engineering elitist. I thought marketing people weren't important. But I learned a really important lesson, which is if you don't look at the needs of the market and match the technology, the technology is for naught. And even if you meet those needs, if you don't understand how to market and sell it, and you don't have distribution
Starting point is 00:29:50 channels that match up, you don't get anywhere, you don't solve any problems. So it's fine if what you're doing is research and discovery, and you're not trying to bring a product to market. But if you're trying to solve a problem, you need all these pieces. I very quickly, within those years at Zilog, moved from being an individual contributor to a first level and then second level. I moved into management pretty quickly. Did you enjoy the transition? Yes. And was it obvious to you that that was the right way to go? I never would have thought. I, as a kid, was not one of these people that thought I wanted to lead, thought I wanted to be an entrepreneur.
Starting point is 00:30:29 I didn't have a lemonade stand. I didn't do any of that stuff. And I found myself in this position, and what I realized very quickly is that I love people. I love people. I love collaborating. I love helping people and helping people develop and leading a team. And, you know, the fact of the matter is, I wasn't a great software engineer. I think what I was best at was the architecture and the thinking about the kind of systems thinking, I never wrote the fastest or best code. And so I ended up lucky that I had the opportunity. By going to Xilog, I also met the person who became my husband, now my ex-husband.
Starting point is 00:31:17 We started Bridge Communications, which was one of the early local area network providers. And so the name Bridge was about interconnecting networks. And it was interesting, our business plan when we started was going to be about interconnecting networks. And very quickly, we realized there are not enough networks to interconnect. And so we started off selling communication servers, which were connecting devices into networks, and then also interconnected those networks. That's awesome.
Starting point is 00:31:48 So presumably you ran engineering and technology at the startup. So you went from being a second-level manager to the buck stops with you. So how was that transition? So Bill was the CEO. I was the executive vice president in charge of engineering for the first couple of years. But, you know, early on, it felt very natural. And because Bill and I were partners, we kind of shared a lot of the responsibility. But the thing I really had to learn was how to make decisions, and not how to make decisions, but how to make decisions
Starting point is 00:32:26 in that kind of environment. And as an engineer who loves solving problems, I wanted to get it right. And sometimes you have to make a decision with the data you have, and you can't know that it's right. And it was a really interesting challenge for me. Bill had a very different style than me. He sometimes would say, go and fire that person or go pound on the table. And there are people who are very effective with an intimidating style. That isn't me. And if I had tried to be that, I wouldn't have been, I don't think, as successful because I wouldn't have been authentic to who I am. Yeah. You know, I think that's really interesting. Like, one of the things that I struggled with when I first became a manager, and I struggled with different flavors of this for a very,
Starting point is 00:33:22 very long time until I was managing hundreds of people. I just didn't understand that in leadership, many, many, many times, maybe more often than not, there isn't a right and a wrong. There's the best you can do at any particular point in time. That's particularly true around people. For a super long time, this is maybe the last big management lesson that I learned, is that I would keep people on the team who were toxic for the culture that I was trying to create just because by the numbers they were doing their job well. And giving myself permission to say, okay, this is my team. There's no right and wrong of it. This is the group of people that I want to surround myself with. And this is the culture
Starting point is 00:34:12 that we want to have in order to go solve these problems. That's sort of okay. But you just used an interesting example, which is there is no right or wrong decision often, but internally you're guided by what is right or wrong. So you used an example of not tolerating behavior because of performance based on you're driven by what your values in terms of what is right or wrong. I've seen leaders who use the excuse of there is no right or wrong decision to actually go in the opposite direction. And so I think that learning how to be able to make those decisions with a combination of your gut and your intellect, with a combination of data and compassion. It's the balance. It's why we're not machine learning algorithms, right? You know, we bring something different to it because there's a lot of nuance. I love what you just said. This whole notion of data plus compassion, I think, actually does lead
Starting point is 00:35:23 to the best decisions. Right. And as a hard thing to get pounded through the head of a computer scientist. Right. I like to say that you can have data-driven management, but you need human-driven leadership. Leadership is not data-driven. Managing is data-driven. And so you somehow have to combine the two of those. So at some point, like your career, you built
Starting point is 00:35:45 this successful business. You're this hugely successful tech executive. And one of the things that you and I have chatted about, you were CTO of Cisco for a while. There are very few people who have been CTOs of big tech companies. And I've gotten a bunch of good advice from you about how to do my job. So what was that transition like for you, you know, going from entrepreneur, like you've got the mission of the company, you're building the team, you know everybody, to holy crap, I'm CTO of Cisco. Yeah, fascinating. So Cisco bought Precept, and I became Cisco's chief technology officer. And that was in 1998. So I was there from 98 to 2000. The company
Starting point is 00:36:27 went from 18,000 to 36,000 people in that timeframe. I was CTO. I had legal M&A also reporting to me. It was fascinating to go from being an entrepreneur who was always fighting the scrappy company, fighting against big companies, to all of a sudden be at this company where everybody returned your phone call. Everybody wanted to see you. And I loved learning a different level of issues. It was also intensely frustrating for me because I was not building. I was the CTO. But you must have learned something in that role about managing at a distance and via influence because you had these incredibly successful and long board runs at Disney and at FedEx. So I was just about to say that I think it was the inverse in some ways.
Starting point is 00:37:22 I had been on the board of FedEx. I went on the board of FedEx in 1989. So I had been on the FedEx board for a long time, which gave me exposure to the issues of a big company, a different type of leadership. I went on the board of Disney the same year I became CTO at Cisco. And I had been on the board of Sun Microsystems for a while. So the board work gave me a sense of scale and innovation and the issues of that size company. It also gave me a sense of how to make an impact without direct line responsibility. And I think you had some really interesting moments in your board tenure. Yeah, that is true.
Starting point is 00:38:08 But I've got to tell you, the opportunity to serve on the FedEx and Disney boards is just phenomenal in terms of helping to build my understanding of a bunch of different things of innovation at scale, what it's like to use technology as opposed to being the developer of the technology, the media business, what I learned at Disney about the media business, and now that there's been a confluence between the technology and the media business.
Starting point is 00:38:38 It's a really important part of my sets of experiences. Awesome. Well, so one last thing before we let you go. I want to talk about some of the stuff that you've been doing more recently. So, and you and I have been having conversations about where both technology and technologists potentially can have impact, both positive and negative, on the public good. You're doing some really interesting thinking and collaborations there. So, tell us a little bit about that. So I'm spending a certain amount of time thinking and writing and collaborating in the unintended and intended consequences of today's business
Starting point is 00:39:27 model of the disruption of disinformation, of addiction to technology. There's a whole set of interrelated issues. My concern comes from often we want to, again, just look at a piece. Is it data privacy? Is it just election hackering? It's a lot of things. And so I've been working with a number of people, both in terms of writing and collaborating, but also some specific
Starting point is 00:39:56 things of, okay, what are some of the things we can do about this? And I really do believe this is similar to big oil or big pharma or big food, where you have an industry that is very successfully focused in an area, but there are consequences. And then what happens is there's opportunity in developing alternative energy. There's opportunity in developing healthy food. And the existing legacy companies have a choice. open up to provide all the wonderful things they provide without some of the harm, or they can deny climate and go in their hole and ultimately they'll get regulated, disrupted, hopefully in the climate case.
Starting point is 00:40:55 But I just think we need to be more aware of all of those consequences. So I'm spending a certain amount of time on that. Which is fantastic. And I'm really, really glad that you are engaged here because one of the real difficulties that I see is that these issues are, I think, unprecedented in terms of the complexity. There's just nothing in human experience that would let you develop an intuitive feel for what's happening under the hood of some of this technology. And it's not all sinister, right? I would put forth that most of this stuff, like the vast majority, especially built by tech companies, is like well-intentioned technology where we haven't thought about the second-order consequences
Starting point is 00:41:41 of what people are going to do with them and a variety of different things. But getting the public to be informed enough about how these systems work so that they can have a degree of agency, they understand what's going on both for themselves and for how they are pushing their own advocacy out in the world for how they would like things regulated and how they would like companies to better serve them. I think it's really, really important. It's tough because of complexity. It is tough. And I think that we just need to keep in mind, and you said it, it's not that people are bad, but we look at the incentives of organizations. When we were talking about making decisions, all of these decisions are made by who are you serving. And so it is easy in a capitalist for-profit environment that you serve shareholders who are pushing you for short-term returns to want to maximize growth.
Starting point is 00:42:44 It's not like you're trying to do harm, but you're measuring the things. We manage what we measure. And if we pick those metrics around things that are focused on short-term growth, short-term profits, it's not like we want to do harm, but we're not measuring the harm. And if you're not measuring the harm, then the people in the organization don't rally against it. And I am fortunate at this point in my life to have the understanding of technology at a broad level, but no longer being in the middle of an organization or set of organizations. So I see if I can bridge that gap without an agenda. So having an understanding of the for-profit world,
Starting point is 00:43:36 having an understanding of those incentives, understanding technology, but now being independent enough to actually be able to look at it and say, you know what, when I was doing that, I had that problem also, but now I can see that we need to think about how to change that because it's doing harm. And so I'm just really fortunate to be at a point where I can at least try to add my voice to the debate. And we're so glad that this is how you're choosing to spend your time. So thank you so much for being on the podcast today. It was my pleasure. Thank you for inviting me.
Starting point is 00:44:19 Well, thanks for joining us for Behind the Tech. What a great conversation we just had with Judy Ashton. Yeah, it was really great hearing the two of you talk and hearing about Judy's career and all the things that she's done and how she's literally been involved with the Internet from basically its inception through the present is so interesting. Yeah. Now, Judy has been an incredible mentor to me. Her story is inspiring. She has so much wisdom. She's such a good technologist. I've learned a ton from her over the years. I was really struck by one of the things
Starting point is 00:44:52 that she said in your conversation where she was talking about growth without the expense of humanity and how that seems easy to say, but seems in practice actually really difficult. I think there are a bunch of things that we can do. She was sort of dead on with this notion of we don't fix what we don't measure. And part of what I think we're learning right now with some of the things that are going on in the tech industry is how to measure some of the things where people are using the technology that we built in these sort of unanticipated bad ways. And I think we will learn very quickly as an industry how to get better on this stuff. But I think it really is going to require us to start thinking about our role as computer scientists and engineers and technologists a little bit differently. And for us to start when we are educating folks to make sure that that human part of the job is just as
Starting point is 00:45:53 well emphasized as the technological part. I went to a liberal arts school to get my computer science degree. And even there at a liberal arts school, things like taking an ethics class weren't mandatory. I wound up taking a philosophy class. We've sort of developed over the many centuries, this whole notion of a liberal arts education is important. What all of us have to like really understand, especially those of us who are in fields where we're building technology that has such a pervasive impact. Do you think having that background in humanities, do you think that helped you as you became a technologist and as you've transitioned into becoming a manager and now an executive? Yeah, I think it has an interesting, unanticipated ways. If you'd asked me the question
Starting point is 00:46:39 20 years ago, probably I would have said not so much. The obvious things I think that it helped with were with just writing, for instance, and being an effective communicator. The thing that I think is really useful, part of this, I think you sort of learn more from being a manager than you do from a liberal arts education is just having compassion for people. Once you put yourself in the mindset that you have to take care of people, it really does change how you make decisions. And I think, if anything, what we need to have a consistent understanding of in the technology industry is, like, that's our job. We have to take care of the people who are using our products. So, Judy talked about her transition from being an engineer and into management and entrepreneurship.
Starting point is 00:47:23 Did you relate to any of that at all? Does any of that kind of resonate with you and your transition from being an individual contributor to leading hundreds and thousands and more people? Yeah, no, totally. I think it's really interesting to see what a consistent experience that is for leaders. I think a lot of what Judy went through, like I went through as well, it's really challenging as an engineer to go from this mindset of like, okay, I'm solving problems that have a right and a wrong to like, okay, now I'm solving problems that don't have a right and wrong. And a bunch of constituencies who are asserting their right to be heard. And you have a bunch of stakeholders involved in the outcome of the decision. You got
Starting point is 00:48:00 to basically make calls on imperfect and incomplete data. And I think that struggle is something that every leader goes through at some point. And it is an interesting transition. And it makes you on many days wish that you just stayed an engineer. That's amazing. But I think what you're just talking about, again, goes back to kind of what I picked up as the thesis of you and Judy's discussion, which is all about thinking about and being aware of consequences as you're building things. Yeah. And given the complexity of the things that you're building, that is easier said than done, but increasingly necessary nonetheless. Well, I'm glad that people like Judy are working
Starting point is 00:48:39 to help other businesses and other people think about those things. And I'm glad that you're aware of those things and are doing that as well because we definitely need all the help we can get. Yeah, I'm super grateful for the folks like Judy who are choosing to use their time in this way to try to benefit everyone. Great conversation, Christina. Looking forward to chatting with you again on the next episode. I can't wait.
Starting point is 00:49:02 Next time on Behind the Tech, we'll talk with Danielle Feinberg. Danielle is Director of Photography at Pixar Studios. Her love of combining computers and art began when she was eight years old. This eventually led her to a BA in Computer Science from Harvard. Today, besides making films for Pixar, she mentors teenage girls, encouraging them to pursue code, math, and science. So be sure to tell your friends about our podcast, Behind the Tech. Hope to see you next time.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.