Big Technology Podcast - What the Ex-OpenAI Safety Employees Are Worried About — With William Saunders and Lawrence Lessig

Episode Date: July 3, 2024

William Saunders is an ex-OpenAI Superallignment team member. Lawrence Lessig is a professor of Law and Leadership at Harvard Law School. The two come on to discuss what's troubling ex-OpenAI safety ...team members. We discuss whether the Saudners' former team saw something secret and damning inside OpenAI, or whether it was a general cultural issue. And then, we talk about the 'Right to Warn' a policy that would give AI insiders a right to share concerning developments with third parties without fear of reprisal. Tune in for a revealing look into the eye of a storm brewing in the AI community. ---- You can subscribe to Big Technology Premium for 25% off at https://bit.ly/bigtechnology Enjoying Big Technology Podcast? Please rate us five stars ⭐⭐⭐⭐⭐ in your podcast app of choice. For weekly updates on the show, sign up for the pod newsletter on LinkedIn: https://www.linkedin.com/newsletters/6901970121829801984/ Questions? Feedback? Write to: bigtechnologypodcast@gmail.com

Transcript
Discussion (0)
Starting point is 00:00:00 An ex-open AI super alignment team member joins us to share his concerns about the company's trajectory, along with his lawyer, the Harvard Law Professor Lawrence Lessig, who will shed light on the lack of protections for those who speak out. All that and more is coming up right after this. Welcome to Big Technology Podcast, a show for cool-headed, nuanced conversation of the tech world and beyond. We have a great show for you today. We're finally going to speak with some of the people behind some of the concerns you've been hearing about the trajectory of open-eastern. AI, especially with regard to the alignment work within the company or really the super alignment work. So we're joined today by a former member of that super alignment team.
Starting point is 00:00:39 William Saunders is here. Welcome, William. Thanks for having me on. Thanks for being here. And it's my great pleasure to welcome Larry Lessig back to the show. He's a professor of law and leadership at Harvard Law School. And he's also representing William Pro Bono here as he goes. And I guess many of his colleagues as well as he goes and speaks out about these issues. Welcome, Larry. Great to be back.
Starting point is 00:01:00 Let's begin just talking a little bit about the vibe within Open AI. William, you left a few months ago. Take us a little bit inside the company so we can understand the environment from which you're coming out of. During my three years at Open AI, I would sometimes ask myself a question. Was the path that Open AI was on more like the Apollo program or more like the Titanic? And, you know, the Apollo program was about like carefully predicting, at assessing risks in doing groundbreaking science, building in enough safeguards to be able to successfully bring astronauts to the moon. And then even when big problems happened, like Apollo 13,
Starting point is 00:01:42 they had enough sort of like redundancy and were able to adapt to the situation in order to bring everyone back safely. Whereas the Titanic, you know, came out of this competitive race between companies to keep building bigger and bigger ships and sort of ships that were bigger than the regulations had been designed for. Lots of work went into making the ship safe and, you know, building watertight compartments so that they could say that it was unsinkable. But at the same time, there weren't enough life votes for everyone. And so when disaster struck, you know, a lot of people died. And OpenAI claimed that their mission was to build safe and beneficial AGI. And I thought that this would mean that they would prioritize, you know, putting safety first. But over time,
Starting point is 00:02:32 it started to really feel like the decisions being made by leadership were more like the White Star Line building the Titanic, prioritizing, getting out newer, shinier products, than, you know, really feeling like NASA during the days of the Apollo program. And I really didn't want to end up working on the Titanic of AI. And so that's why I resigned. It's kind of interesting that you use those examples and not the Manhattan Project, which is kind of the one that's been people have brought up, like the power and the destruction potential of nuclear energy has been something that's been talked about. And I don't know if Sam has compared himself to Oppenheimer, but I've definitely heard
Starting point is 00:03:12 some people make those comparisons. Why did you shy away from that? As long as we're going through the analogy lens, why did you shy away from that one? I think that is another valid analogy. Um, you know, I think this example of the Titanic, you know, makes it sort of clear that there were like, again, like a, uh, you know, safety decisions that could have been made better or something. Right. I think the Manhattan project is more the analogy for like the scope of impact that this technology could have. The companies are claiming this technology will have and are raising, you know, billions and billions of dollars.
Starting point is 00:03:55 based on this premise of the scope of impact. I think it's also a tale of sort of scientists who set out building a technology wanting to do something good in the world, right? The reason the Manhattan Project got started is because scientists looked at what was coming, what was possible, and were terrified that, you know, Adolf Hitler would get the bomb, right? And that this would be, you know, absolutely terrible for the world. And that's why they went to the Americans. Somewhere along the way, you know, at some point, like Hitler, you know, was dead.
Starting point is 00:04:32 Germany had surrendered, and yet the project went on. And again, that's another situation that, like, I would really not like to find myself in. Right. And I was going to ask you whether you think OpenAI is a product or a research company. And obviously, it's both. But the question is what leads? And reading between the lines or maybe just hearing you explicitly, your belief is it's a product company. Is that right?
Starting point is 00:04:53 It's a bit different from just trying to make the products that are most useful today. It's coupled with a vision of the research for how to build towards something called AGI or artificial general intelligence, which is building systems that are as smart as most humans and can do most economically valuable work that humans can do, aka can do most jobs of people. And so it's like the combination of these visions or something is something that I'm more concerned about. It's where that they're like building on a trajectory where they are going to do, like, AGI as stated, will be a tremendous change to the world. And so they're on this trajectory to like change the world. And yet when, you know, they release things, their priorities are more like a product company. And I think that is like what is most unsettling. You came out with this letter recently talking about how there needs to be a right for people within these companies to warn the public about some of the concerns they have.
Starting point is 00:06:07 Here's a quote from the letter. There is no effective government oversight of these corporations. Current and former employees are among the few people who can hold them accountable to the public. yet broad confidentiality agreements block us from voicing our concerns. So we're going to get into those agreements in a bit with Larry, but you obviously signed the agreement. You're here talking today. The main question I think the public has from you and the people who have signed this letter is, did you see something? Like, did you see something concerning enough inside the company to merit speaking out here? And if so, what was it?
Starting point is 00:06:44 To set something's clear. If there was, if there was a group of people, that I knew were being seriously harmed by this technology. First, I still really hope that open AI would like do the right thing and address this if this was like a very clear-cut case. I also personally would, you know, ignore any sort of like considerations of how I might be retaliated against, and I would talk about that plainly. So that's not what I was seeing. It's also, I don't think that I was working on the Titanic. I don't think that GPT4 was the Titanic.
Starting point is 00:07:19 I'm more am afraid that like GPT5 or GPT6 or GPT7 might be the Titanic in this analogy. And so I can talk about maybe like a couple of areas here. So one is the former colleague on the super alignment team, Leopold, Aschenbrenner, has like talked on a podcast about how he was like looking, you know, asking some questions about like how the internal security was working at the company. And then, you know, wrote a document containing some concerns. He shared this around. He then got reprimanded because, you know, some of the concerns offended some people. And, you know, personally, I would have like written it in a different way. And I think, but the disturbing part of this was that, you know, the only response was
Starting point is 00:08:21 reprimanding the person who raised these concerns, right? It might be reasonable to reprimand the person and then be like, okay, but these parts of these concerns, we're taking these seriously and we're going to address them. Right. That was not the response. And then he later talked about, like, this was one of the reasons that was offered for him being fired. So I guess what I'm narrowing, I'll let you do the other example in a second, but like what I'm narrowing down on is the people that have raised concerns within the super alignment team, it's not like they've all seen some powerful, dangerous technology that they don't believe open AI, like in the immediate term, they don't believe open AI is going to handle appropriately. It's more of like what is the path in the future and that's where this right to warn thing comes down. I'll say here at the beginning, we're going to get into right to warn. I'm fully in favor of this right to warn. And I'm really glad that you brought it up. But I think that it's important for the public just to establish that like inside open AI today, there's this group that's just left. It's not like there was this question like, did Ilius see something? Right. Did you guys see something? It's not that there's something immediate and harmful that you've seen. It's more of like
Starting point is 00:09:26 you're concerned about the path that this company can go down. Is that right? Yeah. Yeah. And I think, you know, this right to warn, this is a right to warn responsibly. This is not a right to like cause like unnecessary panic or something. And I am, most of my research was driven by, again, concerns along this trajectory that the companies are going along, have demonstrated progress on, and are raising billions of dollars to keep going along this trajectory. But I do think there could be like, there could be things happening today that we don't know about. And so the sort of scenario that I'm worried about happening today is suppose that there's some group of people that wants to, you know, spread a lot of disinformation on social media.
Starting point is 00:10:08 Let's say they want to manipulate an election or they want to, you know, incite violence against a minority ethnic group. And let's say that they're in a non-English speaking country. And so, you know, again, most of the people at OpenEye speak English. Most of the alignment work is done in English. And so, like, it would be, you know, somewhat harder for the company to notice this. Now, like, the models are, like, safety trained to, like, refuse requests to do things that are inappropriate, you know, or, like, that seemed like they might be harmful. Now, you know, Open AI has, like, caught actors, like, generating disinformation.
Starting point is 00:10:49 And so clearly they were able to either, like, you know, the safety training didn't apply or they were able to bypass it, right? There's this technique of jailbreaking where you, like, change how you frame the request in order to, like, bypass the safety limits. Maybe you just ask it in a different language or you tell some story around it. So now you've got a group of people. They have the ability to get the model to go along, generating lots of disinformation. And then, you know, the other line of defense that you might have here would be monitoring, the company looking at it. And I have, you know, concerns that, like, there might be a lot of ways that monitoring might miss things. Right. So, for example, like some systems that the company has talked about involve using a very, like, small and dumb language model to, like, monitor what a larger language model is doing, and this will, like, clearly mess a bunch of things.
Starting point is 00:11:45 You know, or there are, like, there might be ways to, like, send requests in that, like, go through some pathway that is just not subject to monitoring or, like, you know, the people can't look at what the actual requests and the completions are. just because the company just doesn't store that. And so now you might have a group of people like generating massive amounts of disinformation using like open AI products and you know the company wouldn't know about it. And it's like you know in this in this case if this happens in an English speaking country
Starting point is 00:12:19 you know somebody might notice and eventually tweet about it and the company would like find out through that pathway. Right. But if it's in a non-English speaking country right you know I don't know how big it could get before people would notice and I really would want a company, like, taking this kind of step to, you know, have, you know, this is a scary story, right? Tell me why this can't happen. And, like, you know, have somebody who is outside of the company who can, like, have an independent assessment that can say, like, yes, this can't happen, you know. And then I will, like, be able to rest easy.
Starting point is 00:12:54 But I can't. Yeah, and I want to get to Larry here because we should talk about the broader context here. which is that, first of all, believe it or not, William is the first Open AI employee that we've had on the show in four years. He's the first one expressing criticism of Open AI from within, or like from previously within, for four years. And recently we kind of had a, you know,
Starting point is 00:13:15 at least from the former employee standpoint, we figured out why. And that is because there are these broad nondisclosure agreements that Open AI employees have to sign before they leave. So I think we're going to talk about those first. And then we'll talk a little bit about this, I guess, new regulation or law that you are, you're both advocating for, which is basically a law that's going to allow employees to, or I don't know, a rule that will allow employees to whistleblow, even if there's nothing imminently illegal that's happening within AI. But let's talk about the NDAs first. So when people leave OpenAI, what are they forced to sign?
Starting point is 00:13:55 and why has that made it so difficult for people like William to speak out? Okay, but I want to actually first tag on to something William just said, which I guess I think is really important. Yeah, great. I first got into this space of whistleblowing protection, helping Francis Haugan, who was the Facebook whistleblower. And what William just described, of course, is what happened inside of Facebook with Myanmar Rwungi genocide,
Starting point is 00:14:23 which was basically the technology wasn't able to monitor the hate that was being spread by the government in that country, which led to, you know, tens of thousands of people being murdered. And while this is happening, people in the company are trying to raise the alarms and the company is not willing to devote the resources necessary to address the harm that they are able to demonstrate their company is committing. And the reason why that experience is relevant is it shows exactly why you can't rely on the company alone. When you've got a company in a deeply competitive market that's focused on a, you know, Facebook's had a single dimension,
Starting point is 00:15:05 which is like user engagement. Like, are we continuing to meet our target? And an employee, you know, there are a lot of great engineers in that company who raised concerns that were valid and serious, but if they're inconsistent with the objective of the company, it's not going to do anything about it. And so that's the structure. that's real in Silicon Valley, that you've got to build around.
Starting point is 00:15:28 And that's why what we're talking about, which we'll get to in a second, as you've said, guarantees that any concerns that are raised are not just raised to the company. They're raised to people outside the company who can do something about it. Now, as to what you're bound by, I got connected to this group, this incredible group of Open AI employees, ex-employees. when I read about the struggle that Daniel had gone through. Another ex-open AI employee who didn't sign the agreement. Yeah, who believed as he was leaving that by not signing an agreement,
Starting point is 00:16:04 he was giving up, as New York Times reported, something like $1.7 million in equity. And when I read that, I was like, wow, I mean, I don't know many people who would give up $1.7 million just for the freedom to speak. that's interesting. Like, what is it that you think you need to say? And when he raised that concern, and we began to talk to people in the circle of the company, very quickly the company realized that the agreements they were forcing people to sign were technically just not legal agreements in the state of California. Equity is wage in the state of California. If you earn equity, your vested equity, it's like your wages. And when you leave, they can't say,
Starting point is 00:16:49 oh, here's a bunch of other additional terms you must agree to in order to take what you've already earned. So the non-disbaragement part, any other additional obligations that were demanded were not actually obligations that could be enforced. And right now, the company's in the process of revising and putting together exit packages that are consistent with the law. And I'm optimistic. We're not, nothing settled yet, but I'm not. domestically you get to place that the company's rules are exactly right, that they, you know, they say you're leaving. Remember, you've got secrets. You can't share and don't. But, of course, you're allowed to share secrets with government investigators or people who are doing work with
Starting point is 00:17:38 the government for safety purposes. They're not trying to block any of that. And to the extent they do that, what they're doing is going to be consistent with the law. But we're in a transition right now. and it's not yet fully resolved exactly how much they've accomplished and how much they still needs to be done. It's interesting. So those non-disparagement that employees had to sign or else they could face their equity being clawed back seems like they're both like non-enforceable and potentially being revised, which is very interesting, good news.
Starting point is 00:18:07 I think OpenAI also says that they've never clawed back any equity nor did they ever intend to, but it's making people sign the agreement is strong enough. Yeah, I mean, but that's right, because, I mean, I've spoken to X, employees who've said, look, I've not done X, Y, and Z because I've feared the Clawbacks. So they can say they never enforced it. They didn't need to enforce it to have the effect, which it had for a significant number of people, especially when you've got, you know, people who are being very conscientious about the kind of obligation they're going to accept for themselves or not. And so if they sign something like that, they're going to live up to it. Totally. So it has an effect whether they enforced it or not, and that's the problem with the agreement. Right. And then, so that's once people leave. So the deeper question is what happens if people are inside the company and they see something they don't like and are they able to speak out? Because let me see if I get this right. In a normal whistleblower situation, let's say you're an Enron and you see the company committing tax fraud. That's obviously illegal. You'd be protected on whistleblower status.
Starting point is 00:19:14 statutes. But if you're within, let's say, an open AI and you find that the development of the technology is moving towards artificial general intelligence or superintelligence in a way that you find dangerous, you're not allowed to say anything because we don't have any laws against developing super intelligence. Yeah. So there's actually, so there's two things that go together that's very important in this context. So one thing you're right, there's not a lot of regulations. So there's not an FAA or an FDA sitting on top of the company that has imposed regulations that the company is either living up to or not living up to. But, you know, some agencies like the SEC takes the view that most anything could potentially be the sort of thing
Starting point is 00:20:01 you'd have privilege to complain to the SEC about because it could potentially affect the value of the company. And to the extent it's potentially affecting the value of the company, it raises SEC concerns. So if you you say you're following the following safety regime as the company in order to make sure AGI is safe and then you don't follow that regime. The SEC's view is you can come out and tell the SEC. You can whistleblow to the SEC and the SEC would consider whether that's something to act on. The problem is, this is the second part, engineers inside of companies like this or policy people inside of companies like this need to have confidence that the people they're
Starting point is 00:20:40 talking to know what the hell they're talking about. Right? So it's one thing to imagine, you know, an AI safety institute where you can imagine going to that and talking to people like you, people who have a really good sense of like what the risks are, what the technology is, and explaining here's why you think there's a concern. It's another thing to imagine like calling the SEC and telling the SEC, here are the seven safety related concerns that I have because you're very anxious that they understand it and are able to act on it in the appropriate way. And so that's why this is a kind of unique situation. It's both that there's not adequate regulation. So there's no regulator on the scene and that it's a technical field that doesn't easily open up to like non-technical lawyer types, the sorts that are going to be working at the SEC. And that's why, you know, when I spoke to the employees that I'm representing, it became clear that they wanted to kind of craft something that was different and new. And that's what the structure of the right to mourn is trying to produce.
Starting point is 00:21:45 So are you advocating for both a new regulatory agency and a rule to protect AI whistleblowers? What are you going to try to get at here? Well, my own view, and I won't speak for my clients here, my own view is, yeah, absolutely. There needs to be a regulatory agency that is overseeing this. I'm not sure what the structure of it is. It's kind of academic to talk about it, given the dysfunction of the federal government right now. But yes, other countries are building things like this, and we ought to be doing the same. And if there were such an agency, it itself would have lots of whistleblower protections built in,
Starting point is 00:22:20 and that would maybe obviate a significant chunk of the need for the rule. But the rule that we're talking about is a rule that initially we're trying to get companies to embrace. I think the most interesting part of the right to warn is the third point where it talks about creating a culture of criticism where the company says, look, we want you to criticize us. We want you to tell us what's going wrong. We want to encourage that. We're not going to punish that because that's the way we become the safest kind of company that we can be. And so that's really about the company itself creating that. And then the other part that I think is really critical is that the company says, we agree, if we create, we'll create this structure that says you can
Starting point is 00:23:07 complain to us and to a regulator and to an independent AI like a safety institute. You can do all three of those things, confidentially and anonymously. And if we do that, we expect you will use that channel. And if we don't do that, we acknowledge you can use whatever channel is necessary to make sure that these safety concerns are out there. But that's obviously designed to create a strong incentive for them to to build a channel for warning. But opening, I would say they already have that channel for warning.
Starting point is 00:23:42 So this is what they've said in the press reports. No, they don't. They say they have avenues for employees to express a concern, including an anonymous integrity hotline and a deployment safety board that they run products through. Right, but that's the company alone. Okay. So what I said is it has to be all three of those things together, right? So it's the company and the regulator and the AI Safety Institute.
Starting point is 00:24:05 So that, again, like we saw with Facebook, a lot of complaints were made to Facebook about the safety or the lack of safety of their product, and the company didn't do anything about it. And so the concern here is you need to have external review as well. And that's why the channel has got to be a channel that goes to three of these entities, so that we have some confidence that somebody is going to do something if there's something that has to be done. Well, I'm just curious from your perspective, do you think going to the SEC like Larry described is something that you or your colleagues would consider, given, you know, if there were things that you saw that didn't sort of hold to the safety protocols that opening I had lined out, or is that a non-starter?
Starting point is 00:24:46 Yeah. So what I would really want, if I went to whistleblow, is to, you know, have, like, somebody on the other end of the phone or the other end of the message line who I know really understands the technology. And I don't know to, you know, who at the SEC would be the person who would really understand the technology. I think that, like, a model that I think might work, you know, I think I personally think would work better would be like the model more proposed in California, Senate Bill 1047, where there would be like a, you know, where the law would create, like, you know, the office of the California Attorney General as a place where you could submit whistleblower complaints to.
Starting point is 00:25:30 and you could have, like, you know, if you have employees who understood the technology there and you could talk to them, you know, and ideally, this doesn't need to be, like, ideally this is not a high-stakes conversation. Ideally, you can just, like, call up somebody at the government and say, like, hey, I think this might be going, like, a little bit wrong. What do you think about it? And, like, talk to them, and they can gather the information. And then hopefully they say, like, okay, this isn't actually that bad, you know? And then you can, like, get on with your day. I think the thing to fight for here is like being able to really like, you know, talk about things before they become big problems. And in that circumstance, the SEC is insufficient.
Starting point is 00:26:13 Going to the SEC sounds like very intimidating. And, you know, it sounds like the sort of thing one would only do, you know, like, you know, it would be, again, it would be better to be like, you know, know, again, like, be able to talk to somebody in some agency who understands the technology and understands, you know, what the safety like system should look like. Great. Well, I have a few more questions that build off what some of the former colleagues of William have said within OpenAI. And then more about the nature of the company and where we might be heading.
Starting point is 00:26:52 So let's do that right after this. Hey, everyone. Let me tell you about the Hustle Daily Show, a podcast filled with business, tech news, and original stories to keep you in the loop on what's trending. More than 2 million professionals read The Hustle's daily email for its irreverent and informative takes on business and tech news. Now, they have a daily podcast called The Hustle Daily Show, where their team of writers break down the biggest business headlines in 15 minutes or less and explain why you should care about them. So, search for The Hustle Daily Show and your favorite podcast app, like the one
Starting point is 00:27:23 you're using right now. And we're back here on Big Technology Podcast with William Sounders. He's a former open AI super alignment team member now here with us expressing his concerns. William, I can't thank you enough for being here and being open about this stuff. And we're also here with Larry Lessig, the professor of law and leadership at Harvard Law School, also representing William and some of his former colleagues. So here's like a couple questions that have come up in discussions of this after you've gone public. So let me start with this one. So Jan Lika, who used to run the Open AI Super Alignment team, which you were on. He said that safety, culture, and processes have taken a backseat to shiny products within Open AI.
Starting point is 00:28:07 We've discussed that already here. So there's an argument that's being made online, and I'm just going to put it out there and would love to hear your thoughts on this, William. That basically the argument is that the group, the superalignment group didn't really see anything. and that the company doesn't really expect to see anything super dangerous for a while. And so it's reasonably putting like the 20% of compute that I was going to give to the super alignment team toward product until the time comes where it makes sense to shift that resources back to alignment work. What do you think about that? Again, I don't think the super alignment team saw like, you know, this is a catastrophe and it's like endangering people now. I think what we were seeing is a trajectory that the company is raising billions of dollars to go down
Starting point is 00:28:58 that leads to somewhere with predictable, unsolved technical problems. Like, how do you supervise something that's smarter than you? How do you make a model that can't be jailbroken to do whatever any unethical user wants it to do? And more fundamentally behind this, you know, how do we understand what's going on inside of these? language models, which is what I was working on, you know, for the second part of my career. And I was leading, you know, a team of four people doing this interpretability research. And like, we just fundamentally don't know how they, how they work inside, unlike, you know, any other technology known to man.
Starting point is 00:29:38 And, you know, there's a, there's a research community that is like trying to figure this out and we're making progress. But I'm, like, terrified that we're not going to make progress, you know, fast enough before we have something dangerous. And, you know, what people were talking about at the company in terms of timelines to something dangerous were, like, there were people talking, a lot of people talking about similar things to, like, the predictions of, like, Leopold Ashenbrenner, where it's, like, three years towards, like, you know, wildly transformative AGI. And so, I think, you know, when the company is, like, talking about this, I think that they have a duty to put in the work to prepare. for that. And when, you know, the super alignment team formed and the computer commitment was evade, you know, I thought that, like, maybe they were finally going to take that seriously.
Starting point is 00:30:31 And we could finally, like, get together and figure out the, like, you know, I could concentrate on the hard technical problems we're going to need to get right before we have something truly dangerous. But, you know, that's not what happened. Some people say that this, conversations like this are kind of doing open AI's marketing work for it. That basically, like, if this technology could potentially like level cities within a few years then like i don't know mackenzie is going to definitely get in there and try to contract with gpt4 what do you think about that conversation i certainly don't feel like what i'm saying here is doing marketing for open AI. I think, you know, we need to be able to have like a serious and sober conversation
Starting point is 00:31:24 about the risks. And risks are not certainties. There's a lot of uncertainty about what could happen. But when you are uncertain about what should happen, you should be preparing for worst case scenarios, right? The best time to prepare for COVID was not when like it had spread everywhere, but when you could start seeing it spreading and you could be like there's a significant chance that it will continue spreading. So this is for both you and Larry. So Joshua Aachiam, who's open AI employee currently, he sort of took issue with the letter on a couple of areas.
Starting point is 00:31:58 I'm just going to read from a tweet thread that he put out there. He said the disclosure of confidential information from Frontier Labs, however well-intentioned, can be outright dangerous. This letter asks for a policy that would, in effect, give safety staff carte blanche to make disclosures at will, based on their own judgment. And he says, I think this is obviously crazy. The letter didn't have to ask for a policy so arbitrarily broad and so underdefined,
Starting point is 00:32:22 something narrowly scoped around discussions of risk without confidential material would have been perfectly sufficient. What do you think about that? So what's interesting about that is I think it means that he didn't actually read the full agreement, right to warn that we were talking about. because the right to warn we were talking about actually talked about creating an incentive so that no confidential information would be released to the public. If they had this structure, imagine a portal again, where you can connect with the company
Starting point is 00:32:55 and with a regulator and with something like an AI safety institute together, the deal was that's what you would use and you wouldn't be putting any information out in the public. The only way that you would, the right to warn asks for recognition of the right to speak to the public if that is if that does not exist. So I, when I read that, I was like, wow, it's missing the most important part, which is an incentive to build something that doesn't require information as released to the public so long as there's adequate alternative channel for that information to flow. And what about this idea that getting something like this established might keep safety staff out of product meetings? Here's again from Joshua at GM. He says,
Starting point is 00:33:39 good luck getting product staff to add you to meetings and involve you in sensitive discussions if you hold up a flag that says, I will scuttle your launch or talk shit about it later if I feel morally obligated. I mean, that's like, I guess, sort of traditional Silicon Valley thinking, but I'm curious what you both think about that. This is not something that I want to achieve. This is a right that should be used responsibly. And so that if, you know, you're involved in decision-making and you feel like you don't, you disagree with the outcome, but you feel like a good faith process is followed. You know, you should be willing to, you should be willing to respect that. And I think nailing this down where it, you know,
Starting point is 00:34:27 getting the right balance of, you know, the legal rights on this is going to be tricky. And, you know, I want to get that right, but this is more like starting a conversation of where it should be. And I think that like, you know, again, I think on the other side, companies shouldn't have, you know, carte blanche to like declare any information about possible harms confidential. But yeah, it's going to be any implementation of this, you know, is going to get more detailed and more nuance trying to defend both, you know, the company's legitimate rights to, you know, confidential information that, like, preserves their competitiveness and also the, like, rights of employees to, you know, warn the public when something is going wrong.
Starting point is 00:35:16 The other thing is, I mean, even if there is the dynamic that you described, there's also so within a company, there's also a dynamic between companies. So if a company were to embrace the right to warn the way we've discussed it, there would be a lot of people like William or others who would say, that's the kind of company I want to work for. And so that company would achieve an advantage of talent that might swamp any cost that they're paying because they're being anxious about who they're sharing safety concerns with, number one. Number two, you know, again, inside of Facebook,
Starting point is 00:35:53 of course there were people inside of Facebook because I don't care what we're doing. I don't care what the world is suffering because of what are doing? You know, what do I care about 10,000 people dying in a country? I've never heard of it. Yeah, they're mostly not like that, though. they're mostly not like that. These are really smart, decent people who went to work for these companies because they're trying to make the world better, especially AI companies.
Starting point is 00:36:13 People who went to work for Open AI at the beginning didn't even have any conception of what open AI was going to be like today. The idea that it made the progress it did was a surprise to most people. So these are the very best motivated people that you could imagine. And I'm not worried that you're going to have a bunch of people who are like, I don't care what we do to the world. We're just trying to make sure our stock achieves its maximum return. Right. And so obviously this will take some buy-in from the top of companies. And I mean, this is a particularly interesting one with Sam Altman at the head of Open AI.
Starting point is 00:36:50 So Sam, you know, he has talked often about how he cares about AI safety. There have been some interesting quotes from him. Like early on, he's like, I think there's a good chance that AI is going to wipe us out. But in the meantime, there are a lot of companies that can make some money from it. I'm sure I'm misquoting him, but that was the spirit of the quote. And then, William, you spoke with the New York Times, I believe, talking about your view of Sam and oversight. You said, I do, I'm pretty sure this is you, I do think with Sam Altman in particular, he's very uncomfortable with oversight and accountability. I think it's telling that every group that maybe could provide oversight to him, including the board and the safety and security committee,
Starting point is 00:37:29 Sam Altman feels the need to personally be on and nobody can say no to him. So just curious, like, what your message would be to him and sort of what type of leader do you think he is in, you know, in this moment? You know, I think I don't recall the exact words, but I think Sam Altman has also said, like, no one should be trusted with, you know, this much power. I think he then went on to, like, say, like, oh, I don't think that's happening. But, you know, I think my message would really be like, you know if you want people to trust you like you should have real systems of accountability
Starting point is 00:38:10 and oversight that you know um and like not try to avoid that okay can i just one more question about like what it's like inside open ai because this is sort of like been the message that that we've gotten from you and some of your counterparts who have made these, you know, sort of declarations about what's going on. This idea of like that it's shiny products and safety and culture, safety takes a backseat. Like, how does that manifest internally when there are product launches and things like that? Like, how did you see that actually play out? I was mostly not in the part of the company that was like participating in product launches, I was doing this research to prepare for the problems that are like
Starting point is 00:38:58 coming down the road. But I think, you know, what that can look like is the difference between like, you know, we have a fixed launch date and we'll like rearrange everything to meet that versus, you know, when there's a like safety process, like testing how dangerous the systems are or like, you know, putting things together where there is like not enough time to do this before the launch date being willing to move it, you know. And it's, I do think, you know, now with like the GPT40 voice mode, the company did say that they were like, you know, pushing the launch back. But I think, you know, the, again, the real question here is are the people who are doing the safety work and doing the testing for dangerous capabilities,
Starting point is 00:39:57 like are they actually, you know, able to have the time and support to do their job before the launch? And, you know, I think a company can say that they're like pushing something back for safety, but still, like, not have all the work done by that time. Okay. Last question for both of you. So I think we've established that there's no, like, immediate term like threat to society or like let's say like titanic sinking style event that could happen with AI but what's the time frame that you think that these concerns might start to creep in given the trajectory of this technology we've talked a little bit today about how leopold believes like maybe within three years but i'm curious like yeah what the time frame is and then
Starting point is 00:40:44 is there is there like the like the frog boiling in the water problem where like this might only be become a problem when we've sort of become immune to it because we've heard so much about the dangers here, even as like chat GPT will hallucinate very basic details. Yeah. So I think like Leopold talks about some scenario of like you get AI systems that could sort of like be drop in like remote replacements for remote workers, do anything that you could get a remote worker to do. And then you could start applying this to like, you know, the development of more AI technology and like other science and that sort of thing happening within, you know, like the three-year time frame. And then this coming with, you know, a dramatic increase in the amount of risk that you could
Starting point is 00:41:30 have from either, like, you know, misuse. If anyone can hire an unethical biology PhD student, does it then make it a lot easier for, like, you know, nefarious groups to conduct, like, create biological weapons? Or also just do we like start putting these systems everywhere in our business? and decision-making roles, and then we've, like, put them, you know, in place in our society, and then, like, you know, a scenario that I think about is these systems become very good at deceiving and manipulating people in order to increase their own power relative to society at large and even, you know, the people who are, like, running these companies.
Starting point is 00:42:12 And I'm, you know, not as convinced about Leopold that this is necessarily going to come soon, But I think, you know, there's maybe like a 10% probability that this happens within three years. And then I think in this situation, it is unconscionable to race towards this without doing your best to prepare and get things right. Yeah, and I would add to that by just reflecting on the cultural difference between people who are in the business of setting up regulatory infrastructures to address safety concerns in general. and people who are in this industry. So when people in this industry are saying, look, between three and five years, it's probably a 10, maybe 20, maybe 30% chance,
Starting point is 00:43:00 we're going to have AGI like capabilities, and that's going to create all sorts of risks. In the safety culture world, you know, outside of these tech companies, three to five years is the time it takes just to even understand that there's a problem, right? So anybody who expects you're going to set up an infrastructure of safety regulation in three to five years just doesn't understand how Washington or the real world works, right? So this is why I feel anxious about this. It's not that I'm worried that
Starting point is 00:43:30 in three to five years everything's going to blow up. It's just that I'm convinced that it takes 10 years to get to a place that we have an infrastructure of regulation that we can count on. And if we're talking about 10 years, what is the real estimate of this technology manifesting these very dangerous characteristics seems to be from what people on the inside are saying pretty significant. So that's why even if it's not a problem today or tomorrow or next year or the year after, we have to, you know, it's a huge aircraft carrier. We've got to turn and it takes a long time to get it to turn. And that work has got to begin today. And I'll just add that, you know, in the real world with the Titanic, right, you didn't have a regulation
Starting point is 00:44:18 that guaranteed that you have enough lifeboats until the Titanic actually sunk, right? And I am on the side of we should have regulation before the Titanic sinks. I mean, man, all that money to get on that boat. And then no life jacket. Seems brutal. All right. William, thank you so much for coming here, spending the time addressing some of the criticisms and being forthcoming about what your concerns are. I mean, hearing from you after you spent some time when the inside has been illuminating to me, and I think it will be for our listeners as well. So thanks so much for coming on. Thank you. And Larry, always great speaking with you. Thank you for bringing such great analysis to the show every time you're on. And I hope
Starting point is 00:44:59 we can speak again soon. Every time you ask. Thanks for having me. Okay. Thanks so much. All right, everybody. Thanks so much for listening. We'll be back on Friday breaking down in the week's news with Ron John Roy. Until then, hope you take care. And we'll see. See you next time on Big Technology Podcasts.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.