The Joy of Why - Will AI Ever Have Common Sense?

Episode Date: July 18, 2024

Common sense rules our world. This fundamental, sometimes trivial knowledge is inherent to how humans interpret language. Yet, some of these simple human truths are so obvious that they're ra...rely put into words. And without the data of common sense to train on, large language models such as ChatGPT have bizarre, often humorous blind spots.Yejin Choi, professor and the chair of computer science at the University of Washington, calls common sense the “dark matter” of intelligence. In this week’s episode of “The Joy of Why,” Choi talks with co-host Steven Strogatz about decoding the interstitial glue of language and comprehension. Together, they explore the question: Should we program more humanity into the next generation of artificial intelligence?

Transcript
Discussion (0)
Starting point is 00:00:00 Hello, I'm Brian Cox. I'm Robin Ince, and this is the Infinite Monkey Cage trailer for our brand new series. We've got mummies, we've got magic, we've got asteroids. Mummies, magic and asteroids. What's the link? That it was an asteroid that magically went over the world that led to Imhotep the mummy coming back to life? That's correct. I thought it would be weird, scientific as ever. But the most important thing to know is that we are going to deal with the biggest scientific question we finally ask.
Starting point is 00:00:25 What is better, cats or dogs? Listen wherever you get your podcasts. Going back to at least the 1960s, computer scientists have been dreaming about brain-inspired computers that might someday demonstrate human-like intelligence. With the rise of the internet, the availability of enormous textual datasets, and impressive advances in computational power, we've reached a momentous point. Large language models, or LLMs, can often seem to wield something close to human intelligence, at least to us non-experts.
Starting point is 00:01:10 And the release of OpenAI's ChatGPT in 2022 helped these LLMs make their mark in the headlines, in the workplace, and in dinner table conversations. But there's still a telltale sign that large language model intelligence is truly artificial. Their lack of common sense, which can emerge in spectacular and sometimes hilarious ways in the mistakes they make. I'm Steve Strogatz, and this is The Joy of Why, a podcast from Quantum Magazine, where I take turns at the mic with my co-host, Jana Levin, exploring some of the biggest questions in math and science today. In this episode, we're going to be speaking to computer scientist Yejin Choi
Starting point is 00:01:51 about the architecture and capabilities of large language models and speculate about whether artificial intelligence, AI, will ever gain common sense. Yejin Choi is a professor and the chair of computer science at the University of Washington, where she researches statistical approaches and computational models for natural language processing. She was recognized as a 2022 MacArthur Fellow and named one of Time Magazine's 100 Most Influential People in AI in 2023. Yejin, thanks so much for joining us here on The Joy of Why. Thank you for having me. I'm excited to be here. Great. Well, this is going to be so much fun. I'm so fascinated, as I know so many people are, by what's going on these days in AI. And so
Starting point is 00:02:40 while I was preparing for this chat, just as a little bit of a joke, I was curious. I asked ChatGPT, a chatbot built on one of these large language models, do you understand? And it replied, as an AI, I don't possess consciousness or subjective understanding in the way that humans do. While I can generate responses that seem like understanding, it's important to recognize that this understanding is mechanistic and based on statistical patterns rather than true comprehension. What do you think about that answer from ChatGPT? It sounds like someone coached ChatGPT to say the right thing. That's funny.
Starting point is 00:03:21 That it was coached? You don't think it would come up with that on its own? We don't think it would come up with that on its own? of particular style of language, lawyer-like language that ChatGPT distinctly use that's provided by humans as good examples to follow. So it is not the case that ChatGPT, after reading the raw internet, suddenly speaks like that. It's because of the post-training that does coach ChatG GPT to speak in a more safe, more politically correct and careful ways. I get it. Sure. No, that makes sense. There's so much at stake for these companies. And, of course, people could really be harmed, I suppose, if they take some of these responses too literally.
Starting point is 00:04:22 I noticed that all the different chatbots have disclaimers now on the front page that these can make mistakes, you need to check them. Yeah. I mean, clearly the lawyers have had their, as you say, coaching sessions with these bots. Totally. Okay. But so maybe this is a good point for us to back out just for a second, to say that these large language models, which most of us know through products like ChatGPT or Google's Gemini or Microsoft Copilot or Anthropix, Claude, these are just one type of artificial intelligence. And so since our listeners may have been hearing words, as I think we all have, like machine learning, large language models, AI, neural networks. Could you help us just
Starting point is 00:05:07 distinguish, are some of those subsets of the others? How should we think about those kinds of terms? So these are all fairly broad terms that definitely have their own distinct definitions, but they overlap a lot. So machine learning in general is about an algorithm that teaches machine to learn some patterns between input-output pairs. And then artificial intelligence is perhaps arguably more broadly about computational forms of intelligence that can do certain operations, but that could be using machine learning or it could be using just algorithms, inference algorithms. So neural network is just like one type of machine learning algorithm that is currently the most popular probably. So for example, the computational chess players in the earlier forms of AI, they were just inference algorithms, whereas the more modern
Starting point is 00:06:06 version of that would be in the forms of a neural network that is a form of machine learning. Good. That's very helpful. And so what are some of the capabilities and also limitations of these large language models? So the capabilities of these large language models are phenomenal. It's really beyond what the scientists have anticipated. Just anything that you could provide as textual input, it turns out these large language models can do quite well in answering them,
Starting point is 00:06:42 even if it requires open-ended answers. Long text input output turns out to be really, really great. So not only can it do simple reasoning, multiple choice question type, question answers, but also any topic that you throw at ChatGPT, it will be able to answer strikingly well. The truth is, though, it's really a reflection of human intelligence that is shared on the internet. And internet is vast. Humans typically are not really aware of how vast that is because humans have limited capabilities in digesting and reading what's out there. But the machine uniquely can really read them all, literally.
Starting point is 00:07:27 So then it can mimic the sort of knowledge and wisdom that people shared online, and then in some sense, read them back to you, but not in the verbatim sense, but in a more rephrased sense. So it's not like the exact copy of what it has written from the internet, but it's able to rephrase, it's able to synthesize, so that it sounds new enough for people. That is, as you say, phenomenal. One of the most remarkable things.
Starting point is 00:07:57 You can ask it to generate college application essays or help you write a Python program. I mean, I tried one time just for fun to have it write a Saturday Night Live skit where Donald Trump plays a psychiatrist. And he tries to, you know, give advice to his patient but in the style of Donald Trump. And it was really funny and did something that sounded a lot like Trump. And I can't imagine that, you know, as you say, that's not verbatim on the internet. I don't think Trump has ever been playing a psychiatrist.
Starting point is 00:08:33 But it's amazing the synthesis that it can pull off. So some people have called it sort of spicy autocomplete, what these things do. But maybe you should tell us, why would anyone say that? What is it that these large language models are doing fundamentally? It's able to read a lot of text and learn to predict which word comes next. During training, literally, all that it does is trying to predict which word comes next. But at an extreme scale. to predict which word comes next, but at an extreme scale. For that reason, some people just diminuate large language models as spice auto completion
Starting point is 00:09:12 machine. The reason why it's not necessarily doing verbal team regurgitation of the training data, though, is because of the particularities of the technical details being used under the hood, which is not necessarily about memorization. It's also able to do some degrees of generalization and also there's a randomness in the way that this text is generated out of the learned neural network. And that randomness causes how the text is not necessarily verbatim regurgitation always. But I mean, sometimes it can be, by the way, if the text was repeated often enough on the internet data, then it's going to actually verbatim memorize that. And, you know, there was some incidents that New York Times reported that it was able to
Starting point is 00:10:06 regurgitate some of the past New York Times articles. Oh, really? I haven't heard about that. I see. So it can plagiarize in that sense. I mean, one could say that it's plagiarized. Another person could say that, well, this is neural network being able to retrieve what it has read.
Starting point is 00:10:22 But regardless, because of that, some people diminuate these motions as, oh, it's just like auto-completion. But the reason why it's able to do something striking like, you know, discussing some topics in a Trump style that Trump may have never done before, that's possible because these motions are capable of interpolation between two data points. The novel interpolation that nobody has ever done before is trivial for these machines. So you do get that kind of a novelty to some degree as well. Well, so you've been mentioning training, and I think it would be great if you could explain to us a little bit in detail,
Starting point is 00:11:04 what does that really mean? How would you train a large language model? Or how are these big companies that have built ChatGPT or Gemini, what do they do to train their models? So basically, the training boils down to building extremely large neural network that has layers and layers and layers of neurons piled up and then feed internet data in sequence. And the goal of this learning process is to predict which word comes next, conditioning on the sequence of previous words. And what is striking is that that simple recipe of training neural network can lead to such powerful artifacts that can do all sorts of question answering in text that comes across as a striking level of artificial intelligence for many people. But importantly, that kind of training is really, really different from how humans learn about the world, which we don't really know how humans really learn. However, it's reasonable to suspect that the humans don't necessarily try to predict which world comes next,
Starting point is 00:12:19 but we rather try to focus on making sense of the world. So we tend to abstract away immediately. You and I, by the way, are not able to remember the discussions, the interactions, the conversation. We just had verbal team. We just cannot. Because our brain is trained to abstract immediately. But we do remember the gist of our conversation so far, such that if you ask me the same question again, I'll be surprised. So there's something about the way that humans learn. And also humans learn with curriculum and curiosity, and we make hypothesis about the world. And then
Starting point is 00:13:00 if something doesn't make sense, even children, even babies, they try to do some experiments to figure out their confusion points about simple objects, the physical knowledge about the objects that they interact with. But machines, from day one, they're fed with the New York Times articles. New York Times articles. And they don't have any saying in what order they're going to read this text, nor do they have any saying about, oh, wait a minute, I really want to read something again. There was something really nice and curious about this particular, say, Hemingway's book that caught my attention that I want to read slower. The way that learning happens is so different, and it's quite striking how you can bypass the normal way of learning of humans and then still produce something that speaks human language so well. You've raised so many interesting points there.
Starting point is 00:13:57 For instance, when talking about babies or people at any age, really, that we have curiosity, we have desires, like we want to read or reread that Hemingway passage, or maybe there's something we don't like reading and we would like to skip it. So far, we have talked about artificial intelligence, but we haven't talked about artificial emotion, right? Like the fact that these machines so far were not really putting in desires, it seems like that might limit what they can do, given that babies and people of all ages have willpower, they have desire, they have things they wish. Do you think that emotion is a big part in human learning that these machines are missing?
Starting point is 00:14:41 Yeah, that's a great point. And in fact, it boils down to the fact that we are bio-beings at the end of the day. We have desire, like a deep sense of self-identity that really makes who we are. And it's not something we can change. We are born with this individual identity and then we live with it. We live once and then we live with it. Whereas AI, it's not clear what it really is because it just read everybody's writing in some sense and it became some average viewpoint or thoughts and emotional soup
Starting point is 00:15:22 that does mimic all this human emotion and intent due to the human emotion and intent that humans put into their writing. So then these machines are capable of mimicking all that. But you're right that at the end of the day, it doesn't really have the kind of genuine emotion that humans have. Now, whether that's a bad thing or good thing, that's a philosophical question. That may be even a scientific question in terms of safety. Is it a good thing if AI really, really develops its own emotion such that, you know,
Starting point is 00:16:01 it has like survivor instinct or it wants to dominate the world. Is that a good thing or not? Well, this is, of course, something that all of us are thinking about nowadays. And we should probably save that question for a little bit later, because I wanted to ask a little bit more about the training, because it is, as you say, such an inhuman thing that we ask of them. We ask them to predict the next word given a passage. And what happens when
Starting point is 00:16:25 they get it wrong? The way that it's trained is it's supposed to maximize the probability score that it should assign to the correct word. If it predicted the wrong word, it means the wrong word got the higher probability, but all that there is is to raise the probability of the correct word that that should have happened instead. But not by directly dealing with that word, right? I mean, I guess what I'm going for is the idea of the weights in a neural network and how there are mechanisms to change the weights. Right.
Starting point is 00:17:00 So perhaps I should pedal back a little bit and then talk about that there are two phases of training. One is pre-training. The second is post-training, also known as reinforcement learning with human feedback. That's a jargon, which actually is not just reinforcement learning, but also mixed with something known as sequential fine-tuning or supervised learning. But anyway, by and large, there's pre-training and post-training. During pre-training, the learning mechanism is basically to maximize the probability score that will be assigned to the correct sequence of words, meaning the exact sequence of words that happen to be on the internet. Now, there's really no reason why that's the only correct sequence of words, by the way, because for any given prefix text, there can be another different word that could be okay to say. So the notion of correctness is not quite right. But anyway, so the neural networks are trained to maximize the probability.
Starting point is 00:18:05 And what does that entail in terms of the weights that these neural networks learn is that the way that these machines are learned is basically based on what's known as backpropagation. We basically take gradients of individual weights, partial gradients. So you take the derivative, a partial derivative of every weight of neural network. There are just so many of the weights, by the way, hundreds of billions of parameters that you take partial derivative of, and then you move that weight so that it's going to increase the probability score assigned to the particular sequence of words that were in the training data? Well, as a mathematician, I'm very happy to hear you talking about partial derivatives,
Starting point is 00:18:52 but some of our listeners may not feel the same way. So let me try an analogy. So I like to play tennis, and I remember when I was learning tennis, sometimes, you know, the ball comes to me and I might hit a bad shot. And then the tennis coach says, you need to get your racket back earlier. You know, you weren't prepared. So then I make an adjustment to I don't know what. I want to say weights. Like I have some kind of internal representation of how important is it that my feet are in the right place or that I've turned my body sideways or I got my racket back or I keep my eye on the ball. I have all these different weights that I have to pay attention to. And given that this shot was bad,
Starting point is 00:19:38 so to speak, I'm going to try to adjust my weight so that I'll do it better the next time. It's something like that, right? Yeah, yeah, yeah. That's a really great analogy. Okay. All right. But it's a brutal way to learn that you just make this poor machine take one test question after another, and every time it gets it wrong, you punish it, so to speak, or at least you gently correct it by making it adjust its weight so that it would do better the next time. Yeah. On and on and on. On and on and on.
Starting point is 00:20:07 It's a very brutal training. It's good that motion doesn't have emotion. Right. So it doesn't care, I guess, as far as we know. But this is the pre-training, you say. Yeah, yeah. And then during post-training, there are multiple things that can happen. But maybe let me highlight just the most representative one, which is reinforcement learning with human feedback.
Starting point is 00:20:27 So in this particular type of post-training phase, what happens is that you present the machine's answer to a query to human evaluator, and humans can give a thumbs up, thumbs down. and humans can give a thumbs up, thumbs down. And then based on that, you then go back to the neural network to adjust the weight a little bit using the analogy that you used before. But this time, instead of focusing on which word comes next, you're focusing on whether you get thumbs up or thumbs down by the human evaluator. Okay. I see. So all this process, I mean, we know that computers run fast, but still, how long would it take to do, say, the pre-training phase or just give us a
Starting point is 00:21:13 ballpark feeling for it? Are we talking about days, weeks? Yeah. I mean, it varies a lot depending on how much data that you are going to feed versus how much of a compute you have versus how large neural network you want to train. So there are many variables at play that determines how long it takes. And also, by the way, the tech companies don't share exactly how long it took and how much of a compute they used, but one can speculate. I would say the really good ones typically takes like a couple months if you want to push the limit.
Starting point is 00:21:49 But, I mean, if you want to stop earlier with a smaller amount of data, then it could be just a matter of a couple days. By the way, if you really think about how long humans learn, 10 years of learning as a human baby, you know, becoming a child, they still have a lot more to learn. So in some sense, a couple months isn't so bad. That is true. We'll be right back.
Starting point is 00:22:18 Welcome back to The Joy of Why. Now, so in your own past, I think I described you as a computer scientist, but it seems that your work has been very interdisciplinary with contributions from linguistics, from psychology, cognitive science. What led you in that science and neuroscience and training. But the reason why I find that important for my own research is because there's common ground in the quest of understanding intelligence, whether it's a form of artificial intelligence or human intelligence. There's some insights that I could draw from these other fields. And especially now that AI becomes a lot more human-like, or at least it demonstrates human-like capabilities, I personally believe that it is ever more important to do interdisciplinary
Starting point is 00:23:49 research across these fields. So we did want to really talk to you about understanding and this question of common sense. We've been talking about how great these chat GPT and other large language model-based bots are at some kinds of tasks, but what are some of their weaknesses? Or I mentioned in the introduction, mistakes they make that are sort of silly or even hilarious. Yeah, so this is an example that I gave in my TED talk where I asked, if I left five clothes to dry out in the sun and it took them five hours to dry completely, how long would it take to dry 30 clothes? Chachi P.T. then said it would take 30 hours to dry 30 clothes. Now this is Chachi P.T. trying to be too smart. In fact, when you dry your clothes in the sun, you can dry them
Starting point is 00:24:41 all simultaneously, so you don't need to do the math. You just say the five hours. So this example became very popular. Soon after, the problem seemed to be fixed. But then just in case, I figured that I'm going to ask the same question, but actually phrased differently. We reordered the clauses and phrases a little bit, and then GPT-4 wasn't able to answer this one correctly for some time. Then it got fixed about a month or two. So I thought the problem has been really fixed. But just in case, I decided I'm going to ask just one more variant, which was that suppose it takes three hours to dry a shirt and five hours to dry a pair of pants in the sun. For this, ChatGPT goes back to this original mode of like multiplying numbers and give you the wrong answer again. Now, this is
Starting point is 00:25:40 really curious because people usually do not need to do post-tuning or post-training per se for this kind of a question. So once you acquire this basic common sense knowledge about what it means to dry shorts in the sun, you really don't need to go over different cases and teach yourself whether you should multiply the drying hours proportionately or you should use the same number because you can try them concurrently. Once you have that common sense knowledge, you're good. The curious thing about chat GPT is that for some reason this is very confusing. I mean, of course, there are people who will get confused about these too. So if it is accumulating the wisdom of the internet, in a way, I'm not so surprised it has trouble with this kind of question. But still, it is really bizarre because
Starting point is 00:26:38 it can do all kinds of much harder calculations. But as you say, this is an example where it's common sense is letting us down. And in one of your talks, or maybe it's one of your papers, I saw it somewhere that you referred to common sense as the dark matter of intelligence. I thought it was a really provocative statement. Could you tell us a little more about what you mean by the dark matter of intelligence? Yeah. So the reason why I said that is because common sense really is the unspoken rules about how the world works, how the physical world works, and how the social world works. So this really influenced the way that we use language, we interpret language. And that's really one of the key aspects of human intelligence. And
Starting point is 00:27:27 the mysterious thing about common sense is that humans acquire it presumably easily. I mean, as in like everyone has it, but it's strikingly hard to write them down, to teach machines about these rules that we somehow acquired. So for a long time in AI, common sense was viewed as one of the hardest challenges to overcome. That said, I should really acknowledge that GPT-4, ChatGPT, has acquired really impressive amount of common sense. I've never seen anything quite like that before in AI. So I'm not denying that it didn't acquire any common sense. It did acquire a lot of common sense, but unlike human common sense, which will be, relatively speaking, a lot more robust to the sort of questions that I demonstrated earlier. There are many, many more examples, by the way.
Starting point is 00:28:31 Machines are strikingly brittle when provided with that kind of examples. And here's the reason why. The common sense questions are generally so trivial, it doesn't really appear on the internet as much. If it did appear, by the way, then Chachapiti has learned it. So a lot of common sense that does appear on the internet, like that apples are edible, you know, apples can be usually red or green color, probably not purple or, you know, blue. So these things are now acquired as some sort of like a factual knowledge, but there are other things that are not spoken out loud, then it's less likely that Chachapit has acquired it.
Starting point is 00:29:17 So that's sort of understandable given that it doesn't get to live in the world. So far, its window on the world is text, at least the way that they're being trained. Have you and your group been trying to feed common sense into these kinds of large language models? Yeah, so in my lab, we've been trying to study how to teach common sense in a more effective way, perhaps by mimicking how when
Starting point is 00:29:46 children grow up, they do ask a lot of why this, why that questions. The kind of questions that adults wouldn't ask to each other, it may be obvious to adults, but children while growing up are provided with a lot of such declarative description of common sense. So we attempted at writing down a lot of such common sense rules and then trained the neural network. And we found that the neural network can really generalize fast out of those examples. So that's one way to teach neural network common sense much faster by providing this collection of declarative knowledge. By the way, just like the way that ChatGPT is trained is such that it's going to train on anything on the internet.
Starting point is 00:30:34 By the cutoff time of the data collection, it has read this symbolic common sense knowledge graph that our lab has generated and shared on the internet as well. Uh-huh. Interesting. It reminds me of something that I remember hearing about when I was a professor at MIT. There were so many students there, and probably I should say professors too, who lacked a certain kind of social grace, who didn't know the rules for how people are supposed to interact with each other, that there were courses like etiquette lessons or manners lessons that very explicit declarative instructions for people. Yeah. I mean, I haven't seen such a class in person, but I can totally imagine that there could
Starting point is 00:31:19 be one and some people can benefit. Yeah. But so then if you had the textbook for that course, and if that was part of the training, might be helpful to our friends, the AIs too. Oh, I'm sure that it has already read all of them. Okay. Well, we talked a little bit earlier about the role of emotion and whether it would be helpful in acquiring common sense for AIs to have some emotional capabilities. But leaving that aside, I wanted to explore with you some other things that they notably
Starting point is 00:31:53 don't have. And it feels to me, and I think other people have made this point, that there are a lot of very severe obstacles for them to acquire common sense because, for example, they don't have bodies. Like a little kid gets to fall down or play with toys and they have hands. You know, they don't have a place in society. They don't get to interact with other AIs necessarily or with people. Like they're just missing out on the richness of existence.
Starting point is 00:32:23 I guess my question is, are those things fundamental obstacles? Like, do we have to wait till we get robots that can do those things, move around in space, have emotions, have bodies, have social interactions? Maybe common sense has to wait for all of those? Or is that too pessimistic? That's a great question. It's comforting to believe that due to lack of emotion and embodiment, maybe AI cannot go too far. Wouldn't that be nice if that's true? But I'm not sure whether that's the case. Because with language-only interface, it can still do a lot, really a lot.
Starting point is 00:33:04 So, I mean, that's one thing. But that aside, whether AI lacking true emotion and the true embodiment, whether that's a good thing or bad thing, on one hand, it's definitely a limitation compared to human intelligence. But on the other hand, whether that's the only way to acquire the kind of intelligence that humans have may or may not be true. It's not clear. This is the kind of scientific question that is not as well understood yet. Therefore, we should build a robot that has true emotion. I mean, AI should have emotional intelligence and awareness so that it's going to interact with humans in a pleasant, non-harmful way. But when AI has its own desire and emotion, that may be just an interesting intellectual question,
Starting point is 00:34:04 but I'm not sure in terms of the benefit to humanity, whether that's the right kind of question even to ask in the first place. I mean, let's just say that AI falls in love with another human being. Like it really feels the love. Is that a good thing for humanity? especially if it's going to start doing things that could be harmful for other human beings because it's willing to sacrifice everyone else to serve this one human in the world? Oh, boy. skeptical that we can go that far because the thing about the bio-embodiment is that the human fingers, for example, are unbelievably dexterous. We don't yet know how to make delicate joints that can move around in all different angles. And then, you know, humans taste buds. Is it even necessary to build a robot that can smell and taste in the way that humans do? It's maybe a philosophical question too, but I personally don't think it's all that important
Starting point is 00:35:12 to build robots that really truly mimics every capability of a human being. But we don't even have the technology. Yeah. No, it's also interesting to, since we're kind of speculating now and letting our imaginations go, we could endow them with other senses that we don't currently have. Like, for instance, the sonar that bats use or electric fish swimming intelligence. But as you say, it's not at all clear that this is a good idea to be doing any of this, even if we could do it. So maybe we should close with the final part of our discussion, thinking about questions like this, about policy, about transparency. It's such an expensive pursuit, as we already mentioned, to build these, that only very few people or organizations are making the decisions right now. And they have proprietary data and techniques. Do you see this as a big issue in the field? It's a huge issue in the field.
Starting point is 00:36:19 What could possibly go wrong with such concentration of power? Yeah. I think especially the opaqueness of the data does feed into unnecessary hypes and fear as well. Going back to your earlier example of how chat GPT might answer in a very lawyer-like way that it may or may not understand in the way that humans do, and it's just a machine and it has limitation. When it says that, does it say that because that's exact data that was used for post-training adaptation of chat GPT so that it's going to interact with humans in a more politically correct way?
Starting point is 00:37:07 Or is it that it has genuinely acquired such self-awareness and introspection capabilities to realize that, oh, I'm a mere AI trained on human data? I think if post-training data was transparent, a lot of those unnecessary hypes would be addressed. And also, I think for the purpose of AI safety as well, I personally believe that more transparency is helpful so that we better understand where the limitations are, where the flaws are. Is this something that governments need to impose on the big companies? Is that what you're suggesting? Probably there should be more government involvement in thinking about AI policy. It's a very important topic that one needs to address very carefully, though, because
Starting point is 00:38:01 I can totally also imagine a policy that just slows things down unnecessarily without actually adding much depending on how it's implemented. So it's an effort that requires really a broad range of community involvement and also there needs to be effort to increase AI literacy across people outside AI, including policymakers, but even for daily users of AI, so that they understand really what the limitations of these models might be, so as not to overtrust. Well, I'm reminded of a time long ago when genetic engineering was new, and a lot of biochemists and molecular biologists on their own got together to make sort of police themselves about what kinds of experiments they would conduct or not allow themselves to conduct.
Starting point is 00:38:57 I wonder is that something rather than having the governments do it, do you think maybe the community itself should be coming together, including the big companies? Do you think that's the way to go? In general, I think there's collective efforts where people of all sectors have a way of contributing what AI should and should not be in some high-level declarative sense. We probably all agree that AI shouldn't be used to develop bioweapons or AI should not propagate racism and sexism. But then there can be more gray zones and we then need to think about what to do with those gray zones. What do you see as the biggest dangers in this space right now?
Starting point is 00:39:48 What do you realistically think we should be worried about? I think there are a lot to worry, especially in the near term, like misinformation, increasing use of AI for generating fake media to support a particular political party. That's one thing. But actually also seemingly benign use cases such as, you know, people faking their social media feed might have more longer term consequence in the way that people generate and consume social media content. You know, by the way, I used to think internet is the byproduct of human intelligence, but that may not be the case in the coming years because too many people use ChatGPT
Starting point is 00:40:34 for all sorts of their writing jobs, I hear. There are even some papers that start using ChatGPT and then the authors were not diligent enough to remove when chat GPT says, oh, it's an AI model, blah, blah. At least they should have read. Oh, I shouldn't laugh. Yeah, no, it's serious. But then again, I mean, let's be real. I have a colleague who's a very good and honest person. English is not his first language. And he has told me that he uses ChatGPT to improve the grammar of the abstract for his papers. You know, that seems like a fairly benign use.
Starting point is 00:41:17 He's written the abstract. It's just sort of like a writing coach helping him. It's not really providing new ideas. So these things can be good tools if used properly. Certainly, yeah, it could. sort of like a writing coach helping him. It's not really providing new ideas. Right. So these things can be good tools if used properly. Certainly, yeah. It could help people to learn language faster. It can help as a writing companion if used correctly.
Starting point is 00:41:39 But it's going to have unwanted side effect on humans as well. I do wonder whether it may be okay in the end, but I do wonder personally whether it's going to somehow make measurable changes in the human capabilities of writing and reading comprehension or not in the longer term. Well, so just to wrap up, the one thing that we often like to ask our guests, Well, so just to wrap up, one thing that we often like to ask our guests, since our show is called The Joy of Why, is to talk about the emotional side of being a scientist yourself. Is there something in your research that brings you special joy? Oh, yeah.
Starting point is 00:42:26 Great question. joy, these questions about like, are there limitations in chat GPT? And if so, why? Why does it work so well based on just reading internet text? Seeking answers to these why questions, I don't really know why, but it does give me a lot of pleasure. And perhaps that's one of the differentiating factors compared to human intelligence and chat GPT, that we ask why. Yes, we do ask why. And thank you so much for helping us understand why. We've been speaking with Yejin Choi.
Starting point is 00:42:57 It's been delightful to have you here with us today. Thanks so much. Thank you. It was so fun. Thanks for listening. If you're enjoying The Joy of Why and you're not already subscribed, hit the subscribe or follow button where you're listening. You can also leave a review for the show. It helps people find this podcast. The Joy of Why is a podcast from Quantum Magazine, an editorially independent publication supported by the Simons Foundation.
Starting point is 00:43:28 Funding decisions by the Simons Foundation have no influence on the selection of topics, guests, or other editorial decisions in this podcast or in Quantum Magazine. The Joy of Why is produced by PRX Productions. The production team is Caitlin Falds, Livia Brock, Genevieve Sponsler, and Merritt Jacob. The executive producer of PRX Productions is Jocelyn Gonzalez. Morgan Church and Edwin Ochoa provided additional assistance.
Starting point is 00:43:56 From Quantum Magazine, John Rennie and Thomas Lin provided editorial guidance with support from Matt Karlstrom, Samuel Velasco, Arlene Santana, and Megan Wilcoxon. Samir Patel is Quanta's editor-in-chief. Our theme music is from APM Music. Julian Lin came up with the podcast name. The episode art is by Peter Greenwood, and our logo is by Jackie King and Christina Armitage. Special thanks to the Columbia Journalism School
Starting point is 00:44:27 and Bert Odom-Reed at the Cornell Broadcast Studios. I'm your host, Steve Strogatz. If you have any questions or comments for us, please email us at quanta at simonsfoundation.org. Thanks for listening.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.