The Joy of Why - Will AI Ever Have Common Sense?
Episode Date: July 18, 2024Common sense rules our world. This fundamental, sometimes trivial knowledge is inherent to how humans interpret language. Yet, some of these simple human truths are so obvious that they're ra...rely put into words. And without the data of common sense to train on, large language models such as ChatGPT have bizarre, often humorous blind spots.Yejin Choi, professor and the chair of computer science at the University of Washington, calls common sense the “dark matter” of intelligence. In this week’s episode of “The Joy of Why,” Choi talks with co-host Steven Strogatz about decoding the interstitial glue of language and comprehension. Together, they explore the question: Should we program more humanity into the next generation of artificial intelligence?
 Transcript
 Discussion  (0)
    
                                         Hello, I'm Brian Cox.
                                         
                                         I'm Robin Ince, and this is the Infinite Monkey Cage trailer for our brand new series.
                                         
                                         We've got mummies, we've got magic, we've got asteroids.
                                         
                                         Mummies, magic and asteroids. What's the link?
                                         
                                         That it was an asteroid that magically went over the world that led to Imhotep the mummy coming back to life?
                                         
                                         That's correct.
                                         
                                         I thought it would be weird, scientific as ever.
                                         
                                         But the most important thing to know is that we are going to deal with the biggest scientific question we finally ask.
                                         
    
                                         What is better, cats or dogs?
                                         
                                         Listen wherever you get your podcasts.
                                         
                                         Going back to at least the 1960s, computer scientists have been dreaming about brain-inspired computers
                                         
                                         that might someday demonstrate human-like intelligence.
                                         
                                         With the rise of the internet, the availability of enormous textual datasets, and impressive
                                         
                                         advances in computational power, we've reached a momentous point.
                                         
                                         Large language models, or LLMs, can often seem to wield something close to human intelligence,
                                         
                                         at least to us non-experts.
                                         
    
                                         And the release of OpenAI's ChatGPT in 2022 helped these LLMs make their mark in the headlines,
                                         
                                         in the workplace, and in dinner table conversations.
                                         
                                         But there's still a telltale sign that large language model intelligence
                                         
                                         is truly artificial. Their lack of common sense, which can emerge in spectacular and sometimes
                                         
                                         hilarious ways in the mistakes they make. I'm Steve Strogatz, and this is The Joy of Why,
                                         
                                         a podcast from Quantum Magazine, where I take turns at the mic with my co-host,
                                         
                                         Jana Levin, exploring some of the biggest questions in math and science today.
                                         
                                         In this episode, we're going to be speaking to computer scientist Yejin Choi
                                         
    
                                         about the architecture and capabilities of large language models
                                         
                                         and speculate about whether artificial intelligence, AI, will ever gain common sense.
                                         
                                         Yejin Choi is a professor and the chair of computer science at the University of Washington,
                                         
                                         where she researches statistical approaches and computational models for natural language processing.
                                         
                                         She was recognized as a 2022 MacArthur Fellow
                                         
                                         and named one of Time Magazine's 100 Most Influential People in AI in 2023.
                                         
                                         Yejin, thanks so much for joining us here on The Joy of Why. Thank you for having me. I'm excited to be here. Great. Well, this is going to be so much
                                         
                                         fun. I'm so fascinated, as I know so many people are, by what's going on these days in AI. And so
                                         
    
                                         while I was preparing for this chat, just as a little bit of a joke, I was curious.
                                         
                                         I asked ChatGPT, a chatbot built on one of these large language models, do you understand?
                                         
                                         And it replied, as an AI, I don't possess consciousness or subjective understanding in the way that humans do.
                                         
                                         While I can generate responses that seem like understanding, it's important to recognize
                                         
                                         that this understanding is mechanistic and based on statistical patterns rather than true comprehension.
                                         
                                         What do you think about that answer from ChatGPT?
                                         
                                         It sounds like someone coached ChatGPT to say the right thing.
                                         
                                         That's funny.
                                         
    
                                         That it was coached?
                                         
                                         You don't think it would come up with that on its own?
                                         
                                         We don't think it would come up with that on its own? of particular style of language, lawyer-like language that ChatGPT distinctly use that's provided by humans as good examples to follow.
                                         
                                         So it is not the case that ChatGPT, after reading the raw internet, suddenly speaks
                                         
                                         like that.
                                         
                                         It's because of the post-training that does coach ChatG GPT to speak in a more safe, more politically correct and careful ways.
                                         
                                         I get it. Sure. No, that makes sense. There's so much at stake for these companies.
                                         
                                         And, of course, people could really be harmed, I suppose, if they take some of these responses too literally.
                                         
    
                                         I noticed that all the different chatbots have
                                         
                                         disclaimers now on the front page that these can make mistakes, you need to check them.
                                         
                                         Yeah. I mean, clearly the lawyers have had their, as you say, coaching sessions with these bots.
                                         
                                         Totally. Okay. But so maybe this is a good point for us to back out just for a second,
                                         
                                         to say that these large language models, which most of us know through products like ChatGPT
                                         
                                         or Google's Gemini or Microsoft Copilot or Anthropix, Claude, these are just one type
                                         
                                         of artificial intelligence. And so since our listeners may have been hearing words, as I think
                                         
                                         we all have, like machine learning, large language models, AI, neural networks. Could you help us just
                                         
    
                                         distinguish, are some of those subsets of the others? How should we think about those kinds of
                                         
                                         terms? So these are all fairly broad terms that definitely have their own distinct definitions,
                                         
                                         but they overlap a lot. So machine learning in general is about an algorithm that
                                         
                                         teaches machine to learn some patterns between input-output pairs. And then artificial intelligence
                                         
                                         is perhaps arguably more broadly about computational forms of intelligence that
                                         
                                         can do certain operations, but that could be using machine learning or it could be using just algorithms, inference algorithms. So neural network is just like one type of machine learning
                                         
                                         algorithm that is currently the most popular probably. So for example, the computational
                                         
                                         chess players in the earlier forms of AI, they were just inference algorithms, whereas the more modern
                                         
    
                                         version of that would be in the forms of a neural network that is a form of machine learning.
                                         
                                         Good.
                                         
                                         That's very helpful.
                                         
                                         And so what are some of the capabilities and also limitations of these large language models?
                                         
                                         So the capabilities of these large language models are phenomenal.
                                         
                                         It's really beyond what the scientists have anticipated.
                                         
                                         Just anything that you could provide as textual input,
                                         
                                         it turns out these large language models can do quite well in answering them,
                                         
    
                                         even if it requires open-ended answers.
                                         
                                         Long text input output turns out to be really, really great.
                                         
                                         So not only can it do simple reasoning, multiple choice question type, question answers, but
                                         
                                         also any topic that you throw at ChatGPT, it will be able to answer strikingly well.
                                         
                                         The truth is, though, it's really a reflection of
                                         
                                         human intelligence that is shared on the internet. And internet is vast. Humans typically are not
                                         
                                         really aware of how vast that is because humans have limited capabilities in digesting and reading
                                         
                                         what's out there. But the machine uniquely can really read them all, literally.
                                         
    
                                         So then it can mimic the sort of knowledge and wisdom that people shared online,
                                         
                                         and then in some sense, read them back to you, but not in the verbatim sense,
                                         
                                         but in a more rephrased sense.
                                         
                                         So it's not like the exact copy of what it has written from the internet,
                                         
                                         but it's able to rephrase, it's able to synthesize,
                                         
                                         so that it sounds new enough for people.
                                         
                                         That is, as you say, phenomenal.
                                         
                                         One of the most remarkable things.
                                         
    
                                         You can ask it to generate college application essays
                                         
                                         or help you write a Python program.
                                         
                                         I mean, I tried one time just for fun to have it write a Saturday Night Live skit
                                         
                                         where Donald Trump plays a psychiatrist.
                                         
                                         And he tries to, you know, give advice to his patient but in the style of Donald Trump.
                                         
                                         And it was really funny and did something that sounded a lot like Trump.
                                         
                                         And I can't imagine that, you know, as you say, that's not verbatim on the internet.
                                         
                                         I don't think Trump has ever been playing a psychiatrist.
                                         
    
                                         But it's amazing the synthesis that it can pull off.
                                         
                                         So some people have called it sort of spicy autocomplete, what these things do.
                                         
                                         But maybe you should tell us, why would anyone say that?
                                         
                                         What is it that these large language models are doing fundamentally?
                                         
                                         It's able to read a lot of text and learn to predict which word comes next.
                                         
                                         During training, literally, all that it does is trying to predict which word comes next.
                                         
                                         But at an extreme scale.
                                         
                                         to predict which word comes next, but at an extreme scale. For that reason, some people just diminuate large language models as spice auto completion
                                         
    
                                         machine.
                                         
                                         The reason why it's not necessarily doing verbal team regurgitation of the training
                                         
                                         data, though, is because of the particularities of the technical details being used under the hood,
                                         
                                         which is not necessarily about memorization. It's also able to do some degrees of generalization
                                         
                                         and also there's a randomness in the way that this text is generated out of the learned neural network. And that randomness causes how the text is not necessarily verbatim regurgitation always.
                                         
                                         But I mean, sometimes it can be, by the way, if the text was repeated often enough on the
                                         
                                         internet data, then it's going to actually verbatim memorize that.
                                         
                                         And, you know, there was some incidents that New York Times reported that it was able to
                                         
    
                                         regurgitate some of the past New York Times articles.
                                         
                                         Oh, really?
                                         
                                         I haven't heard about that.
                                         
                                         I see.
                                         
                                         So it can plagiarize in that sense.
                                         
                                         I mean, one could say that it's plagiarized.
                                         
                                         Another person could say that, well, this is neural network being able to retrieve what
                                         
                                         it has read.
                                         
    
                                         But regardless, because of that, some people diminuate these motions as, oh, it's just
                                         
                                         like auto-completion.
                                         
                                         But the reason why it's able to do something striking like, you know, discussing some topics
                                         
                                         in a Trump style that Trump may have never done before, that's possible because these
                                         
                                         motions are capable of interpolation between two data points. The novel
                                         
                                         interpolation that nobody has ever done before is trivial for these machines. So you do get
                                         
                                         that kind of a novelty to some degree as well. Well, so you've been mentioning training,
                                         
                                         and I think it would be great if you could explain to us a little bit in detail,
                                         
    
                                         what does that really mean?
                                         
                                         How would you train a large language model?
                                         
                                         Or how are these big companies that have built ChatGPT or Gemini, what do they do to train their models?
                                         
                                         So basically, the training boils down to building extremely large neural network that has layers and layers and layers of neurons piled up and then feed
                                         
                                         internet data in sequence. And the goal of this learning process is to predict which word comes
                                         
                                         next, conditioning on the sequence of previous words. And what is striking is that that simple recipe of training neural network can lead to such powerful artifacts that can do all sorts of question answering in text that comes across as a striking level of artificial intelligence for many people. But importantly, that kind of training is really, really different from how
                                         
                                         humans learn about the world, which we don't really know how humans really learn. However,
                                         
                                         it's reasonable to suspect that the humans don't necessarily try to predict which world comes next,
                                         
    
                                         but we rather try to focus on making sense of the world. So we tend to abstract away immediately.
                                         
                                         You and I, by the way, are not able to remember the discussions, the interactions, the conversation.
                                         
                                         We just had verbal team.
                                         
                                         We just cannot.
                                         
                                         Because our brain is trained to abstract immediately.
                                         
                                         But we do remember the gist of our conversation so far, such that if you ask me the
                                         
                                         same question again, I'll be surprised. So there's something about the way that humans learn. And
                                         
                                         also humans learn with curriculum and curiosity, and we make hypothesis about the world. And then
                                         
    
                                         if something doesn't make sense, even children, even babies, they try to do some experiments to figure out their confusion points about simple objects, the physical knowledge about the objects that they interact with.
                                         
                                         But machines, from day one, they're fed with the New York Times articles.
                                         
                                         New York Times articles. And they don't have any saying in what order they're going to read this text, nor do they have any saying about, oh, wait a minute, I really want to read something again.
                                         
                                         There was something really nice and curious about this particular, say, Hemingway's book that caught
                                         
                                         my attention that I want to read slower. The way that learning happens is so different, and it's quite striking how you can bypass
                                         
                                         the normal way of learning of humans and then still produce something that speaks human
                                         
                                         language so well.
                                         
                                         You've raised so many interesting points there.
                                         
    
                                         For instance, when talking about babies or people at any age, really, that we have curiosity,
                                         
                                         we have desires, like we want to read or reread that
                                         
                                         Hemingway passage, or maybe there's something we don't like reading and we would like to skip it.
                                         
                                         So far, we have talked about artificial intelligence, but we haven't talked about
                                         
                                         artificial emotion, right? Like the fact that these machines so far were not really putting
                                         
                                         in desires, it seems like that might limit what they can do,
                                         
                                         given that babies and people of all ages have willpower, they have desire, they have things
                                         
                                         they wish. Do you think that emotion is a big part in human learning that these machines are missing?
                                         
    
                                         Yeah, that's a great point. And in fact, it boils down to the fact that we are
                                         
                                         bio-beings at the end of the day. We have desire, like a deep sense of self-identity that really
                                         
                                         makes who we are. And it's not something we can change. We are born with this individual identity
                                         
                                         and then we live with it. We live once and then we live with it.
                                         
                                         Whereas AI, it's not clear what it really is
                                         
                                         because it just read everybody's writing in some sense
                                         
                                         and it became some average viewpoint
                                         
                                         or thoughts and emotional soup
                                         
    
                                         that does mimic all this human emotion and intent due to the
                                         
                                         human emotion and intent that humans put into their writing.
                                         
                                         So then these machines are capable of mimicking all that.
                                         
                                         But you're right that at the end of the day, it doesn't really have the kind of genuine
                                         
                                         emotion that humans have.
                                         
                                         Now, whether that's a bad thing or
                                         
                                         good thing, that's a philosophical question. That may be even a scientific question in terms of
                                         
                                         safety. Is it a good thing if AI really, really develops its own emotion such that, you know,
                                         
    
                                         it has like survivor instinct or it wants to dominate the world.
                                         
                                         Is that a good thing or not?
                                         
                                         Well, this is, of course, something that all of us are thinking about nowadays.
                                         
                                         And we should probably save that question for a little bit later,
                                         
                                         because I wanted to ask a little bit more about the training,
                                         
                                         because it is, as you say, such an inhuman thing that we ask of them.
                                         
                                         We ask them to predict the next word given a passage.
                                         
                                         And what happens when
                                         
    
                                         they get it wrong? The way that it's trained is it's supposed to maximize the probability score
                                         
                                         that it should assign to the correct word. If it predicted the wrong word, it means the wrong word
                                         
                                         got the higher probability, but all that there is is to raise the probability of the correct
                                         
                                         word that that should have happened instead.
                                         
                                         But not by directly dealing with that word, right?
                                         
                                         I mean, I guess what I'm going for is the idea of the weights in a neural network and
                                         
                                         how there are mechanisms to change the weights.
                                         
                                         Right.
                                         
    
                                         So perhaps I should pedal back a little bit and then talk about that there are two phases of training. One is pre-training. The second is post-training, also known as reinforcement learning with human feedback. That's a jargon, which actually is not just reinforcement learning, but also mixed with something known as sequential fine-tuning or supervised learning.
                                         
                                         But anyway, by and large, there's pre-training and post-training.
                                         
                                         During pre-training, the learning mechanism is basically to maximize the probability score that will be assigned to the correct sequence of words,
                                         
                                         meaning the exact sequence of words that happen to be on the internet.
                                         
                                         Now, there's really no reason why that's the only correct sequence of words, by the way,
                                         
                                         because for any given prefix text, there can be another different word that could be okay to say.
                                         
                                         So the notion of correctness is not quite right.
                                         
                                         But anyway, so the neural networks are trained to maximize the probability.
                                         
    
                                         And what does that entail in terms of the weights that these neural networks learn is that the way
                                         
                                         that these machines are learned is basically based on what's known as backpropagation.
                                         
                                         We basically take gradients of individual weights, partial gradients. So you take the derivative, a partial derivative of
                                         
                                         every weight of neural network. There are just so many of the weights, by the way,
                                         
                                         hundreds of billions of parameters that you take partial derivative of, and then you move that
                                         
                                         weight so that it's going to increase the probability score assigned to the particular
                                         
                                         sequence of words that were in the
                                         
                                         training data? Well, as a mathematician, I'm very happy to hear you talking about partial derivatives,
                                         
    
                                         but some of our listeners may not feel the same way. So let me try an analogy. So I like to play
                                         
                                         tennis, and I remember when I was learning tennis, sometimes, you know, the ball comes to me and I might hit a bad shot.
                                         
                                         And then the tennis coach says, you need to get your racket back earlier.
                                         
                                         You know, you weren't prepared.
                                         
                                         So then I make an adjustment to I don't know what.
                                         
                                         I want to say weights.
                                         
                                         Like I have some kind of internal representation of how important is it that my feet are in the right place or that I've turned my body sideways or I got my racket back or I keep my eye on the ball.
                                         
                                         I have all these different weights that I have to pay attention to. And given that this shot was bad,
                                         
    
                                         so to speak, I'm going to try to adjust my weight so that I'll do it better the next time.
                                         
                                         It's something like that, right? Yeah, yeah, yeah. That's a really great analogy.
                                         
                                         Okay. All right. But it's a brutal way to learn that you just make this poor machine
                                         
                                         take one test question after another, and every time it gets it wrong, you punish it, so to speak,
                                         
                                         or at least you gently correct it by making it adjust its weight so that it would do better
                                         
                                         the next time.
                                         
                                         Yeah. On and on and on.
                                         
                                         On and on and on.
                                         
    
                                         It's a very brutal training.
                                         
                                         It's good that motion doesn't have emotion.
                                         
                                         Right.
                                         
                                         So it doesn't care, I guess, as far as we know.
                                         
                                         But this is the pre-training, you say.
                                         
                                         Yeah, yeah.
                                         
                                         And then during post-training, there are multiple things that can happen.
                                         
                                         But maybe let me highlight just the most representative one, which is reinforcement learning with human feedback.
                                         
    
                                         So in this particular type of post-training phase, what happens is that you present the machine's answer to a query to human evaluator, and humans can give a thumbs up, thumbs down.
                                         
                                         and humans can give a thumbs up, thumbs down.
                                         
                                         And then based on that, you then go back to the neural network to adjust the weight a little bit using the analogy that you used before.
                                         
                                         But this time, instead of focusing on which word comes next,
                                         
                                         you're focusing on whether you get thumbs up or thumbs down by the human evaluator.
                                         
                                         Okay. I see.
                                         
                                         So all this process, I mean, we know that computers
                                         
                                         run fast, but still, how long would it take to do, say, the pre-training phase or just give us a
                                         
    
                                         ballpark feeling for it? Are we talking about days, weeks? Yeah. I mean, it varies a lot depending on
                                         
                                         how much data that you are going to feed versus how much of a compute you have versus how
                                         
                                         large neural network you want to train.
                                         
                                         So there are many variables at play that determines how long it takes.
                                         
                                         And also, by the way, the tech companies don't share exactly how long it took and how much
                                         
                                         of a compute they used, but one can speculate.
                                         
                                         I would say the really good ones typically takes like a couple months
                                         
                                         if you want to push the limit.
                                         
    
                                         But, I mean, if you want to stop earlier with a smaller amount of data,
                                         
                                         then it could be just a matter of a couple days.
                                         
                                         By the way, if you really think about how long humans learn,
                                         
                                         10 years of learning as a human baby, you know, becoming a child,
                                         
                                         they still have a lot more to learn.
                                         
                                         So in some sense, a couple months isn't so bad.
                                         
                                         That is true.
                                         
                                         We'll be right back.
                                         
    
                                         Welcome back to The Joy of Why. Now, so in your own past, I think I described you as a computer scientist, but it seems that your work has been very interdisciplinary with contributions from linguistics, from psychology, cognitive science.
                                         
                                         What led you in that science and neuroscience and training.
                                         
                                         But the reason why I find that important for my own research
                                         
                                         is because there's common ground in the quest of understanding intelligence,
                                         
                                         whether it's a form of artificial intelligence or human intelligence.
                                         
                                         There's some insights that I could draw from these other fields.
                                         
                                         And especially now that AI becomes a lot more human-like, or at least it demonstrates
                                         
                                         human-like capabilities, I personally believe that it is ever more important to do interdisciplinary
                                         
    
                                         research across these fields. So we did want to really talk to you about understanding and this
                                         
                                         question of common sense. We've been talking about how great these chat GPT and other large language
                                         
                                         model-based bots are at some kinds of tasks, but what are some of their weaknesses?
                                         
                                         Or I mentioned in the introduction, mistakes they make that are sort of silly or even hilarious.
                                         
                                         Yeah, so this is an example that I gave in my TED talk where I asked,
                                         
                                         if I left five clothes to dry out in the sun and it took them five hours to dry completely, how long would it
                                         
                                         take to dry 30 clothes? Chachi P.T. then said it would take 30 hours to dry 30 clothes. Now this is
                                         
                                         Chachi P.T. trying to be too smart. In fact, when you dry your clothes in the sun, you can dry them
                                         
    
                                         all simultaneously, so you don't need to do the math. You just say the five hours. So this example became very popular. Soon after, the
                                         
                                         problem seemed to be fixed. But then just in case, I figured that I'm going to ask the same question,
                                         
                                         but actually phrased differently. We reordered the clauses and phrases a little bit, and then
                                         
                                         GPT-4 wasn't able to answer this one correctly for some time. Then it got fixed about a month or two.
                                         
                                         So I thought the problem has been really fixed. But just in case, I decided I'm going to ask just
                                         
                                         one more variant, which was that suppose it takes three
                                         
                                         hours to dry a shirt and five hours to dry a pair of pants in the sun. For this, ChatGPT goes back
                                         
                                         to this original mode of like multiplying numbers and give you the wrong answer again. Now, this is
                                         
    
                                         really curious because people usually do not need to do post-tuning or post-training per se for this kind of a question.
                                         
                                         So once you acquire this basic common sense knowledge about what it means to dry shorts in the sun,
                                         
                                         you really don't need to go over different cases and teach yourself whether you should multiply the drying
                                         
                                         hours proportionately or you should use the same number because you can try them
                                         
                                         concurrently. Once you have that common sense knowledge, you're good. The curious
                                         
                                         thing about chat GPT is that for some reason this is very confusing. I mean, of course, there are people who will get confused
                                         
                                         about these too. So if it is accumulating the wisdom of the internet, in a way, I'm not
                                         
                                         so surprised it has trouble with this kind of question. But still, it is really bizarre because
                                         
    
                                         it can do all kinds of much harder calculations. But as you say, this is an example where it's
                                         
                                         common sense is letting
                                         
                                         us down. And in one of your talks, or maybe it's one of your papers, I saw it somewhere that you
                                         
                                         referred to common sense as the dark matter of intelligence. I thought it was a really provocative
                                         
                                         statement. Could you tell us a little more about what you mean by the dark matter of intelligence? Yeah. So the reason why I said that is because
                                         
                                         common sense really is the unspoken rules about how the world works, how the physical world works,
                                         
                                         and how the social world works. So this really influenced the way that we use language,
                                         
                                         we interpret language. And that's really one of the key aspects of human intelligence. And
                                         
    
                                         the mysterious thing about common sense is that humans acquire it presumably easily. I mean,
                                         
                                         as in like everyone has it, but it's strikingly hard to write them down, to teach machines about these rules that we somehow acquired.
                                         
                                         So for a long time in AI, common sense was viewed as one of the hardest challenges to overcome.
                                         
                                         That said, I should really acknowledge that GPT-4, ChatGPT, has acquired really impressive
                                         
                                         amount of common sense. I've never seen anything quite
                                         
                                         like that before in AI. So I'm not denying that it didn't acquire any common sense. It did acquire
                                         
                                         a lot of common sense, but unlike human common sense, which will be, relatively speaking, a lot more robust to the sort of questions that I demonstrated earlier.
                                         
                                         There are many, many more examples, by the way.
                                         
    
                                         Machines are strikingly brittle when provided with that kind of examples.
                                         
                                         And here's the reason why.
                                         
                                         The common sense questions are generally so trivial, it doesn't really appear on the internet as much. If it did
                                         
                                         appear, by the way, then Chachapiti has learned it. So a lot of common sense that does appear on
                                         
                                         the internet, like that apples are edible, you know, apples can be usually red or green color,
                                         
                                         probably not purple or, you know, blue. So these things are now acquired as some sort of like a factual knowledge,
                                         
                                         but there are other things that are not spoken out loud,
                                         
                                         then it's less likely that Chachapit has acquired it.
                                         
    
                                         So that's sort of understandable given that it doesn't get to live in the world.
                                         
                                         So far, its window on the world is text,
                                         
                                         at least the way that they're being trained.
                                         
                                         Have you and your group been trying to feed common sense
                                         
                                         into these kinds of large language models?
                                         
                                         Yeah, so in my lab, we've been trying to study
                                         
                                         how to teach common sense in a more effective way,
                                         
                                         perhaps by mimicking how when
                                         
    
                                         children grow up, they do ask a lot of why this, why that questions. The kind of questions that
                                         
                                         adults wouldn't ask to each other, it may be obvious to adults, but children while growing up
                                         
                                         are provided with a lot of such declarative description of common sense.
                                         
                                         So we attempted at writing down a lot of such common sense rules and then trained the neural
                                         
                                         network.
                                         
                                         And we found that the neural network can really generalize fast out of those examples.
                                         
                                         So that's one way to teach neural network common sense much faster by providing this collection of declarative knowledge.
                                         
                                         By the way, just like the way that ChatGPT is trained is such that it's going to train on anything on the internet.
                                         
    
                                         By the cutoff time of the data collection, it has read this symbolic common sense knowledge graph that our lab has generated and shared on the internet as well.
                                         
                                         Uh-huh. Interesting. It reminds me of something that I remember hearing about when I was a
                                         
                                         professor at MIT. There were so many students there, and probably I should say professors too,
                                         
                                         who lacked a certain kind of social grace, who didn't know the rules for how people are supposed
                                         
                                         to interact with each other, that there were courses like etiquette lessons or manners lessons that very explicit declarative instructions
                                         
                                         for people.
                                         
                                         Yeah.
                                         
                                         I mean, I haven't seen such a class in person, but I can totally imagine that there could
                                         
    
                                         be one and some people can benefit.
                                         
                                         Yeah.
                                         
                                         But so then if you had the textbook for that course,
                                         
                                         and if that was part of the training, might be helpful to our friends, the AIs too.
                                         
                                         Oh, I'm sure that it has already read all of them.
                                         
                                         Okay. Well, we talked a little bit earlier about the role of emotion and whether it would be
                                         
                                         helpful in acquiring common sense for AIs to have some emotional capabilities.
                                         
                                         But leaving that aside, I wanted to explore with you some other things that they notably
                                         
    
                                         don't have.
                                         
                                         And it feels to me, and I think other people have made this point, that there are a lot
                                         
                                         of very severe obstacles for them to acquire common sense because, for example, they don't
                                         
                                         have bodies.
                                         
                                         Like a little kid gets to fall down or play with toys and they have hands.
                                         
                                         You know, they don't have a place in society.
                                         
                                         They don't get to interact with other AIs necessarily or with people.
                                         
                                         Like they're just missing out on the richness of existence.
                                         
    
                                         I guess my question is, are those things fundamental obstacles?
                                         
                                         Like, do we have to wait till we get robots that can do those things, move around in space, have emotions, have bodies, have social interactions?
                                         
                                         Maybe common sense has to wait for all of those?
                                         
                                         Or is that too pessimistic?
                                         
                                         That's a great question. It's comforting to believe that due to lack of emotion and embodiment, maybe AI cannot go too far.
                                         
                                         Wouldn't that be nice if that's true?
                                         
                                         But I'm not sure whether that's the case.
                                         
                                         Because with language-only interface, it can still do a lot, really a lot.
                                         
    
                                         So, I mean, that's one thing. But that aside,
                                         
                                         whether AI lacking true emotion and the true embodiment, whether that's a good thing or bad
                                         
                                         thing, on one hand, it's definitely a limitation compared to human intelligence. But on the other
                                         
                                         hand, whether that's the only way to acquire the kind of intelligence that humans have may or may not be true. It's not clear. This is the kind of scientific question that is not as well understood yet.
                                         
                                         Therefore, we should build a robot that has true emotion.
                                         
                                         I mean, AI should have emotional intelligence and awareness so that it's going to interact with humans in a pleasant, non-harmful way.
                                         
                                         But when AI has its own desire and emotion,
                                         
                                         that may be just an interesting intellectual question,
                                         
    
                                         but I'm not sure in terms of the benefit to humanity, whether that's the right kind of question even to ask in the
                                         
                                         first place. I mean, let's just say that AI falls in love with another human being. Like it really
                                         
                                         feels the love. Is that a good thing for humanity? especially if it's going to start doing things that could be harmful for other human beings because it's willing to sacrifice everyone else to serve this one human in the world?
                                         
                                         Oh, boy.
                                         
                                         skeptical that we can go that far because the thing about the bio-embodiment is that the human fingers, for example, are unbelievably dexterous. We don't yet know how to make delicate
                                         
                                         joints that can move around in all different angles. And then, you know, humans taste buds.
                                         
                                         Is it even necessary to build a robot that can smell and taste in the way that humans do?
                                         
                                         It's maybe a philosophical question too, but I personally don't think it's all that important
                                         
    
                                         to build robots that really truly mimics every capability of a human being.
                                         
                                         But we don't even have the technology.
                                         
                                         Yeah. No, it's also interesting to, since we're kind of speculating now and letting our imaginations go, we could endow them with other senses that we don't currently have. Like, for instance, the sonar that bats use or electric fish swimming intelligence. But as you say, it's not at all clear that this is a good idea to be doing any of this, even if we could do it.
                                         
                                         So maybe we should close with the final part of our discussion, thinking about questions like this, about policy, about transparency.
                                         
                                         It's such an expensive pursuit, as we already mentioned, to build these, that only very few people or organizations are making the decisions right now.
                                         
                                         And they have proprietary data and techniques.
                                         
                                         Do you see this as a big issue in the field?
                                         
                                         It's a huge issue in the field.
                                         
    
                                         What could possibly go wrong with such concentration of power?
                                         
                                         Yeah.
                                         
                                         I think especially the opaqueness of the data does feed into unnecessary hypes and fear
                                         
                                         as well.
                                         
                                         Going back to your earlier example of how chat GPT might answer in a very lawyer-like
                                         
                                         way that it may or may not understand in the way that humans do,
                                         
                                         and it's just a machine and it has limitation. When it says that, does it say that because
                                         
                                         that's exact data that was used for post-training adaptation of chat GPT so that it's going to interact with humans in a more politically correct way?
                                         
    
                                         Or is it that it has genuinely acquired such self-awareness and introspection capabilities
                                         
                                         to realize that, oh, I'm a mere AI trained on human data? I think if post-training data was transparent, a lot of those unnecessary
                                         
                                         hypes would be addressed. And also, I think for the purpose of AI safety as well,
                                         
                                         I personally believe that more transparency is helpful so that we better understand where the
                                         
                                         limitations are, where the flaws are. Is this something that governments need to impose on the big companies?
                                         
                                         Is that what you're suggesting?
                                         
                                         Probably there should be more government involvement in thinking about AI policy.
                                         
                                         It's a very important topic that one needs to address very carefully, though, because
                                         
    
                                         I can totally also imagine a policy that just slows things down unnecessarily without actually adding much depending on how it's implemented. So it's an
                                         
                                         effort that requires really a broad range of community involvement and also there needs to be
                                         
                                         effort to increase AI literacy across people outside AI, including policymakers, but even for daily users of AI,
                                         
                                         so that they understand really what the limitations of these models might be,
                                         
                                         so as not to overtrust. Well, I'm reminded of a time long ago when
                                         
                                         genetic engineering was new, and a lot of biochemists and molecular biologists
                                         
                                         on their own got together to make sort of police themselves about what kinds of experiments
                                         
                                         they would conduct or not allow themselves to conduct.
                                         
    
                                         I wonder is that something rather than having the governments do it, do you think maybe
                                         
                                         the community itself should be coming together,
                                         
                                         including the big companies? Do you think that's the way to go? In general, I think there's
                                         
                                         collective efforts where people of all sectors have a way of contributing what AI should and
                                         
                                         should not be in some high-level declarative sense. We probably all agree that
                                         
                                         AI shouldn't be used to develop bioweapons or AI should not propagate racism and sexism.
                                         
                                         But then there can be more gray zones and we then need to think about what to do with those gray
                                         
                                         zones. What do you see as the biggest dangers in this space right now?
                                         
    
                                         What do you realistically think we should be worried about?
                                         
                                         I think there are a lot to worry, especially in the near term, like misinformation, increasing
                                         
                                         use of AI for generating fake media to support a particular political party. That's one thing. But actually
                                         
                                         also seemingly benign use cases such as, you know, people faking their social media feed
                                         
                                         might have more longer term consequence in the way that people generate and consume social media
                                         
                                         content. You know, by the way, I used to think internet is the byproduct of human intelligence,
                                         
                                         but that may not be the case in the coming years
                                         
                                         because too many people use ChatGPT
                                         
    
                                         for all sorts of their writing jobs, I hear.
                                         
                                         There are even some papers that start using ChatGPT
                                         
                                         and then the authors were not diligent enough to remove
                                         
                                         when chat GPT says, oh, it's an AI model, blah, blah. At least they should have read.
                                         
                                         Oh, I shouldn't laugh. Yeah, no, it's serious. But then again, I mean, let's be real. I have a
                                         
                                         colleague who's a very good and honest person. English is not his first language.
                                         
                                         And he has told me that he uses ChatGPT to improve the grammar of the abstract for his papers.
                                         
                                         You know, that seems like a fairly benign use.
                                         
    
                                         He's written the abstract.
                                         
                                         It's just sort of like a writing coach helping him.
                                         
                                         It's not really providing new ideas.
                                         
                                         So these things can be good tools if used properly. Certainly, yeah, it could. sort of like a writing coach helping him. It's not really providing new ideas. Right.
                                         
                                         So these things can be good tools if used properly.
                                         
                                         Certainly, yeah.
                                         
                                         It could help people to learn language faster.
                                         
                                         It can help as a writing companion if used correctly.
                                         
    
                                         But it's going to have unwanted side effect on humans as well.
                                         
                                         I do wonder whether it may be okay in the end,
                                         
                                         but I do wonder personally whether it's going to somehow make measurable changes in the human
                                         
                                         capabilities of writing and reading comprehension or not in the longer term.
                                         
                                         Well, so just to wrap up, the one thing that we often like to ask our guests,
                                         
                                         Well, so just to wrap up, one thing that we often like to ask our guests, since our show is called The Joy of Why, is to talk about the emotional side of being a scientist yourself.
                                         
                                         Is there something in your research that brings you special joy?
                                         
                                         Oh, yeah.
                                         
    
                                         Great question. joy, these questions about like, are there limitations in chat GPT?
                                         
                                         And if so, why?
                                         
                                         Why does it work so well based on just reading internet text?
                                         
                                         Seeking answers to these why questions, I don't really know why, but it does give me a lot of pleasure.
                                         
                                         And perhaps that's one of the differentiating factors compared to human intelligence and chat GPT, that we ask why.
                                         
                                         Yes, we do ask why.
                                         
                                         And thank you so much for helping us understand why.
                                         
                                         We've been speaking with Yejin Choi.
                                         
    
                                         It's been delightful to have you here with us today.
                                         
                                         Thanks so much.
                                         
                                         Thank you.
                                         
                                         It was so fun.
                                         
                                         Thanks for listening. If you're enjoying The Joy of Why
                                         
                                         and you're not already subscribed, hit the subscribe or follow button where you're listening.
                                         
                                         You can also leave a review for the show. It helps people find this podcast.
                                         
                                         The Joy of Why is a podcast from Quantum Magazine, an editorially independent publication supported by the Simons Foundation.
                                         
    
                                         Funding decisions by the Simons Foundation have no influence on the selection of topics, guests, or other editorial decisions in this podcast or in Quantum Magazine.
                                         
                                         The Joy of Why is produced by PRX Productions.
                                         
                                         The production team is Caitlin Falds, Livia Brock, Genevieve Sponsler,
                                         
                                         and Merritt Jacob.
                                         
                                         The executive producer of PRX Productions
                                         
                                         is Jocelyn Gonzalez.
                                         
                                         Morgan Church and Edwin Ochoa
                                         
                                         provided additional assistance.
                                         
    
                                         From Quantum Magazine,
                                         
                                         John Rennie and Thomas Lin
                                         
                                         provided editorial guidance
                                         
                                         with support from Matt Karlstrom,
                                         
                                         Samuel Velasco, Arlene Santana,
                                         
                                         and Megan Wilcoxon. Samir Patel is Quanta's editor-in-chief. Our theme music is from APM
                                         
                                         Music. Julian Lin came up with the podcast name. The episode art is by Peter Greenwood,
                                         
                                         and our logo is by Jackie King and Christina Armitage. Special thanks to the Columbia Journalism School
                                         
    
                                         and Bert Odom-Reed at the Cornell Broadcast Studios.
                                         
                                         I'm your host, Steve Strogatz.
                                         
                                         If you have any questions or comments for us,
                                         
                                         please email us at quanta at simonsfoundation.org.
                                         
                                         Thanks for listening.
                                         
