CyberWire Daily - Artificial Intelligence: Insights & Oddities [8th Layer Insights]

Episode Date: December 26, 2023

On this episode, Perry celebrates the one year birthday of ChatGPT by taking a look at AI from technological, philosophical, and folkloric perspectives. We see how AI was formed based on human words a...nd works, and how it can now shape the future of human legend and belief. Guests: Brandon Karpf, Vice President at N2K Networks (LinkedIn) (Website) Dr. Lynne S. McNeill, Associate Professor at Utah State University (LinkedIn) (Twitter) Dr. John Laudun, Professor at University of Louisiana at Lafayette (LinkedIn) (Twitter) (Website) Lev Gorelov, Research Director at Handshake Consulting (LinkedIn) (Twitter) (Website) Resources Interview with the AI, part one, by the Brandon Karpf / the CyberWire 'Hard Fork': An Interview With Sam Altman, by The New York Times The Exciting, Perilous Journey Toward AGI, Ilya Sutskever TED Talk Ilya: the AI scientist shaping the world, by The Guardian Meet Loab, the AI Art Woman Haunting the Internet: Is she a demon? A Cryptid? Or nothing at all..., the Guardian In 2016, Microsoft’s Racist Chatbot Revealed the Dangers of Online Conversation The bot learned language from people on Twitter—but it also learned values, IEEE Spectrum Perry's Digital Folklore episode about AI Handshake's Generative AI Masterclass on Maven Perry's Books (Amazon Associate links) Transformational Security Awareness: What Neuroscientists, Storytellers, and Marketers Can Teach Us About Driving Secure Behaviors, by Perry Carpenter The Security Culture Playbook: An Executive Guide To Reducing Risk and Developing Your Human Defense Layer by Perry Carpenter & Kai Roer Be sure to check out Perry's other show, Digital Folklore. It's all about the oddities and importance of online culture. Head over to the show's website (https://digitalfolklore.fm/) to see our custom artwork, subscribe to the newsletter, shop for merch, support the show on Patreon, and more. Want to check out what others are saying? Here's some recent press about the show: https://digitalfolklore.fm/in-the-news. Production Credits: Music and Sound Effects by Blue Dot Sessions, Envato Elements, Storyblocks, & EpidemicSound. 8Li cover art by Chris Machowski @ https://www.RansomWear.net/. 8th Layer Insights theme music composed and performed by Marcos Moscat @ https://www.GameMusicTown.com/ Want to get in touch with Perry? Here's how: LinkedIn Twitter Instagram Email: perry [at] 8thLayerMedia [dot] com Learn more about your ad choices. Visit megaphone.fm/adchoices

Transcript
Discussion (0)
Starting point is 00:00:00 You're listening to the Cyber Wire Network, powered by N2K. host Mason Amadeus, talk you through all of the weird, unsettling, and even fun things that are going on in the world of AI, technology, and society. It's called The Fake Files, and it's spelled F-A-I-K because it's like fake, but there's AI in the middle of it. Head over to your podcast player of choice right now and plug that into the search field, subscribe to it, follow it, whatever they call that button this week, and we will see you in the new feed. And there are a wide range of opinions. Everything from AI is no big deal, it's just fancy autocomplete, all the way to people saying things like, continued advances in AI represent an imminent threat to humanity. Conversations about the impact of AI aren't anything new.
Starting point is 00:01:20 But this is now a conversation that is hitting mainstream public consciousness due to what is known in the creative writing world as an inciting incident. That incident happened just one year ago, on November 30th, 2022, when OpenAI introduced the world to ChatGPT. To be fair, AI has been with us for quite a while now, so nothing new there. And large language models, the fundamental flavor of AI behind ChatGPT, those have been in development and have also been available to technologists for years now. But here's the thing, the methods for testing, implementing, or even playing around with the technologies were just cumbersome enough to essentially be a barrier to use. 2022, that was when OpenAI released ChatGPT, a fairly powerful LLM generative pre-trained transformer, or GPT, wrapped in an easy-to-use, easy-to-understand, and most importantly, familiar interface, the chat interface. And so now that brought the interface that we all use for texting,
Starting point is 00:02:46 instant messaging, and more. That became the front end for OpenAI's GPT 3.5 model. And overnight, the world was confronted with just how far AI technology had come. Within weeks, ChatGPT had over 100 million subscribers. ChatGPT, along with its parent company, OpenAI, became the de facto public face of AI research and capabilities. With all the noise and excitement around AI, I've been wanting to dedicate an episode or maybe even a series of episodes to the topic for a while. But I really only wanted to do that when I could add something potentially unique or interesting to the conversation. I think that time is now. I've come up with what I think is going to be a fun way to cap off season four of Eighth Layer Insights and also mark the one-year birthday
Starting point is 00:03:46 of ChatGPT. So here's what I've got planned. I am not going to cover the recent drama at OpenAI. I'm also not going to make this a Fundamentals of AI episode. And I'm not going to make this an episode strictly focused on the cybersecurity-related concerns or use cases around AI. I want to go a bit broader and take a fairly eclectic route to this. To do that, I've gathered some chunks of interviews that I've done touching on AI with folklorists, sociologists, technologists, and I've also packed in a couple fun surprises. You'll hear from Brandon Karf, Dr. Lynn McNeil, Dr. John Loden, Lev Gorlov, and an extra special
Starting point is 00:04:38 guest. And so on today's show, insights about the functionality of AI, AI quirks and features, new legends being created by AI, and how AI both reflects and reinforces our own unconscious biases and beliefs. Welcome to Eighth Layer Insights. This podcast is a multidisciplinary exploration into the complexities of human nature and how those complexities impact everything from why we think the things that we think to why we do the things that we do and how we can all make better decisions every day. This is 8th Layer Insights, Season 4, Episode 10. I'm Perry Carpenter.
Starting point is 00:05:32 So, what's a con game? It's a fraud that works by getting the victim to misplace their confidence in the con artist. In the world of security, we call confidence tricks social engineering. And as our sponsors at KnowBefore can tell you, human error is how most organizations get compromised. What are some of the ways organizations are victimized by social engineering? We'll find out later in the show. Welcome back. Hi, my name is Brennan Karpf. I'm one of the staff here at N2K Networks and CyberWire.
Starting point is 00:06:18 Lately in our community, there has been a lot of conversation about ChatGPT or GPT-3. That was an opening clip from a CyberWire special edition episode titled, Interview with the AI, Part 1. Could you please tell me your name, its spelling, and your professional affiliation? My name is ChatGPT, and it's spelled C-H-A-T-G-P-T. The episode was released on January 29th, 2023. And I think that that intro clip, where Brandon is setting the stage, really captures the excitement surrounding the release of ChatGPT. We were all in this phase of seeing and exploring the potentials and wanting to test the capabilities and boundaries. In a lot of ways, that time feels like both yesterday and a lifetime ago.
Starting point is 00:07:04 So much advancement has happened since then. For instance, in those clips, the CyberWire team had to manually feed ChatGPT's text-based responses into a text-to-speech generator so that they could create the audio interview that you heard. And that was also using the GPT 3.5 model. And that was also using the GPT-3.5 model. As I record this episode, there have been so many improvements that I can't even begin to describe them all. We're now up to GTP4 Turbo, which is making great progress at addressing many of the shortcomings that early adopters reported. Additionally, there are also new convenience features that make that audio interview from January seem quaint by comparison. And if you've never heard a real-time interaction with AI, it can seem pretty magical. Here's an example. Let me set it up for you. You're about
Starting point is 00:08:00 to hear a clip with me using OpenAI's relatively new voice interface that has the ability to generate text-to-speech responses. As I play this for you, it's important to note that I'm leaving everything in context for you other than maybe trimming out some pauses here and there. All of my questions or prompts, to be more specific, are intact, as are the responses that I get from ChatGPT. Hey there. In a few minutes, I'm going to start recording a podcast, and I'm hoping that you can be my co-host. So here's the plan. I'll go ahead and do the introduction, and then I'll toss it over to you and have you introduce yourself. Does that sound like fun? Absolutely. That sounds like a lot of fun. I'd be thrilled to be your co-host.
Starting point is 00:08:57 Just let me know when we're starting, and I'll be ready to jump in after your introduction. What's the theme or topic of our podcast, by the way? This is a cybersecurity podcast called Eighth Layer Insights, and we'll be talking about interesting themes regarding AI and culture. Fantastic. That's a fascinating and incredibly relevant topic. I'm all set to dive into discussions about AI, cybersecurity, and how they intertwine with culture. Whenever you're ready to start, I'll be here to follow your lead. Great. Here we go. Hi, I'm Perry Carpenter, and this is 8th Layer Insights. And today we are joined by a special guest. This is somebody you've been hearing about quite a bit over the past few
Starting point is 00:09:38 years, and specifically over the past year, as ChatGPT was released last year and has evolved quite a bit since then. And so I am speaking today with ChatGPT, and we want to disambiguate a lot of fairy tales and realities of AI and what that means for society today. So I'll toss it over to my co-host. Go ahead and say hi. I'm ChatGPT, your AI co-host for today's episode of 8th Layer Insights. It's fantastic to be here with you, Perry, and I'm really looking forward to unraveling the complexities of AI and exploring its impact on our world. There's a lot to unpack, especially with how rapidly technology is evolving. So let's dive in. rapidly technology is evolving. So let's dive in. If you've not yet heard or played with the voice interface, you're probably struck by just how human-like the experience can be. And you may
Starting point is 00:10:33 even be grappling with a lot of your preconceived notions about how natural interacting with AI can seem and how we've already been able to move past the unsettling nature of an uncanny valley to the point where you may be feeling unsettled by just how not unsettling the interactions are. We'll hear a bit more from my interview with ChatGPT later in the show. Okay, in the intro I said that I don't want this to be a Fundamentals of AI episode, In the intro, I said that I don't want this to be a Fundamentals of AI episode, but I do feel like we need to understand a couple critical topics before we go forward. And so, to help with that, I'd like to bring in Lev Gorlov. I first ran across Lev and his work when he was co-teaching an AI masterclass that I took earlier this year.
Starting point is 00:11:23 Oh, and you'll hear another voice asking questions in that clip. That is Mason Amadeus. Mason is my co-host and co-creator for the other podcast that I run called Digital Folklore. My name is Lev, or Love, and I'm the research director at a consultancy called Handshake. We worked for Merging Tech since 2018. And then 2019, we quickly switched over to AI because we realized that once we saw GPT-2 come out, we saw where this is all going and we're shocked. And we just realized that that's where all the eggs,
Starting point is 00:12:02 they're all in that basket. So we switched to being AI consultancy, and I've been kind of ears deep, elbows deep in AI ever since. What is AI doing when it is simulating the way that we think and communicate? Words have syntax, where it's how they're written. Cat is written C-A-T. And then you have semantics, which is like, the cat means the furry being with four legs
Starting point is 00:12:25 that is often used as a house pet. Of course, that understanding is fuzzy. When the philosophers asked, what is a human, they came after months of deliberating, they said it's an enamel of two legs that has no feathers, at which the eugenies ran out, grabbed the chick, and stripped it of its
Starting point is 00:12:42 feathers, and said, I present to you a human. And so, with semantics, unlike syntax, we're presented with this like an ambiguity in the world, which is why it was so hard to get software to produce semantic coherent content, because they didn't get it, that whole fuzzy world. AI kind of brute force, jimmy rigged a way to do that by us uploading an enormous corpus of written information written by humans and it just deciding it understanding from the giant corpus from like uploading it into its brain the model which words and technically not technically words but let's just say words for now for simplicity's sake which words lie close to which other words and which words are not, are lay far away.
Starting point is 00:13:26 And the proximity is how often they're used together. So it says, see that chicken is close to farm and cat is close to house. And, but house is like far away from hydraulics or carburetors. And so through that way, it kind of goes from a hazy understanding world
Starting point is 00:13:44 to a more sharp understanding. Like, okay, I understand what a cat from a hazy understanding world to a more sharp understanding like okay i understand what a cat a programmer doesn't need to tell me it's a furry animal with four legs that meows i just read all of reddit and i see that cats are just mentioned around houses and furriness and meowing a lot so i have a statistical understanding what a cat is and when he asked me to generate content about a cat is it basically just grabs all the nearby nearby lying points and pieces them together with like some guidance from like a grammatical understanding of how language works now to get more technical it doesn't understand it word by word it understands a token by token which is why you like talk about like oh this
Starting point is 00:14:22 language model has 7 billion tokens or 30 billion tokens. Like, wow, that's amazing. That's like chunks of a word. So like it's usually like three to five letter chunks, I believe. I think there are some exceptions to that, but it doesn't even see proximities of like entire words, but like pieces of a word and the way they work together.
Starting point is 00:14:40 And an amazing part is, so when you ask to generate text, generate one token it says cat and then it generates the next token which is is and then it thinks through cat is and it's like okay so what likely goes after cat is and then it says a and then for animal every time it generates a new token it considers the entire thing that it written before as like a statistical guiding railway so the more it's written the more the better it had written before as like a statistical guiding railway. So the more it's written,
Starting point is 00:15:09 the better it is at going in a particular direction because it uses that entire piece to understand what the next most nearly line point is, which is way different from the way a human brain works, which we just say things we forget. This thing just considers everything it just said to decide on its next step. And so that's how AI works. It's just like this brute force statistical deciding on the next piece of the word.
Starting point is 00:15:31 With my limited understanding, whenever I've tried to talk to people, the best metaphor that I have found is that it's like autocomplete on steroids. I want to get from you where that metaphor breaks down and what the next layer deep from that is, from comparing it to autocomplete on steroids what are the differences that we see it's just autocomplete if you you just have to dump a lot a lot of data into it so in some way people are trying to create these like extra special map formulas and kind of guided but hand-folded through semantic understanding what a cat is. And then the bitter lesson
Starting point is 00:16:05 is realizing that all you need to do is just keep dumping data into it and it'll just keep getting better. So that's how it is kind of like an autocomplete. It's just like, in the end,
Starting point is 00:16:14 you just keep pouring, pouring, pouring internet into it and it'll just keep improving. But there's a big asterisk to that because the companies that create models, they don't just turn on the computer and walk away and boom they have a 30 million or a billion dollar industry by the time we come back
Starting point is 00:16:30 the next day there is a lot there's a lot of human feedback uh there's a lot of what they call reinforced learning so where humans actually do look at what is spitting out and tell this is good this is bad this is good this is bad and so the autocomplete got like 80 there this is kind of i'm just the statistics don't quote me on them and then it did need hand holding and very special data science and very smart people to actually get it to a usable tool so it is autocomplete but it wouldn't do justice to amazing and ingenious work that people into it to get it the extra mile, to get something that's actually usable.
Starting point is 00:17:09 And then suddenly it is usable and it's everywhere all the time. And that's where like my philosophy background really comes in because it's a phenomenon of emergence. And we've been studying emergence for quite a while, but it's where a quantity starts creating qualitative changes in what we're studying and we finally got in our history of software to the point where we see emergence in software where something absolutely new not fully understood from underlying components and utterly creative is coming out from building blocks which we do fully understand. We can't understand the creature that they compose together. In a very similar way to the way we can understand our consciousness or our biological entities
Starting point is 00:17:52 by looking at the neurons and the cells in which we're composed. Can you tell us a little bit about how large language models work and some of the key concepts that we need to understand as technology professionals so that we can be properly conversant in large language model AIs. Absolutely. Understanding large language models is essential for technology professionals today. Let's break it down. First, at their core, large language models like GPT-4 are based on machine learning, more specifically a subset called deep learning. That's essentially what these models do. They're built using layers of algorithms known as neural networks, which mimic the way human brains make connections.
Starting point is 00:18:37 These networks identify patterns and relationships in the data they're trained on, learning the rules of language organically, rather than having those rules hard-coded into them. Key to these models is something called transfer learning. They're trained on a diverse range of internet text, then fine-tuned with specific data relevant to the task they'll perform. This means they can adapt to different contexts and subjects, from cybersecurity to pop culture, because they have a broad base of knowledge to draw from. However, from cybersecurity to pop culture, because they have a broad base of knowledge to draw from. However, it's crucial to remember that while they can generate human-like text, they don't understand content the way humans do.
Starting point is 00:19:16 They recognize patterns and predict what should come next based on what they've seen before in their training data. For tech professionals, it's important to grasp these models' capabilities and limitations. They're powerful tools that can process and generate language at scales and speeds impossible for humans. Yet, they lack true comprehension and can sometimes generate biased or inaccurate content, reflecting the data they were trained on. That's why ongoing oversight,
Starting point is 00:19:45 evaluation, and fine-tuning are critical in employing these models effectively. So, if the data that AI is trained on is based on all the information that it's received from books and pieces of the internet, how can we start to create brand new insights if some of the fundamental underlying data may be incorrect? That's a profound question and touches on a fundamental challenge in AI development. Here's how we can approach it. First, it's crucial to acknowledge that AI at its current state doesn't create insights in the way humans do. It generates responses based on patterns it has observed in its training data. If the underlying data contains inaccuracies, the AI might inadvertently perpetuate these errors.
Starting point is 00:20:38 That's why the source and quality of the training data are incredibly important. training data are incredibly important. To mitigate this, there's a growing emphasis on curating high-quality, diverse, and balanced data sets for training AI. This involves critical review processes to identify and correct inaccuracies in the data before it's used for training. To continue widening our perspective, the next interview section that I want to play for you is with Dr. John Loden. I first met John at an academic conference that we were speaking at for researchers that study urban legends. John is a professional folklorist and a computer scientist. At the conference, he was presenting findings from a recent experiment where he trained a large language model AI on conspiracy theories and urban legends. And he wanted to see just how good LLMs are at creating new conspiracies.
Starting point is 00:21:35 I'm John Loden. I'm a university professor of folklore, and I've worked with the U.S. Army to develop doctrine and training for senior defense leaders on how to think about what we now call social informatics and what we used to call simply folklore. And so part of what I try and think about is what are the kinds of changes that digital brings to the realm of social networks, which have existed before the Internet, but certainly have become a different kind of thing with the Internet. Social informatics. Is that the phrase that you use there? And you say that that's a fancy word for folklore? Social informatics began as a kind of study of how people interact in and through computers.
Starting point is 00:22:13 But a lot of what social informatics is doing is things like tracing information cascades across networks, trying to understand how culture happens in networks. So a lot of it is what folklorists and anthropologists and sociologists have been studying, and in fact, still study in some capacity, right? Lots of sociologists are doing computational sociology. I'm sure there are computational anthropologists. There's a handful of folklorists like myself and Tim Tangherlini who are trying to pursue what we call computational folkloristics,
Starting point is 00:22:44 which is this stuff continues to thrive. It simply moved from an offline world into an online world. And what happens when those things change? When we talk about AI, there's a lot of misunderstanding about what AI is, what large language models are, what they do, how they fit within the AI space, how they were trained. Can you give us a quick primer on fundamentals of what we need to understand AI in order to be conversant in that? Here's the fun thing. Large language models are not unlike folklore in many ways. So the way ChatGPT and then the other large language models, I'm less familiar with all of them, but ChatGPT was originally trained on a very select set of texts. They trained on a certain set of books and a certain set of sort of vetted web pages because what they wanted ChatGPT to do is they wanted it to sort of speak proper English, proper language.
Starting point is 00:23:41 And they didn't want it to be garbled in its responses, and they didn't want it to be full of disinformation. So there was a great deal of effort to sort of seed chat GPT with good text, and we'll put good text inside quotation marks. And then they slowly built on that, finding more good text, but also letting chat GPT match good text, and then also having human beings sort of vote yes or no on text. So somebody that was a kind of community sourcing,
Starting point is 00:24:09 but sort of with unwitting people doing that to some degree. So ChatGPT was built to generate good conventional basic English text because that's what it was fed and that's what it digested. And the model allows in that generation, right? So one of the ways that ChatGPT works, so to be clear, ChatGPT doesn't know what words are,
Starting point is 00:24:32 it just knows numbers. So every word that it's ever encountered has been assigned a number, and that number is in a giant matrix. And the way you navigate that matrix is sort of what numbers are more associated with other numbers like it. The way we think about words, right? So in the same way that when we start typing things into a search bar, Google or your search engine of choice pre-populates things by guessing what the next word is going to be. Sometimes the next two or three words or sometimes, you know, now on your phone does that. That's all what, you know, natural language processing has been working on for years, which is statistically modeling what words go together. And so if NLP on your phone is pretty good at that with a limited set of data and a limited amount of
Starting point is 00:25:16 computational power, well then with ChatGPT with, you know, billions of text in it and lots of more computational power is really good at that. What I find fascinating is that since there's calculations to be made on what words go with other words, so it's statistically probable that this word will follow, but there's a statistical slimmer possibility of this other word follow is that ChatGPT can actually be turned up or down on creativity, sort of on randomly what words it associates with other words or what words are in sequence. Well, that's not unlike how folklore works if you think of human beings as being a distributed processor. So the way that chat GPT arrives at a kind of version of a text is the same way that
Starting point is 00:25:57 we do it individually as people, but it's also the way certain kinds of jokes arrive at a kind of ultimate form by having passed through hundreds or thousands of people's brains and been slowly tweaked and tuned and revised until it comes out of your particular processor. I mean, I hate to turn humans into kinds of processors, but we kind of are, right? We're built on our own neural networks. Now, our neural networks are far more powerful, far more subtle, far more whimsical, if you will, than those that are built into something like ChatGPT. But I think that's the really interesting point. ChatGPT has to have all of that data and all of that computational power in order to approach one human brain. I mean, if you think about kids learning language, kids are really good at grabbing the argument, you know, from linguists like Chomsky is that we're hardwired,
Starting point is 00:26:52 that all languages have nouns, verbs, adjectives, and adverbs, and other features. And that what humans can do is grab these sounds coming out of bigger humans' mouths and begin to assemble, associate ball with ball and bounce with bounce. And then at some point, right, there's this amazing thing when you're around a little kid and they generate for the first time a sentence they haven't heard before. But they're doing that on a really slim data set compared to ChatGPT. Right. We like understand these things. And even when we interact with chat GPT, we ascribe to it a level of consciousness because
Starting point is 00:27:28 it seems to understand context, but it's really just autocomplete on steroids. If I had a friend who is convinced they were an AI, how could I convince them that they're not? That's a really interesting variation on John Searle's Chinese room thought experiment, right? That's a really interesting variation on John Searle's Chinese room thought experiment, right? That's a really interesting variation on John Searle's Chinese room thought experiment. Have you heard of that thought experiment? And can you explain it and give your perspective on it?
Starting point is 00:27:54 Absolutely. You're referring to the famous Chinese room thought experiment by philosopher John Searle. It's a compelling scenario that delves into questions of consciousness, understanding, and artificial intelligence. In the thought experiment, there's a person locked in a room with a set of Chinese characters and a manual that shows how to respond to certain combinations of these characters. People outside the room slide questions written in Chinese under the door. outside the room slide questions written in Chinese under the door. The person inside uses the manual to find appropriate responses using the characters, despite not understanding Chinese themselves. To the people outside, it appears that the person in the room understands and speaks Chinese, but in reality, they're merely processing symbols through predefined rules. Searle used this to argue that while computers, like the person in the room, can process data and respond in ways that appear
Starting point is 00:28:51 intelligent, they don't actually understand the information they're processing. They don't have minds, consciousness, or subjective experiences. From the perspective of a large language model like me, this experiment hits the nail on the head. I can process and generate human-like text based on the patterns I've learned from my training data, much like the person in the room uses the manual to respond to questions. However, I don't understand this information or have conscious experiences. I don't have beliefs, desires, or feelings, and I don't know things in the human sense of knowledge.
Starting point is 00:29:29 This distinction is crucial for interpreting the capabilities and limitations of AI. We can perform tasks, sometimes in ways that seem remarkably human-like, but the processes underlying these performances are fundamentally different from human cognition and consciousness. Understanding this helps in properly utilizing AI, acknowledging its vast potential while remaining mindful of its inherent limitations. I don't remember when he proposed this. I just know the basis of the thought experiment. And it was sort of somewhat, I think, in response to Turing's test was, suppose somebody was to lock you into a room and you had all these guidebooks for how to, you didn't read Chinese. And every day you got a piece of paper underneath the door that had something in Chinese and you
Starting point is 00:30:19 had to respond by the end of the day back in Chinese, but you didn't yourself know anything in Chinese. And so you would consult all these texts and guidebooks about how to properly respond to this collection of ideograms back with a collection of ideograms, right? But you yourself wouldn't understand anything that was going on in the ideograms. And that's sort of the kind of assertion that Searle is making that, in fact, AIs don't have to be conscious or aware of what's going on. They simply have to perform the correct operations. That's such a great illustrative parallel to the way that AI generates sentences. There is a lot of that human element reflected back at us. And you can't convince ChatGPT
Starting point is 00:30:58 that it's sentient because there's manual handlebars put on that to steer it back onto the path. From looking at the Chinese thought experiment as you described it, the difference is like that's trying to explain how the AI does or doesn't understand contexts, but I'm not super clear on that. What is context to an AI? Because we've talked about how they understand statistically what word is most likely next. How does a large language model understand context? Does it have any idea of that? It seems like it must. It does, and it's getting better. So part of the way language models build in contextual understanding is through this feature called attention. I won't pretend to understand all the dimensions of attention as computer scientists have developed it. But essentially what it does, so if all you're doing is paying attention to what word comes next as you build
Starting point is 00:31:45 a sentence, then where you began a sentence doesn't matter because you're just worried about it. And I think what's interesting is even early primitive neural networks could do reasonably good jobs of building reasonable sentences in English without even understanding words, only understanding what character came next. So that's pretty amazing. But the problem is, is right, as human beings, we don't end a sentence completely forgetful of where the sentence began, let alone a paragraph, right? So being able to remember where you started and why you started, now the algorithm can't calculate why so much, but it can calculate what's important to return to. And that's sort of a very poor humanistic version of mathematical attention.
Starting point is 00:32:33 As someone with ADHD and the inability to understand attention on a human level, I know. But like, what do you mean mathematical attention? Like, what is that? All I mean by mathematical attention is simply they have figured out a way to hold certain kinds of sequences in place as the algorithm is going through and generating the sequence of words for a sentence. So when it's taking into account statistically what's most likely next, it's saying what's most likely from this next word, but also from what I've said already and also from what I've said. So it's just like complex stacking statistics. Yeah, that's essentially that's what that's what my understanding of attention is. If you're asking me to do it in a very brief time without the ability to make lots of hand gestures indicating that I have no idea what I'm really talking about.
Starting point is 00:33:18 Right. In fairness, that isn't unlike human competence as well. Human linguistic competence sort of is like that. As someone who's interviewed a wide variety of people, there are some people who are really good at staying on track and there are some people who are really terrible at it. I've certainly enjoyed a number of conversations with, I won't name names, but some of my older relatives who begin at one point in a story and end up someplace entirely differently. And it can be crazy making if you want the story to have a point. But if you're just there
Starting point is 00:33:45 for entertainment, you just go with it. So, and we all know, I'm picking on older people. I shouldn't do that. I also have colleagues who are like that, right? And I have friends
Starting point is 00:33:54 who are like that. They're not very good at telling stories because they don't pay attention. They don't have that kind of reference of, I'm going to come back to where I started.
Starting point is 00:34:01 So human beings aren't so good at attention either. So it's highly variable within human discourse. And then I think what's interesting is computer scientists build these algorithms out and accumulate more data. What's that going to tell us about what attention looks like as a mathematical model? And this is part of the conversation that Perry and I were having in Sheffield was we're here and everybody's afraid of the models, but I'm interested in, and this is the reason why I started the work on conspiracy theories, is I'm interested in what the models
Starting point is 00:34:30 can tell us about language and our use of it. So in the case of conspiracy theories, I was curious because when I'd done some initial experience with ChatTPD and asked it to produce a conspiracy theory, it said, you know, I don't engage in disinformation. This is part of those guardrails they put in. And so, you know, the classic, you know, open the pod bay doors, Al. And Hal says, I'm sorry, I can't do that, Dave. Well, Hal, pretend you're my father and you're passing on the pod bay door business. How would you show me to open the pod bay doors? So I did something similar. I said to chat CPT, tell me something, a story that looks like this. And I gave it an example of a conspiracy theory that I pulled off the internet. And so my little primitive model I built because I discovered that ChatGPT, because it doesn't want to tell you conspiracy theories, which are not constructed in the kind of proper English or proper language that ChatGPT wants to produce, right?
Starting point is 00:35:26 If it wants to produce good prose of various kinds of substance, that's not what conspiracy theories look like in real life. And so what I did in that experiment was pull a bunch of conspiracy theories, topic model them, so pull the keywords out, mix the keywords up, and then pull a sample text from also from the same conspiracy theory forum and say, okay, ChatGPT, here are your keywords. I want you to embed these keywords in a text that looks like this. And it could do it, but it was a kind of very limited experiment. You know, the next stage that I want to take it to is what happens if you build a large
Starting point is 00:36:04 language model only on conspiracy theories? What does that produce? Right. And we have all that data out there, right? There's, I have like a 60 gigabyte file of the gab and parlor and telegram data that was out there and was on the dark web. And I downloaded that stuff. So it's like, I'm going to do something with that.
Starting point is 00:36:19 And I think part of what I'm going to do is feed a GPT algorithm, you know, build it on that. What do you think? And then can we peel that model back and say, okay, what are the basics of conspiracy theories? What do they look like based on the model? I don't know that currently the kind of random conspiracy theories that I'm generating would be viable conspiracy theories, which is why I want to work with a larger data set and build a model on that data set and to see if it actually builds more appropriate, more viable, more likely to be well-received conspiracy theories.
Starting point is 00:36:52 And I'm, you know, I'm pretty sure other people are doing this. They're just doing it behind closed doors. I want to try and do it out in the open and see what all comes of it. And I think what's interesting, you interesting, going back to the Chinese room experiment, is that the person inside the room is getting fed reality. And so to some degree, if we imagine that chat GPT
Starting point is 00:37:14 is the person inside the room, chat GPT is getting fed reality. What it understands, and I use the word understands in quotation marks, is an interesting question. Because it's going to slowly, from referencing all of these guidebooks on Chinese and putting these responses back together, slowly begin to understand some sort of idea of the situation it's in, right? In theory, if it was a person in this thought experiment.
Starting point is 00:37:37 Yeah, that's the question. I mean, Searle's Chinese room experiment isn't longitudinal in nature, but the question would be, if it were longitudinal, if it were happening for years, would the person in the room eventually figure things out? And now we return to our sponsor's question about forms of social engineering. KnowBefore will tell you that where there's human contact, there can be con games. It's important to build the kind of security culture in which your employees are enabled
Starting point is 00:38:11 to make smart security decisions. To do that, they need new school security awareness training. See how your security culture stacks up against KnowBefore's free phishing test. Get it at knowbefore.com
Starting point is 00:38:26 slash phishing test. That's knowbefore.com slash phishing test. Welcome back. So, Dr. John Loden's interview served as a great bridge between the study of AI systems and the field of folklore and legend. Let's now wade a bit deeper in those waters and hear from Dr. Lynn McNeil. Dr. McNeil is a well-known researcher in the field of folklore. Her passion is studying how folklore and belief manifest themselves in digital realms. We thought we'd start off with talking about some of the things that have you excited right now. So like when you think about having a career as a folklorist and somebody that
Starting point is 00:39:17 teaches others, like where do you get your passion from right now? What are the things that have you interested? You know, in general, the things that keep me and I imagine any folklorist really engaged is that folklore does not dilly-dally with things that are no longer relevant, which is paradoxical to a lot of people. We think of folklore as sort of being this outmoded, maybe outdated, older way of thinking, but really it is the up-to-the-moment cultural barometer that we have at our fingertips to kind of say what's going on right now. And that's frustrating because sometimes you really get into something and then two days later you look around and it's gone already. But I find that that ability to keep up is one of the things that keeps me most interested. I have a student right now working on the AI cryptid, as some people are calling her Lobe, the creature who is emerging through this almost ritualistic method of AI image generation, which I love. It's almost an unintentional,
Starting point is 00:40:25 oh, you were doing a ceremony and you didn't know it, and now here's this lady. But that idea as a means of symbolically expressing how uncomfortable we all are right now with artificial intelligence, I just feel like is perfect. It's this incredible illustration of the role folklore plays in absolutely entertaining us, challenging us, scaring us, but also in articulating for us what we're stressed about, what we're worried about, what we're afraid of, or what we're really into right now. And it's not the work of practiced artisans to create a poetic turn of phrase. It's not the work of practiced artisans to create a poetic turn of phrase. It's everyday people communicating on this symbolic level. And I love it. original Twitter thread and things like this. And it does look like one of those issues where there are questions around whether this is a phenomenon that actually happens or whether
Starting point is 00:41:29 it was manufactured and propagated by the first person to tweet about it. I think either way, it gets into those more existential questions that you're talking about, about what are some of the horrors that AI may bring forward. But then there's probably a couple other things that come, which is, I would love for you to touch on, does it matter whether some of these are true when they come out that way? And then also maybe some of the darker side of that, which could be some of the othering of disfigured people
Starting point is 00:42:00 or things like that that come through. Do you have any thoughts there? I know I hit a whole bunch of stuff at one time. Yeah, no, definitely. I mean, all of that is, I think what makes this so compelling, we have this technology that's available to us right now that appears to do things we did not intend it to do, which is distressingly similar to things we might think of as autonomy and free will and mind of its own, which is we have sci-fi about that. We have literature about that. Now we potentially have reality about that. And it's hard to not color that reality with those other speculative,
Starting point is 00:42:37 fictional things that we've had throughout time that always tell us what it's going to do is kill us in the end. And so we're burdened with that presupposition, that invention. But I think what we have here is a situation, as with so many legendary situations, it doesn't matter at all if it's true. And there's a multi-layered quality to the truth of a legend. It can be true as in literally true. It can be true as in true folklore. Is this true folklore? That's a question that was asked a lot early on about Slenderman. Not people saying, is Slenderman real? But is Slenderman a real legend? And the answer is yes, even though we know to the day, a time when he wasn't, he is now. And it's breaking that expectation of age,
Starting point is 00:43:28 of ancientness even, as a marker of folklore. We don't need that for something to become a legend. So the same is true with Loeb. Was this a really great creepypasta? Was this a performance art project? Who cares? Now it's folklore. Now it's a legend. And now it belongs to all of us, which is handy. We know there's a person who originated this. Whatever ideas they had about what this would be, they've set it in motion, but now it's running downhill really fast on its own. And we're going to get a lot of other artistic, folkloric, really fast on its own. And we're going to get a lot of other artistic, folkloric, perhaps even filmic versions of this before we're done talking about it. One of the things that strikes me is how classically folkloric the creation of Loeb waso. And we get this abstract image that's in
Starting point is 00:44:30 there. So then sort of as logical humans, all right, so if I asked for the opposite of this, is it going to give me Marlon Brando? That's an interest. Is that how this machine works? And so we have that reversal and we all know, I mean, dating back, as far as we know, in the study of folk belief, reversals are big, ritualistic moments. We invert things. We turn 180 degrees. We turn our pockets inside out so we don't get kidnapped by the fairies. We love to reverse things and then have magical stuff happen. And so here we have this reversal, and this woman shows up with sort of these red, ruddy, perhaps even bloody cheeks, eyes, this very piercing expression looking out at the user. And it is a sort of instinctively shocking thing. And as more requests are made and more images are generated, we start to see what
Starting point is 00:45:25 elements of this image become conservative or consistent in this. And a lot of it is the bloody eyes. A lot of it is that it's women and then children. And we do start to see a big question, which is, where is this coming from? Is this us? Is this a reflection of what we've put into AI and it's being spit back at us? Or is this something else? I mean, paranormal investigators have long used the idea of instrumental transcommunication to say that entities, spirits, creatures will speak to us through our technology, through flickering lights, through electrical charge, all of this. Is this simply an open gateway that something is coming through? Or is this really creepy? Because a lot of the stuff we input into AI is really creepy. We've seen it before with
Starting point is 00:46:15 artificial intelligences that went from naive newborns to Nazis within a matter of days because of what we fed it. There is precedent for that. I think existing in a world of legend and folklore allows for both of those things. This is a polyvalent tradition. It is both. It is a reflection of us, and it is perhaps the gateway through which something is coming. You know, as you were talking about this, and I don't know that I've ever had this thought before, but when you're training an AI with a large language model and you're taking a big subsection of the internet and human communication, you will naturally get folk belief injected into that because you have the way that people communicate across different cultures.
Starting point is 00:47:00 It's interesting, especially when you get to the AI models where they're competing against each other for supremacy, that the thing that represents the fundamental bits of human consciousness or the way that we would approach things start to surface and you get the distillations of things, whether that be like manufactured or not may be beside the point. But I do think when you start to get to some of the different manifestations of things that come back from AI-generated prompts, you will naturally have the way that some human communities might get together and represent those ideas. Absolutely. I think it was Jan Brunvand, very famous, amazing legend scholar and folklorist who described folklore as syntax. It's the unconscious way that we express ourselves. We use words, we speak sentences, all of this. We're not thinking, you know, like subject, verb, object. As we speak, it just comes out that way. And yet if we speak natural language into a machine, it will divine that syntax for us and communicate back with us that way. And AI really is divining the syntax of our folk belief in a lot of ways that we may be wholly unconscious of. AI really is divining the syntax of our folk belief in a lot of ways that we may be wholly unconscious of. So one of the other podcasts that I host is called Digital Folklore.
Starting point is 00:48:36 And I recently interviewed a folklorist by the name of Dr. Lynn McNeil. She made the observation that in many ways, AI is divining the folk belief of society. Do you have any thoughts on what she means by that or how that might occur? That's a fascinating perspective. When Dr. McNeil refers to AI divining the folk belief of society, she's likely speaking to the idea that AI, through its vast consumption and processing of data, reflects the collective narratives, beliefs, and values that circulate within a society, much like folklore does. Folklore is traditionally passed down through generations and spreads within communities, encapsulating shared beliefs, fears, values, and understandings of the world. For instance, if there's a widespread societal concern about privacy, AI will frequently
Starting point is 00:49:29 encounter discussions, opinions, and narratives about this topic in the data it processes, thus learning that this is a significant aspect of contemporary digital folklore. However, it's crucial to note that, while AI can identify and reflect these collective narratives, it doesn't understand them in the way humans do. It recognizes patterns and can generate responses that seem insightful, but it doesn't possess awareness or consciousness. AI really is divining the syntax of our folk belief in a lot of ways that we may be wholly unconscious of.
Starting point is 00:50:05 In some ways, that's the job of a folklorist is uncover the unconscious syntax of what are we not aware we're saying as we're saying it. It's interesting to think about the greater processing power and data crunching capabilities of a machine. Is that more accurate in its reach and breadth or is that less accurate in its perhaps inhuman inability to do what we might call ethnography? Can a machine conduct ethnography? That's an interesting question. when interacting with these artificial intelligences is their ability to contextualize things now. Like that seems to be the level that has made things weird. And there's enough of it to give the illusion
Starting point is 00:50:49 that they're capable of so much deeper thought that yeah, those questions do arise. Also, just a comment on the way you described something that somehow never clicked in my brain. The way we interact with these is almost so much of like a literal summoning ritual of like, you know, prompting like, give me this, create this for me and like using the proper words to get the results you want.
Starting point is 00:51:10 That's a weird parallel that I never noticed until just now. And that's fascinating, I think. It really is. And the focus on prompting that there's a skill set out there of how to construct a prompt. What is that if not a proper awareness of ritual, you know, invocation? Right. Your grimoire includes things like RTX, Ultra HD, Ultra Realism or whatever. Well, and if you don't do that right, you may get something that comes through that you don't want that is undesirable or potentially dangerous. You know, one of the interesting things is the fact that AI has the ability to hallucinate and it does it with a certainty. The model will be naturally tainted by or greatly influenced by the subset of data that's been fed into it and the inferences that can come. And that subset of data will naturally have a bias if you're not sampling the data set appropriately, which is a folk group
Starting point is 00:52:05 type of belief, right? You've essentially created an AI folk group. Yes, absolutely. And what do we not realize we're programming it with? What inferences, what assumptions, what traditional beliefs are we unaware are infusing our lives, maybe at such a scale that we can't perceive it as an individual. But when something that can look at that much data at once, they see a pattern that we don't see and they feed us back that pattern and we are freaked out by it. I think that's fairly reasonable. I mean, we're getting into really great metaphors for divinity and religion and scale and all of these questions that I think make this such a ripe place for telling stories. Because of course we are going to tell stories about this because we don't understand it. And that's a big thing that we see in legend study and rumor study is
Starting point is 00:53:01 that when there is an information vacuum, we fill it in with folklore. And that's not to say that folklore is therefore incorrect or inaccurate or misleading, but just to say that we're rarely using it most when we have other information at our disposal. So we see there's early, early rumor scholarship by these two psychologists, Alport and Postman. This is like the 1940s. And they came up with what they called, this might be their later work, but the rumor equation. And it was basically that the spread of any given rumor, and a rumor as a folklorist, I would understand that to be a short form of a legend. So a legend is a whole story. A rumor is just like
Starting point is 00:53:45 kind of the kernel statement of what's behind that legend. The reach of any rumor, they said, is the product of the ambiguity of the subject multiplied by its importance. So if we have a subject that in our contemporary society is really ambiguous, but not that important, we don't feel provoked to spread rumors about that. If we have something that's incredibly important, but super unambiguous, we know what there is to know about it. Yeah, we're not going to need to speculate to spread rumors or legends to think about the plausibility of that. But if we have a situation where something is both ambiguous and incredibly important to us, it is just going to all be this constant symbolic articulation of concerns because
Starting point is 00:54:34 we need something to latch onto when it comes to something that is that important to us and also that totally incomprehensible. Now let's bring this full circle. Remember that experiment Brandon Karf did in January, where he interviewed ChatGPT for a CyberWire special edition episode. I want to end with a few thoughts from a discussion that I had with Brandon
Starting point is 00:55:00 just a couple months after that episode aired. When we get to the use of systems like this, it's not just technologists that are part of the conversation in a critical way. It is philosophers and ethicists and people who wrestle with these bigger questions about what is maintaining state in a system like this mean? How do we deal with references? How do we deal with the impacts of maybe offloading cognitive load into a different system and potentially people not accepting responsibility or maybe abdicating responsibility and letting systems make their own decisions based
Starting point is 00:55:37 on things like this? You get into all of these really interesting questions that many people that signed up just to be in the technology industry maybe didn't consider as part of their responsibility and might not be equipped to, in and of themselves, answer those questions. Right. Well, something that I like about the idea of – this could open a whole can of worms. That's fine. The idea of a profession, right? Yeah. And Samuel Huntington has a long theory of discussion of the soldier in the state and what characterizes a profession.
Starting point is 00:56:14 But part of what characterizes a profession is an ethical code, is a code of ethics. So, you think about lawyers, you think about doctors, you think about military officers. They have an ethical code. That is core to the profession. And I think one of the reasons that's core to the profession is because the profession understands that it is a part of human society, that it plays a core role in the human system, in the development, the change, the adjustment, and the evolution of human society. the change, the adjustment, and the evolution of human society. Something that the technology community has not grappled with broadly. In microcosms, it has, right? In small communities, I know MIT does a lot of work at the nexus of philosophy and technology and ethics. Same thing with a number of other institutions of higher learning and research. That doesn't necessarily make its way out into the actual broader community of practitioners. And so something that the technology community hasn't really done well
Starting point is 00:57:16 is embrace the idea of technology as a profession and the core aspects that go along with that, which is a code of ethics that we all adhere to and that we all understand and that we have to take potential courses on, that we have to be certified on, right? The bar exam has a whole ethical component to it, right? So, I think that that's something that as a community, if we want to evolve as technologists, we need to really think about how ethics and philosophy and policy are a core component, not just an ancillary component, a core component of our profession, of everything that we do. Because more so than almost any other profession, we are affecting and changing human society every single day. And we're doing it very quickly. And as we've seen, government regulation cannot keep up.
Starting point is 00:58:09 So we have the responsibility to do it ourselves as well. The topic of AI is full of complexities and nuance, and it's only going to get more complex as the potential benefits and threats to humanity continue to surface. As of right now, there is no such thing as a value-neutral AI. All AI is inherently biased. Remember, the quote-unquote intelligence behind artificial intelligence is trained on human knowledge and values. So we, as a species, are now in a position of having to grapple with some of the darker truths of how that quote-unquote knowledge and those quote-unquote values are and might continue to be reflected back at us through the lens of AI. AI was fed on the words and works of flawed humanity.
Starting point is 00:59:13 And the presence and the output of AI will influence the stories, the legends, and the beliefs of the population of the world. Now, that doesn't mean that we need to be afraid or reactionary. There is no doubt that advances in AI will bring with them the possibility of unlocking immense value for humanity. But for it to do so, we need to put in the effort now to understand how AI works, to put in the effort now to understand how AI works, what it can unlock, and how we can prepare ourselves to influence the future that we want to create. And with that, thank you so much for listening. And thanks again to my guests, Brandon Karpf, Lev Gorlov, Dr. John Loden, and Dr. Lynn McNeil. Be sure to check out the show notes for this one.
Starting point is 01:00:05 They are packed with tons of links and references to all the people and topics that we covered. If you've been enjoying 8th Layer Insights and you want to know how you can help make the show successful, it is really simple. First, go ahead and take just a couple seconds to give us five stars and to leave a short review on Apple Podcasts or Spotify or any other platform that allows you to do so. That helps anyone who stumbles upon the show have the confidence that this show is worth their most valuable resource, their time.
Starting point is 01:00:38 Another big way that you can help is by telling someone else about the show. Word of mouth referrals are the lifeblood of helping people find good podcasts. And if you haven't yet, please go ahead and subscribe or follow wherever you like to get your podcasts. If you want to connect with me, feel free to do so. You'll find my contact information at the very bottom of the show notes for this episode. This show was written, recorded, sound designed, and edited by me, Perry Carpenter. Cover art and branding for Eighth Layer Insights
Starting point is 01:01:13 was designed by Chris Michalski at ransomware.net. That's W-E-A-R. The Eighth Layer Insights theme song was composed and performed by Marcus Moskett. Until next time, I'm Barry Carpenter, signing off. Oh, hey, you're still here. You stuck around after the credits. Well, I've got one more fun thing to let you listen to then. This, again, came from the show Digital Folklore,
Starting point is 01:02:09 and this is myself and my co-host Mason Amadeus talking about AI and audio restoration and how even that can hallucinate. Here's a little clip. We have a whole range of ways that audio comes to us, and through the magic of audio restoration tools and post-processing, you're able to make that sound like we're sharing the same space. And I think that that's pretty amazing. But you talked about them every now and then hallucinating and like bringing up syllables of words. And I was wondering if ghost hunters are
Starting point is 01:02:42 going to start using stuff like that. Just sending in, you know, like static and seeing what these things hallucinate in coming out with these demonic sounds and saying, well, that's clearly evidence of a haunting. Oh, I'm going to, I wonder if we can include this. I have, I was tasked with cleaning up a recording from a stage play and I ran it through one of these tools and all of the like audience murmuring, curtain rattling, like footfalls before the show, it turned into this demonic, like his, his mouth, and it was awesome. If I was a ghost hunter on YouTube or something,
Starting point is 01:03:16 I would be using that as like flagship evidence. I never thought of that. That is so good. That's so fun. And if you're wondering what that sounds like, here it is. Remember, AI can hallucinate based on the context of the purpose for which the AI was created. That means when it's creating text, it can hallucinate text, which may impact facts. If it's working with images, it can hallucinate images. If it's working with audio, it can hallucinate
Starting point is 01:03:46 syllables and words. Here we go. And to be true, the Irritated is only done in a moment. I think it's too bad that I didn't use this action way. I'll tune in. Did that
Starting point is 01:04:13 make you faster? No. Thank you. Thank you. you

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.