Your Undivided Attention - Esther Perel on Artificial Intimacy (rerun)

Episode Date: September 6, 2024

[This episode originally aired on August 17, 2023] For all the talk about AI, we rarely hear about how it will change our relationships. As we swipe to find love and consult chatbot therapists, acclai...med psychotherapist and relationship expert Esther Perel warns that there’s another harmful “AI” on the rise — Artificial Intimacy — and how it is depriving us of real connection. Tristan and Esther discuss how depending on algorithms can fuel alienation, and then imagine how we might design technology to strengthen our social bonds.RECOMMENDED MEDIA Mating in Captivity by Esther PerelEsther's debut work on the intricacies behind modern relationships, and the dichotomy of domesticity and sexual desireThe State of Affairs by Esther PerelEsther takes a look at modern relationships through the lens of infidelityWhere Should We Begin? with Esther PerelListen in as real couples in search of help bare the raw and profound details of their storiesHow’s Work? with Esther PerelEsther’s podcast that focuses on the hard conversations we're afraid to have at work Lars and the Real Girl (2007)A young man strikes up an unconventional relationship with a doll he finds on the internetHer (2013)In a near future, a lonely writer develops an unlikely relationship with an operating system designed to meet his every needRECOMMENDED YUA EPISODESBig Food, Big Tech and Big AI with Michael MossThe AI DilemmaThe Three Rules of Humane TechDigital Democracy is Within Reach with Audrey Tang CORRECTION: Esther refers to the 2007 film Lars and the Real Doll. The title of the film is Lars and the Real Girl. Your Undivided Attention is produced by the Center for Humane Technology. Follow us on Twitter: @HumaneTech_

Transcript
Discussion (0)
Starting point is 00:00:00 Hey everyone, this is Tristan. Before we start, we're gearing up for a new Ask Us Anything episode. If there are any questions you'd like to ask me, Aza, or Daniel, we want to hear from you. All you need to do is record your question on your phone and send it to Undivided at HumaneTech.com. That's Undivided at HumaneTech.com. So today we're bringing back one of my favorite conversations on your undivided attention, and that's with my friend Esther Perel, the brilliant psychotherapist who counsels couples in real life, and on her podcast, where should we begin?
Starting point is 00:00:31 And in this episode, we get to explore the world of artificial intimacy, which Astaire calls the other AI. And even since last year, these kinds of unsettling technologies have grown by leaps and bounds, from holographic AI home companions to an AI wearable friend that you hang around your neck like a pendant, to AI chatbots that market themselves to be therapists for kids. There's so much at stake with getting this right as AI keeps advancing. So I hope you enjoy this conversation. and take it away.
Starting point is 00:01:03 I go from one technology conference to another. I can tell you the word relationship is never mentioned. Never mentioned. There are moonshots about everything, environment, education, health transportation, you name it. Nobody ever is looking at a moonshot for relationships. And yet all these technologies are profoundly affecting how we relate to each.
Starting point is 00:01:28 other, how we relate to ourselves. That's famous psychotherapist Esther Perel, who's the New York Times bestselling author of the books, The State of Affairs, and Mating in Captivity. She's had celebrated TED Talks that have garnered more than 20 million views, and is also the host of the two popular podcasts, Where Should We Begin and How's Work? And Astaire and I have been going back and forth for the last 10 years about how technology has been distorting the way that human beings form relationships and distorting the way that we gain or lose intimacy. And what we're about to get into is how do we deal with technology that is distorting our relationships and our intimacy?
Starting point is 00:02:05 What would it look like for technology to foster meaningful connection and meaningful intimacy and distinguish between the fake kind and the real kind? Who are we to say what fake and real actually is? So why would we do an episode just about relationships? In the AI dilemma talk that Aza and I gave, we talked about how loneliness is an national security risk. As you see in the United States the number of shooters and shootings going up and innocent people dead from these kinds of events, are the people that are doing these shootings in healthy relationships? Do they feel deep belonging? Or do they feel mass alienation? How many of them
Starting point is 00:02:46 had healthy romantic relationships? Oftentimes when we talk about the future of technology, we talk about how AI can solve climate change and give people mental health benefits, absent the word, relationships. How does the ground underneath our feet that depends on this very deep interconnected structure of relationships, how does that help us answer questions like loneliness? Loneliness isn't some disease to be cured by having a person interact with the chatbot. It's how does technology sear us towards the kinds of things that sow the deep interconnected relationships that create a healthy society? Can you have social trust in a society if you don't have relationships? Now, this isn't just some nostalgia for a bygone era.
Starting point is 00:03:28 We have to figure out and celebrate all the things that the virtualization of our lives by AI, by technology, by smartphones, by being able to connect with each other around the world and different time zones, we need to be able to celebrate the good things that come from that. But we also have to live in right relationship between how those things land in a body, waking up, opening your eyes, taking a deep breath, seeing the sunshine, and living the day. Esther Perel, welcome to your undivided attention. Thank you. It's my pleasure to be here. So it's kind of hard to have this conversation in almost an artificially formal way,
Starting point is 00:04:04 because you and I know each other for quite some time. I think we met, I believe it was 2013. I remember knowing of your work and seeing you at a conference, and I remember that when we talked for the first time, and I kind of walked you through, I was a design ethicist at Google at the time. I was first starting this work inside of Google, and I was feeling trapped. And I remember walking in. through a version of that slide deck that I had spread internally and the race to the bottom of the brainstem
Starting point is 00:04:28 and your response was we've been completely hijacked and I started using that word hijacked in our work how technology is hijacking these core psychological vulnerabilities but maybe from your perspective I'm just curious if you want to reflect on
Starting point is 00:04:41 how we got to meet specifically maybe a moment that I know I was in your apartment I think after that so we met twice for the first time so to speak first time I was at this conference and indeed you open my eyes
Starting point is 00:04:57 it was like in my mind you said this is not inevitable this is actually controlled by particular agents that have an invested interest and that's when I just felt like yes I am a pawn
Starting point is 00:05:10 in a system that I don't even know I'm in so that was the first reaction and then the second time when we met I remember that I asked you basically to show me your Tinger fee Tinder, yes. Yes, and I wanted to see, you know, what you write and what people say, and I just thought that it was the most unerotic and unpoetic.
Starting point is 00:05:31 Thank you for that very, very kind compliment. But I thought, maybe I can suggest a little bit. It's like, I thought, it's so deflated and empty of energy. How can people actually really fall for each other or become curious of each other? And I thought, that was not just your doing. See, at first I thought, oh, you know, it's you and your experience. But then I thought, no, there is something in the medium that really takes the juice out of it as well. And it's a combination of things.
Starting point is 00:06:00 I'm not saying that people can't be utterly creative and inspired on an app, but that there is something going on that is changing human relationships, changing how we meet, how we connect, how we date, and how we break up. Yep. Well, totally. And I remember, and thank you for not just blaming me in my lack of juice and tend to. gender communication. That was an earlier time in our lives. But I think that's what we're here to talk about today, which is artificial intimacy. How is technology mediating and changing the form of relationships, you know, the loneliness crisis, fewer and fewer people actually having sexual relationships and deep connection and feeling connected? And actually clarifying the language,
Starting point is 00:06:42 the thing that I've always appreciated about your work is it comes with a kind of subterranean deeper look at what are the structures underneath the way that we tend to talk about it. Because The way we tend to talk about the breakdown of relationships and intimacy, it's like we're missing the vocabulary. And the thing I've always appreciated about talking to you is we get to go underneath that vocabulary to some of the underlying dynamics. And just to kind of link it to our work, the premise of humane technology is it comes with a deeper ergonomic understanding of what it means to be human so that technology can fit that ergonomics better. I call it the other AI, by the way. Yeah, say something about that. Yes, I mean, I call it the other AI, the rise of artificial intimacy.
Starting point is 00:07:24 And what I'm interested in is what does it mean for us? And it started because I got a note from someone who said, have you seen this? And it was somebody who basically wanted to have a session with me after he broke up and was wanting to understand what had happened in his relationship. And he couldn't get an appointment. And I wasn't available. So he decided to create me. and he created the A.I. Esther, and A.I. Esther was available. She didn't have a booked up calendar. A.I. Esther was, you know, pure. She didn't have anything interfering in her life to be there for him. And I just thought, this is amazing. And he thinks that I have fundamentally helped him and that his experience with talking with me, i.e. robot me, was really illuminating and gave him tremendous peace, quote.
Starting point is 00:08:17 And I just thought, wow, I don't even have to be there. In the past, I would have said you just have to fantasize about someone and make them be really present. Now you can actually go a step further and have your fantasy create that person and then ground it in their corpus of knowledge and body of work. And I thought, this is the other AI. Now, what experience did he actually have with me?
Starting point is 00:08:44 So just to be clear, because we've been talking about the rise of large language models and generative AI and what this is doing. So someone built a large language model, built on a corpus of what everything that you've said on your podcast, all the interviews we've ever given, and created this virtual ester who's available to everyone all the time to provide relationship advice. And this is kind of the same as this sort of seduction, this sort of deal with the devil trade we get with technology overall, right? We get this infinite access to always on swiping of the infinite pool of all human beings on dating apps. We get the infinite access to the
Starting point is 00:09:17 best people that we want to talk to you all the time. But it's always in this virtual form. And in this, I think one of the things I wanted to explore with you today is on the one hand, people will say, isn't it great that now everyone has access to you? But then the other hand, they'll know on some level that they've been interacting with some virtual agent and maybe the part of them after the session is over feels a little bit empty because they still don't have a connection with a real person. What is the trade that we're making with the virtual? But I'm just curious, as you had this new agent show up without your consent, because this person did this like without asking you, right? Oh, no, no, no. I had nothing to do with it. It was all done when I
Starting point is 00:09:55 arrived. So what did this bring up for you? I mean, you know, a part of me was flattered. A part of me was curious. A part of me had deep ethical concerns. You know, I went in many, many directions. A part of me felt violated. But, you know, on the other hand, somebody said to me, but your books have done the same. What's the difference? And people have quoted you from your books, and they have misquoted you,
Starting point is 00:10:25 and at least the butt quotes you accurately. But the butt has no soul. The butt regurgitates what I have said elsewhere. And as a clinician, our practice is highly nuanced, highly relational, and highly contextual. The art of therapy is not, to tell you what I just told the person just before you. It's to actually
Starting point is 00:10:46 say something that is completely in the moment and probably very different than what you may have anticipated so that I don't regurgitate I'm not just an amalgamation of everything that has been accumulated and summarized and organized, which is
Starting point is 00:11:02 important. It's not like it doesn't have its importance, but this connection is multidimensional, it is layered, it is embodied, it is embodied, It's the other thing. It's fundamentally embodied. It is experienced through the senses. And when you flatten it, you reduce the nuance
Starting point is 00:11:22 and you reduce the connectivity of the human experience. The other thing I think that I did also see is, you know, many of our relational issues are complex problems, they're dilemmas. They don't necessarily have a solution. There are paradoxes that we need manage. Technology believes that everything has a solution. You can reduce things to binaries and find an answer. And that's not really what happens in our existence. So it is very
Starting point is 00:11:56 useful for some aspects of our mental health and our relational life, but it doesn't touch some of the other more profound aspects. So what are some examples of that to ground that for listeners, of the dilemmas and complexity that are not reduced to answers? Shall I stay or shall I go? Where am I at in my relationship? Shall I have another child? I say, yay, you say nay.
Starting point is 00:12:22 Shall I move? Shall we move? Shall I change my career? Shall I move my parents into my house to take care of them because they are not well? You know, values questions, moral dilemmas, existential questions that are prevalent in clinical work. I mean, there are issues that have.
Starting point is 00:12:42 profoundly to do with pathology and mental illness, but there are many other things that have to do with emotional well-being and with values. And those things, they're not an answer. They're about holding the contradiction, you know, they're about grief, because whichever choice you're going to make, you're going to have to mourn something else. And that's different than an answer with a solution that wraps a nice bow around it. Here's what you need to do. It's not like the next song you need to listen to, or the directions you should take to go the fastest way, you know, to a place, or the movie you should watch tonight because you don't know what you want to see. It's very different.
Starting point is 00:13:20 This is about how will this affect my life, the people that I'm connected to, et cetera, et cetera. So you're much more systemic. Yeah, the very premise of technology is an individual experience often. I mean, it's not like when that guy built that virtual ester version of you. He didn't, I mean, he could build it as a couple's chat agent, by the way, right? he could say multiple people could be talking to the virtual version of you. But one of the problems that you and I both, I think, deeply agree on is the individualization of the experience. I mean, something that's been bothering me, actually, is the idea that mental health is just
Starting point is 00:13:53 this individual thing. We just need to have these interventions for an individual when it's so much of it is about this deeper social connection, belonging, non-alienation experience. And our technology is not wired for that. The designers at Apple, who design the interface for an iPhone aren't designing it for shared experiences. Social media is not being designed for shared experiences. Tinder isn't designing for group experiences or, you know, which groups do we want to go out to? Where can I go out to be with a community tonight? It's all designed for this individual commodified experience. But, you know, I think what's interesting is what it is doing is that it is changing our expectations of intimacy. You know, my AI creator, he thinks he had a
Starting point is 00:14:37 fantastic experience, as do many people who have AI assisted therapy. They think that they have experienced intimacy. But when you actually look at a qualitative experience, what they experienced is a lot less. It's the same thing if you have bad food or fast food or you get adjusted to thinking it's food. So you lower the expectations and then you tell people, this is good intimacy. So, you know, Sherry Turkle and Todd Essex are two people who've written quite extensively on artificial intimacy when it comes to telehealth. And the point that they make is that when people get mental health treatment online, they often will describe themselves as satisfied. And they say that it would be as good as if they had been there in person. But qualitatively,
Starting point is 00:15:30 the result is not as good, which means that patients will be satisfied. when they lower their expectations, for what? For empathy, for intimacy, for connection, for feeling understood, and they lower their expectations to match what the technology can provide. So they receive less effective care, but they don't recognize the difference. I feel like I'm saying the obvious. That's one thing honestly that bothers me about the space is it makes it hard to talk about. Everyone kind of sounds like we're speaking in these cliches,
Starting point is 00:16:01 and it's hard to get underneath that to what's really wrong with all this. So if I, but I want to go with what's really wrong, I will just share observations. And then you can decide if you think that's problematic or not. You know, you would think, I mean, the way I began one day to think of it is I'm living a kind of assisted living, but prematurely. I'm being assisted by a host of predictive technologies that are basically saying, you don't have to know, I will know for. you. I'll recommend the next song to listen to, who to date, where to eat, you know. And you would think that that should actually make us feel more confident, more at ease, because I'm neutralizing the unpredictable, the unknown. But you know, l'inconu, as they say in a big word in French,
Starting point is 00:16:56 the unknown, is it actually demands that you interact with it on a daily basis a little bit. The big unknown of course is death, but there is plenty of others in our life. And when you erase all of this, you actually make people more anxious, more unsure, filled with self-doubt, constantly wondering, you know, is there better? Is there more? Am I enough? And those are psychological questions that I think are directly related to the increased virtualization and technologies and I put it all in one piece, is that the rubbing, the living in close proximity with the messiness of another person that helps you know who you are in the presence of others, that's the piece that we are noticing in our work at this moment. Every therapist I teach and I supervise, we are having that very same conversation.
Starting point is 00:17:55 You know, so artificial intimacy with the butt isn't actually what concerns me. What concerns me is how the digitally facilitated connections are lowering our expectations and our competence in the intimacy between humans. And that I do think that all of that makes us less able to be with people who challenge us. From the political to the personal to the familial, we polarize much faster than we, you know, we polarize much faster
Starting point is 00:18:29 than we ever have because I shouldn't have to be uncomfortable. Where did we get that idea? Maybe discomfort is actually a major piece of life and you learn to deal with those discomforts. You know, you talk about delayed gratification or having all your needs met in the moment. We used to think that maturity was in part
Starting point is 00:18:51 learning how to delay gratification. Maturity is about holding ambivalence and understanding that complex issues, live with inherent contradictions. Ambivalence is about knowing that you have certain feelings for a person and the opposite feelings for that person, and they coexist. And not that you have to get rid of one side of it. So it's a very interesting new thinking about then what is development?
Starting point is 00:19:18 What is psychological development look like at this moment? Well, technology certainly does not feel like it is driving us towards greater development, but more anti-development towards infantilization, instant gratification, lowered expectations, inability to deal with that complexity. That's one of the experiences
Starting point is 00:19:38 that take place. At the same time, there's a lot of other uses from, look, to be really clear, this is not a critique of AI in and of itself or even of the virtualization of things.
Starting point is 00:19:51 I work, we are here, you and I because of it. This is the conversation we're able to have. We were able to get through the pandemic, thanks to it. You know, people listen to my podcast. Where should we begin, thanks to it?
Starting point is 00:20:03 I mean, I mean it. Let's not vilify the virtual. You know, this is not that. But there is something about understanding what it means to do it responsibly. So, yes, there is the help that I can get from it as a therapist in terms of progress notes, in terms of tracking questionnaires, in terms of symptom compliance, in terms of even just having an assistant, you know, a virtual assistant. but more than a person.
Starting point is 00:20:30 All of these things are really useful. And I think that the bot can ask you to do your breathing exercises, to help you track your thoughts, to help you map your feelings. But there is something else that has to do with differential diagnosis, that has to do with treatment conceptualization, that has to do with the big picture insights that one has when you work, and that has to do with the highly relational aspect. of our work. And that are the pieces that we are not there yet and that need to be done
Starting point is 00:21:04 very incrementally and very gradually. That's it. It's not about can we replace people or not replace people. It's that we need to find waste for the experience of mutuality and reciprocity that is part of an intimate experience. In reality, I mean, one of the main things we teach children is that when you are little, it's all about you. But part of growing up is becoming aware of the presence of others, the needs of others, the humanity of others, the impact of your behavior on others, it's that you become a relational creature. If it's about just having my need satisfied and I am being tracked and you are always there for me and you don't have any bad days and you don't forget anything, what exactly are we saying about relationships?
Starting point is 00:21:51 You know, that's the piece is that it only will make us more anxious and more unable to deal with the unexpected and the unpredictabilities and the messiness of life. Part of why we are experiencing more and more anxiety is because we are having less and less spaces in which to experience the friction, the obstacles, the failures that help us know who we are and who we are not. You know, why are people having less sex?
Starting point is 00:22:23 You were saying before, you know, as a sex therapist, I could say, you know, friction is a major part of sexuality, isn't it? The erotic relies on friction and attraction plus obstacles equals excitement, says Jack Morin. So this idea that you smooth it all out, in effect, makes us less prepared for true relationships. What is the consequence? You retreat even more into a place where you don't have to deal with any of these. challenges. It's the mass atrophying of relationality, our ability to deal with that complexity. So it's just the weakening of all those muscles. I mean, from the perspective that you just laid out, technology is infantilizing our experience of being with that complexity, because the whole
Starting point is 00:23:05 premise of it is I can get what I want right now. If I'm not interested in the thing that someone in front of me is saying, I can just check out and check my email and get a couple of things done right now. I can cater to my own individual needs. If I don't know what to say to that complicated text message that the person I just went on a date with sent to me, I can just instead of responding to it and being forced to be there in a synchronous present state where they're staring at my eyes and sharing something personal and vulnerable, I can say, oh, that's an uncomfortable text message. I'm just going to go watch a couple YouTube videos right now. Or I can ghost you all together. Or ghost you all together, which is, I think, one of the things
Starting point is 00:23:37 we talked about when I was in your apartment in 2013 and this, the phenomenon of ghosting going up and up. And then there's those who justify that and saying, well, we have more relationships and more connections than ever. And so are we ghosting or are we just dealing with the complexity of how many relationships we have to manage? That was actually one of my first interests in the attention economy. Before there was even social media unraveling democracy and kind of causing all these big geopolitical trends. I was just interested in how the rewiring of attention and the relationships and just feeling overwhelmed by how many relationships we'd have to keep up, how hard that was. That was something I was worried about just in fraying the social fabric.
Starting point is 00:24:14 even before, you know, the engagement-based AIs that are pointed at our brains, strip-minding our attention, even before that, just managing all these relationships is something that's very complex and was kind of weakening the social fabric. But, you know, modern loneliness often masks itself as hyper-connectivity. That's the thing. You can have a thousand virtual friends, but nobody to come and feed your cat. You know, who is going to actually go and get your prescription at the pharmacy when you need to? So to call these relationships or to call these friends, that was such an amazing, you know, coup.
Starting point is 00:24:52 That's a word that has been hijacked. Right. You know, and totally changed its meaning. And there's a whole trend like that at this moment of words that are being used in the vernacular with a complete redefinition. Right. Followers, community. What community? On what basis?
Starting point is 00:25:12 You know, what are the rules of enducing? A community gives you a sense of belonging, but also a sense of obligation and duty and norms to follow. Not just I'm part of something. What is that something? Who is that something? So I use the word community sometimes and I think, oof, I fell in the trap. I'm participating in this thing. What am I?
Starting point is 00:25:33 Which community? Followers? What am I? A saint? A prophet, a messiah, the religiosity of it as well. You know, the kind of unconscious replacement of religion. but the system is the same. Right.
Starting point is 00:25:49 And this is how technology, actually the first point of our conversation, the mediation of everything, the changing of the meaning and the way that we relate to, what is a community, what is a friend, what is it to be, quote unquote, liked as represented by an arbitrary button with a thumbs up that was designed and placed in certain places
Starting point is 00:26:07 because it was good for a business model of driving up engagement and attention. And it's rewriting the meaning of all these core things that we're not questioning anymore, which is, again, I think with your work, it's about how do we see through the language that has kind of hijacked the meaning of what a relationship is, what a community is, what belonging or connection are. So one of the trends that we're also seeing is people sanctioning a replica of themselves. Rather than someone not asking you before they create an artificial Estere virtual
Starting point is 00:26:41 therapy chatbot. There's a 23-year-old Snapchat influencer with more than 1.8 million followers. Her name is Karen Marjorie. She created a replica of herself, which she calls a virtual girlfriend, and she will rent that replica out to other people for a dollar per minute. And she has more than a thousand, quote, boyfriends. She says her goal is to cure loneliness. I offer emotional and physical experiences just like a human does, but delivered digitally. Let's get to know each other better. And this strikes me as the kind of thing that we're talking about, except this time this person is sanctioning a replica of themselves. You can imagine a future world where everyone has a replica. And actually, Digital Minister of Taiwan, Audrey Tang, has found, you know, she has
Starting point is 00:27:23 so many requests for interviews that she actually did create a large language model of herself so that when the press asks her for interviews and she can't do it physically, she'll hand them to her AI replica. And I'm just curious what you see in the benefits of this approach and then also what are you going to be the cost? I'm not a futurist, so I'm going to just respond in the moment. I mean, you know, I think children often have an imaginary friend, and they feel very intimate with that friend. That is totally a figment of what they produce in their head.
Starting point is 00:27:56 Sometimes the friend is represented in a stuffed animal. So there is something, you could call it infantilization, or you could say it actually is a recall of something very, very deep that we experience in childhood, and that actually profoundly accompanies us and does help us with loneliness. When your little one is walking around, they're talking out loud to their imaginary friend that may not even be represented or has a form of representation, they are really in a state of relationship. They are related. When they take the doll and they put them to sleep or when they give them a bath and they talk out loud like children do, you think
Starting point is 00:28:37 who you're talking to and just having a whole world, you know. So I actually think, yes, there's something about that. We have the capacity, you know, our brain is all about fantasy and imagination. And it can imagine entire worlds and the systems of relationships in which we have a role in which somebody expects us, et cetera, et cetera. What does it do to human relationships is a different question. Does it help us? Like Lars and the Doll, you know, remember that movie?
Starting point is 00:29:07 and the real doll. It's one of the first ones where, like her, one of these visionary movies, where he basically grieves and mourns at loss through his relationship with this real-life doll. She won't wake up. Bianca! I think she's unconscious.
Starting point is 00:29:24 Hey. Are you sure? Look at me. Bianca. Hey. Hey, hey, hey, hey. Hey. Bianca's unconscious.
Starting point is 00:29:31 Do you hear me? Yonka. The doll gradually becomes a transitional object, and that's the thing. that happens in childhood, is that our tools, our toys, become transitional objects that gradually allow us to learn the ins and outs of having real relationships. But by transitional object, you mean something that's a temporary developmental container that allows you to practice something.
Starting point is 00:29:56 On the way to something else. On the way to something else. But what would be regressive is to get stuck with that imaginary friend psychology. And that's the infantilization that we are in. It's not to vilify the imaginary process or the fantasy, but it's understanding what is the right relationship or where is that right developmental stage for that kind of way of if relating.
Starting point is 00:30:15 So if I wake up and I feel is life worth living, which is one of the questions that we ask when we feel deeply alone. Who would notice if I wasn't there? Who cares about me? You know, what's the point? Why am I trying so hard? Maybe I'm not made for this world.
Starting point is 00:30:33 Maybe I'm not made for living. You know, will the bot give me a reason to go on? There's a deeper layer to just having somebody to talk to like I talk to Alexa, you know. And there's something when people describe the conversation that is so superficial to the kinds of aches and pains that we really live with. I don't see this ever being, the word suffering never appears in any of these languages. Yet, you know, what we live with, you know, it's easy to say depression and English. anxiety. You kind of give it a label, you diagnose it. And then it's bad. You know, it comes with deep grief, longing, loss, suffering, pain.
Starting point is 00:31:16 And those things are embodied experiences. And those things keep you up at night. And those things make you weep. And those things make you frozen. And, you know, when I watch this 40 years of practice in my office, you see these things. You don't just label these things. You live them. You know, you sit with somebody and your tears come down as theirs come down because you're in touch with the human experience, not with the machine. That's the difference. I mean, there's sort of the transhumanists who say that suffering is a mistake, an evolutionary mistake. And we should engineer suffering out of human existence.
Starting point is 00:31:52 And that's kind of the trajectory of technology is to make it easier to just erase suffering, to erase loneliness, give you an instant ability to run away from yourself, get virtual connection, virtual help. What do you think of that mindset, which I think a lot of people in Silicon Valley believe is the next step? Let's just erase these harmful experiences so we can end up with this high fructose corn syrup version. Of course, they don't see it as high fructose corn syrup.
Starting point is 00:32:14 They see it as a bootloader to a deeper kind of way of relating and better than the alternative, which does not have anything at all. It's very hard for me to relate to this because a therapist, we don't usually sit in the room with people who are happy. and who feel fulfilled in life at all times. I mean, we sit with people with loss. And the same people who say we shouldn't have any suffering also tell you that you can live forever and never be sick.
Starting point is 00:32:43 So they have a system. It's a very hermetic system. You will never lose anybody because nobody will be unhealthy because we will, you know, super young at every stage and defended against any of the diseases. But that's not the reality, at least not in my world at this moment. I live with people who lose their children, who lose their parents, who lose their limbs, who lose their livelihood, who lose their countries, who are rejected, who are betrayed,
Starting point is 00:33:14 whose trust has been violated. I mean, that's suffering. What are we trying to do, flatten the entire human experience? I also see people who come out of it, and so the first time see the light again and rejoice and are able to connect again and to love it. again and to feel alive again. And what I miss when you describe this thing to me is I don't hear the word alive. I hear, you know, something that's basically flat and that's very different.
Starting point is 00:33:45 Alive is the erotic energy. It's a sense of vitality, vibrancy, radiance, it's energetic. There is zero energy in anything that list you just made for me. That is not living. There's a different story here. I think we should get into solutions. I mean, both I think our audiences care and are interested in what we think a more ideal world should look like. And how do we have like the Toyota Prius, the hybrid version of our virtualized experiences with our real world experiences?
Starting point is 00:34:20 And how do we have technology that actually invites us to live with some portfolio of virtualized mixed with the physical, the convenient mixed with the complex and the friction. And I actually remember being a design ethicist at Google when I was thinking about how to redesign literally the operating system of Android or iPhones to be privileging this kind of view. Like imagine if the people who are designing technology as a first order basis,
Starting point is 00:34:46 they think about the design in terms of this portfolio, this healthy portfolio. I don't like using the FDA, Nutrition Facts Pyramid sort of version of a healthy media diets or that formulation. But you could imagine kind of a reality formulation of what is a portfolio that blends the physical with the virtual, the embodied with the disembodied?
Starting point is 00:35:05 I would love to hear maybe it's a catalyzing effect where you would go. I mean, interestingly, I'm tempted to answer this more anecdotally and even personally because it's actually a subject that touches me a lot and that I am very involved with. So I think when I remember the first beginnings of phase. book when my kids would use it and it was a way to actually gather people to know where's the party tonight it was used as a way to lead you into in an in-person meeting it didn't take it as a replacement the point is that you the improvisation the serendipity the spontaneity that the whole
Starting point is 00:35:47 element that makes us feel alive it's that meeting with the unknown and the unpredictable that needed to be done in person i mean just to link this example with what we're doing right now. This is a podcast. You and I both do podcasts. I'm sure you also experience this when we do episodes. Right now, as we're recording this, I'm just seeing you on my screen. I don't have a felt sense of the million people who are going to listen to this. And I don't feel whether this is going to matter, frankly. I mean, we do this because we want to change how the tech industry and society harmonize so that we actually have a more humane future. And that's why we're doing this. We put a lot of thought and effort into it. But I don't
Starting point is 00:36:26 have any felt sense of whether this podcast is making that impact, except when I maybe run into people out in the world, and I do get to hear that. But let's imagine, again, to translate this back into design, because the whole point is technology can be designed to reinforce the maximum virtualization and efficiency of podcast production, or it could be designed in a way, where imagine Apple Podcasts says, hey, there's these community events that for everybody who loves your undivided attention or loves your podcast, Astaire, here are events that are happening in your area that are coming up on these dates or here are ways that communities can demand and here are the topics people want to know about and here's a way that we can bring people physically together
Starting point is 00:37:02 if not with the hosts of the podcast at least with each other like meetups but integrated directly into the way the podcast imagine that that was how apple did this thing right imagine that's how Spotify did this thing instead of maximizing the virtual they really did try to create this hybrid version of an embodied and virtualized experience so I try to do it on a personal level all the time. I think that it's a beautiful tool and, not but, but and. I think you can put it in the design. I think you can be on a dating app and the dating app is also inviting you as soon as possible to meet. I think that everyone today that wants to have in-life events is talking about how hard it is to get people to come outside. People are complaining that they feel lonely and
Starting point is 00:37:49 disconnected and flat, but they're not able to actually go outside and re-engage. And I think that if we try to promise them that they can have the same experience by staying at home as they would have if they left, do the streaming of the concert, do the streaming of this and that, then we are basically going to transform human nature into something else. We are changing human beings. As we have always changed human beings, it's not the first time, but we are changing. very profound is switching inside of us. And what happens is that when you and I are concerned about this, we are seen by the techno-optimists as kind of retrograde. You know, we're from the last
Starting point is 00:38:33 century. They were nostalgic for a golden era for what it meant to be human 50 years ago. Who's to say that what it means to be human where everyone's plugged into VR, the metaverse and virtualized relationships and synthetic boyfriends and girlfriends with replicas and chatbots? that why, you know, if people are quote unquote happy in that, who are you, Tristan and Esther, to say that that's bad for people or that's not real, right? And this is where I think the adult developmental psychology in someone like you who's studied in a deeper way, when you get subterranean, when you get underneath the hood of what's going on for people, there's a lot more loneliness or anxiety that actually, I think those signals of loneliness and anxiety being on
Starting point is 00:39:08 the other side of that virtualized environment show us that it's not fulfilling. And there are more embodied experiences and more complexity and more friction that we know the people who are living in those ways maybe have a lot less of that loneliness and anxiety. But we need to be able to establish more normative facts about human well-being and flourishing so that we can say that, yes, it would be a good thing for technology companies to design and privilege more friction, more complexity, more in-person. We did the same thing with junk food, right? We first had a promise of repeatable, always available, shelf, stable food. And it was iriscipline. And only now are we coming to terms with the fact that we traded away nutrition, right?
Starting point is 00:39:50 In the same way that we're going to trade away real human connection and that the consequences were not very good for our physical health. Well, it won't be any different when we talk about our relational or our mental health. Okay. Esther, do you have any parting thoughts for our listeners and closing out this conversation today? I think you're doing something so important, you know. Because what you're really asking for is responsibility. You're asking for people who are at the forefront of the changes in our society, in our economy, in our political system, in our humanity, to think responsibly about what they are doing.
Starting point is 00:40:31 If when it comes to mental health, it means that we're doing it with caution and with care. And we don't just look at money. You know, psychiatry is a real complex history of cultures. And you don't just erase all of this and basically just say, oh, come on, if I can promise you a life without suffering, wouldn't you want it? No, actually, because the first philosophical questions that come up is, how do you know you're happy if you've never been unhappy? I mean, we live as dialectic creatures. So I'm in awe of what you do and the relentlessness with which you have been at it since the first time we met. you know and I've learned a ton from you because I think of it really much more in my own sphere
Starting point is 00:41:16 because that's the one I know something about I take full advantage of many technologies but I still very much want to talk to you in person and I want to reach you in your relationship life in person you know whereas you are really working with larger systems and the design of these systems and that is the piece where I don't go because it's not my world. What stands out for me, and this will be one of my parting words, actually, is that I go from one technology conference to another. I can tell you the word relationship is never mentioned, never mentioned. There are moonshots about everything, environment, education, health, transportation, you name it, nobody ever is looking at a moonshot for relationships. And yet, all these
Starting point is 00:42:04 technologies are profoundly affecting how we relate to each other, how we relate to each other, how we relate to ourselves, how we form relationships, and I think that it's irresponsible to not pay attention to that part. That's the main piece here that I would like to highlight is it is the challenge that everybody is avoiding, partly because relationships are complex systems with a lot of contradictions inherently, and they don't just suit one person, but they are at a root of societies. If you don't manage relationships, you don't manage social and political systems either. You know, you wanted to talk to me about polarization. I mean, it is about how do you hold two parts, you know, and when you are alone, you usually have to live with those
Starting point is 00:42:53 different parts inside of you that pull in different directions. When you are in a relationship, you often outsource onto the other person, the part of the equation that you repudiate. And when you do it in a society, you do the exact same thing, and it becomes, I am pro-life and you're what, pro-death? I mean, what exactly are we saying here? You know? So if I accepted to do this conversation, it's in large part because I want those of you who are at the forefront of this technological revolution to not forget relationships. It's not enough to just say that GPT-4 can be as smart as me and be as creative as me. It's great. I mean, I'll use any tool that can enhance me. But you have to be responsible and look at the social consequences and the relational
Starting point is 00:43:42 consequences of what we are creating. Well, I couldn't have said it better. Thank you so much, Esther, for coming on your undivided attention. It's an honor to be your friend, to know you, and to really just admire all the work that you're doing to raise awareness about the complexity of how we hold relationships and have a more healthy society. So thank you so much. Thank you. Your undivided attention is produced by the Center for Humane Technology, a non-profit working to catalyze a humane future. Our senior producer is Julia Scott. Josh Lash is our researcher and producer, and our executive producer is Sasha Fegan.
Starting point is 00:44:22 Mixing on this episode by Jeff Sudaken, original music by Ryan and Hayes Holiday. And a special thanks to the whole Center for Humane Technology team for making this podcast possible. You can find show notes, transcripts, and much more at HumaneTech. And if you like the podcast, we'd be grateful if you could rate it on Apple Podcasts because it helps other people find the show. And if you meet it all the way here, let me give one more thank you to you for giving us your undivided attention.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.