Your Undivided Attention - Esther Perel on Artificial Intimacy

Episode Date: August 17, 2023

For all the talk about AI, we rarely hear about how it will change our relationships. As we swipe to find love and consult chatbot therapists, acclaimed psychotherapist and relationship expert Esther ...Perel warns that there’s another harmful “AI” on the rise — Artificial Intimacy — and how it is depriving us of real connection. Tristan and Esther discuss how depending on algorithms can fuel alienation, and then imagine how we might design technology to strengthen our social bonds.RECOMMENDED MEDIA Mating in Captivity by Esther PerelEsther's debut work on the intricacies behind modern relationships, and the dichotomy of domesticity and sexual desireThe State of Affairs by Esther PerelEsther takes a look at modern relationships through the lens of infidelityWhere Should We Begin? with Esther PerelListen in as real couples in search of help bare the raw and profound details of their storiesHow’s Work? with Esther PerelEsther’s podcast that focuses on the hard conversations we're afraid to have at work Lars and the Real Girl (2007)A young man strikes up an unconventional relationship with a doll he finds on the internetHer (2013)In a near future, a lonely writer develops an unlikely relationship with an operating system designed to meet his every needRECOMMENDED YUA EPISODES Big Food, Big Tech and Big AI with Michael MossThe AI DilemmaThe Three Rules of Humane TechDigital Democracy is Within Reach with Audrey Tang Your Undivided Attention is produced by the Center for Humane Technology. Follow us on Twitter: @HumaneTech_ 

Transcript
Discussion (0)
Starting point is 00:00:00 I go from one technology conference to another. I can tell you the word relationship is never mentioned. Never mentioned. There are moonshots about everything, environment, education, health, transportation, you name it. Nobody ever is looking at a moonshot for relationships. And yet, all these technologies are profoundly affecting how we relate to each other, how we relate to ourselves. That's famous psychotherapist Esther Perel,
Starting point is 00:00:34 who's the New York Times bestselling author of the books, The State of Affairs, and Mating in Captivity. She's had celebrated TED Talks that have garnered more than 20 million views, and is also the host of the two popular podcasts, Where Should We Begin, and How's Work? And Astaire and I have been going back and forth for the last 10 years about how technology has been distorting the way that human beings form relationships and distorting the way that we gain or lose intimacy.
Starting point is 00:00:59 And what we're about to get into is how do we deal with technology that is distorting our relationships and our intimacy? What would it look like for technology to foster meaningful connection and meaningful intimacy and distinguish between the fake kind and the real kind? Who are we to say what fake and real actually is? So why would we do an episode just about relationships? In the AI dilemma talk that Aiz and I gave, we talked about how loneliness is a national security risk. As you see in the United States the number of shooters and shootings going up
Starting point is 00:01:34 and innocent people dead from these kinds of events, are the people that are doing these shootings in healthy relationships? Do they feel deep belonging? Or do they feel mass alienation? How many of them had healthy romantic relationships? Oftentimes when we talk about the future of technology, we talk about how AI can solve climate change and give people mental health benefits, absent the word, relationships. How does the ground underneath our feet that depends on this very deep interconnected structure of relationships, how does that help us answer questions like loneliness? Loneliness isn't some disease to be cured by having a person interact with the chatbot. It's how does technology sear us towards the kinds of things that sow deep interconnected relationships
Starting point is 00:02:18 that create a healthy society? Can you have social trust in a society if you don't have relationships? Now, this isn't just some nostalgia for a bygone era. We have to figure out and celebrate all the things that the virtualization of our lives by AI, by technology, by smartphones, by being able to connect with each other around the world and different time zones, we need to be able to celebrate the good things that come from that. But we also have to live in right relationship between how those things land in a body, waking up, opening your eyes, taking a deep breath, seeing the sunshine, and living the day.
Starting point is 00:02:54 Esther Perel, welcome to your undivided attention. Thank you. It's my pleasure to be here. So it's kind of hard to have this conversation in almost an artificially formal way, because you and I know each other for quite some time. I think we met. I believe it was 2013. I remember knowing of your work and seeing you at a conference, and I remember that when we talked for the first time, and I kind of walked you through, I was a design ethicist at Google at the time. I was first starting this work inside of Google, and I was feeling trapped. And I remember walking you.
Starting point is 00:03:24 you through a version of that slide deck that I had spread internally and the race to the bottom of the brain stem. And your response was, we've been completely hijacked. And I started using that word hijacked in our work, how technology is hijacking these core psychological vulnerabilities. But maybe from your perspective, I'm just curious if you want to reflect on how we got to meet, specifically maybe a moment that I know I was in your apartment, I think, after that. So we met twice for the first time, so to speak. first time I was at this conference and indeed you open my eyes
Starting point is 00:03:58 it was like in my mind you say this is not inevitable this is actually controlled by particular agents that have an invested interest and that's when I just felt like yes I am a pawn in a system that I don't even know I'm in
Starting point is 00:04:13 so that was the first reaction and then the second time when we met I remember that I asked you basically to show me your Tinder feed I think I was using Tinder, yes. Yes, and I wanted to see, you know, what you write and what people say. And I just thought that it was the most unerotic and unpoetic. Thank you for that very, very kind compliment.
Starting point is 00:04:36 But I thought, maybe I can suggest a little bit. It's like, I thought, it's so deflated and empty of energy. How can people actually really fall for each other or become curious of each other? And I thought, that was not just your doing. See, at first I thought, oh, you know, it's you. your experience, but then I thought, no, there is something in the medium that really takes the juice out of it as well. And it's a combination of things. I'm not saying that people can't be utterly creative and inspired on an app, but that there is something going on that
Starting point is 00:05:08 is changing human relationships, changing how we meet, how we connect, how we date, and how we break up. Yep. Well, totally. And I remember, and thank you for not just blaming me in my lack Lack of juice and Tinder communication. That was an earlier time in our lives. But I think that's what we're here to talk about today, which is artificial intimacy, how is technology mediating and changing the form of relationships, you know, the loneliness crisis,
Starting point is 00:05:36 fewer and fewer people actually having sexual relationships and deep connection and feeling connected? And actually clarifying the language, the thing that I've always appreciated about your work is it comes with a kind of subterranean deeper look at what are the structures underneath the way that we tend to talk about it. Because the way we tend to talk about
Starting point is 00:05:53 the breakdown of relationships and intimacy, it's like we're missing the vocabulary. And the thing I've always appreciated about talking to you is we get to go underneath that vocabulary to some of the underlying dynamics. And just to kind of link it to our work, the premise of humane technology is it comes to the deeper ergonomic understanding
Starting point is 00:06:11 of what it means to be human so that technology can fit that ergonomics better. I call it the other AI, by the way. Yeah, say something about that. Yes, I mean, I call it the other AI, the rise of artificial intimacy. And what I'm interested in is what does it mean for us? And it started because I got a note from someone who said, have you seen this? And it was somebody who basically wanted to have a session with me after he broke up
Starting point is 00:06:40 and was wanting to understand what had happened in his relationship. And he couldn't get an appointment. And I wasn't available. So he decided to create me. and he created the A.I. Esther, and A.I. Esther was available. She didn't have a booked up calendar. A.I. Esther was, you know, pure. She didn't have anything interfering in her life to be there for him. And I just thought, this is amazing. And he thinks that I have fundamentally helped him and that his experience with talking with me, i.e. robot me, was really illuminating and gave him tremendous peace, quote. And I just thought, wow, I don't even have to be there. In the past, I would have said you just have to fantasize about someone and make them be really present.
Starting point is 00:07:28 Now you can actually go a step further and have your fantasy create that person and then ground it in their corpus of knowledge and body of work. And I thought, this is the other AI. Now, what experience did he actually have with me? So just to be clear, because we've been talking about the rise of large things, language models and generative AI and what this is doing.
Starting point is 00:07:51 So someone built a large language model, built on a corpus of what everything that you've said on your podcast, all the interviews we've ever given, and created this virtual ester who's available to everyone all the time to provide relationship advice. And this is kind of the same as this sort of seduction, this sort of deal with the devil trade we get with technology overall, right? We get this infinite access to always on swiping of the infinite pool of all human beings on dating apps. We get the infinite access to the best people.
Starting point is 00:08:18 that we want to talk to all the time, but it's always in this virtual form. And in this, I think one of the things I wanted to explore with you today is, on the one hand, people will say, isn't it great that now everyone has access to you? But then the other hand, they'll know on some level that they've been interacting with some virtual agent and maybe the part of them after the session is over feels a little bit empty because they still don't have a connection with a real person. What is the trade that we're making with the virtual? But I'm just curious, as you had this new agent show up without your consent,
Starting point is 00:08:50 because this person did this without asking you, right? No, no, no. I had nothing to do with it. It was all done when I arrived. So what did this bring up for you? I mean, you know, a part of me was flattered. A part of me was curious. A part of me had deep ethical concerns.
Starting point is 00:09:12 You know, I went in many, many directions. A part of me felt violated. But, you know, on the other hand, somebody said to me, but your books have done the same. What's the difference? And people have quoted you from your books, and they have misquoted you, and at least the butt quotes you accurately.
Starting point is 00:09:29 But the butt has no soul. The butt regurgitates what I have said elsewhere. And as a clinician, our practice is highly nuanced, highly relational, and highly contextual. The art of therapy is not, to tell you what I just told the person just before you. It's to actually say something that is completely in the
Starting point is 00:09:49 moment and probably very different than what you may have anticipated so that I don't regurgitate I'm not just an amalgamation of everything that has been accumulated and summarized and organized which is important. It's not like it doesn't have its importance
Starting point is 00:10:05 but this connection is multidimensional, it is layered it is embodied, it is embodied, it's the other thing. It's fundamentally embodied. It is experienced through the senses. And when you flatten it, you reduce the nuance and you reduce the connectivity of the human experience. The other thing, I think that I did also see is, you know, many of our relational issues are complex problems. They're dilemmas. They don't necessarily have a solution. They are paradoxes that we need to manage. technology believes that everything has a solution
Starting point is 00:10:47 you can reduce things to binaries and find an answer and that's not really what happens in our existence so it is very useful for some aspects of our mental health and our relational life but it doesn't touch some of the other more profound aspects so what are some examples of that to ground that for listeners of the dilemmas and complexity that are not reduced to answers shall I stay or shall I go?
Starting point is 00:11:15 Where am I at in my relationship? Shall I have another child? I say, yay, you say nay. Shall I move? Shall we move? Shall I change my career? Shall I move my parents into my house to take care of them because they are not well?
Starting point is 00:11:33 You know, values questions, moral dilemmas, existential questions that are prevalent in clinical work. I mean, there are issues that have profoundly to do with pathology and mental illness, but there are many other things that have to do with emotional well-being and with values, and those things, they're not an answer. They're about holding the contradiction. You know, they're about grief, because whichever choice you're going to make, you're going to have to mourn something else.
Starting point is 00:12:03 And that's different than an answer with a solution that wraps a nice bow around it. Here's what you need to do. It's not like the next song you need to do. need to listen to or the directions you should take to go the fastest way, you know, to a place or the movie you should watch tonight because you don't know what you want to see. It's very different. This is about how will this affect my life, the people that I'm connected to, et cetera, et cetera. So you're much more systemic. Yeah, the very premise of technology is an individual experience often. I mean, it's not like when that guy built that virtual ester
Starting point is 00:12:35 version of you. He didn't, I mean, he could build it as a couple's chat agent, by the way, right? He could say multiple people could be talking to the virtual version of you. But one of the problems that you and I both, I think, deeply agree on is the individualization of the experience. I mean, something that's been bothering me, actually, is the idea that mental health is just this individual thing. We just need to have these interventions for an individual when it's so much of it is about this deeper social connection, belonging, non-alienation experience. And our technology is not wired for that. The designers at Apple, who design the interface for an iPhone, aren't designing it for shared experiences. Social media is not being designed for shared experiences.
Starting point is 00:13:14 Tinder isn't designing for group experiences or, you know, which groups do we want to go out to? Where can I go out to be with a community tonight? It's all designed for this individual commodified experience. But, you know, I think what's interesting is what it is doing is that it is changing our expectations of intimacy. You know, my AI creator, he thinks he had a fantastic experience, as do many people who have AI assisted therapy, they think that they have experienced intimacy. But when you actually look at a qualitative experience, what they experienced is a lot less. It's the same thing if you have bad food or fast food or you get adjusted to thinking it's food. So you lower the expectations and then you tell people, this is good intimacy.
Starting point is 00:14:06 So, you know, Sherry Turkle and Todd Essex are two people who've written quite extensively on artificial intimacy when it comes to telehealth. And the point that they make is that when people get mental health treatment online, they often will describe themselves as satisfied. And they say that it would be as good as if they had been there in person. But qualitatively, the result is not as good, which means that patients will be satisfied. When they lower their expectations, for what? For empathy, for intimacy, for connection, for feeling understood, and they lower their expectations to match what the technology can provide. So they receive less effective care, but they don't recognize the difference.
Starting point is 00:14:53 I feel like I'm saying the obvious. That's one thing honestly that bothers me about the space is it makes it hard to talk about. Everyone kind of sounds like we're speaking in these cliches, and it's hard to get underneath that to what's really wrong with all this. So if I, but I want to go with what's really wrong, I will just share observations. And then you can decide if you think that's problematic or not. You know, you would think, I mean, the way I began one day to think of it is
Starting point is 00:15:23 I'm living a kind of assisted living, but prematurely. I'm being assisted by a host of predictive technologies that are basically saying, you don't have to know, I will know for you. I'll recommend the next song to listen to, who to date, where to eat, you know. And you would think that that should actually make us feel more confident, more at ease, because I'm neutralizing the unpredictable, the unknown. But you know, l'in canu, as they say in a big word in French, the unknown, is it actually demands that you interact with it on a daily basis a little bit.
Starting point is 00:16:04 The big unknown, of course, is death, but there is plenty of others in our life. And when you erase all of this, you actually make people more anxious, more unsure, filled with self-doubt, constantly wondering, you know, is there better? Is there more? And am I enough? And those are psychological questions that I think are directly related to the increased virtualization and technologies and areas. I put it all in one piece, is that the rubbing, the living in close proximity with the messiness of another person that helps you know who you are in the presence of others, that's the piece that we are noticing in our work at this moment. Every therapist I teach and I supervise, we are having that very same conversation.
Starting point is 00:16:56 So artificial intimacy with the butt isn't actually what concerns me. What concerns me is how the digitally facilitated connections are lowering our expectations and our competence in the intimacy between humans. And that I do think that all of that makes us less able to be with people who challenge us. From the political to the personal to the familial, we, you know, we polarize much faster than we. ever have because I shouldn't have to be uncomfortable. Where did we get that idea? Maybe discomfort is actually a major piece of life and you learn to deal with those discomforts. You know,
Starting point is 00:17:45 you talk about delayed gratification or having all your needs met in the moment. We used to think that maturity was in part learning how to delay gratification. Maturity is about holding ambivalence and understanding that complex issues live with inherent contradictions. Ambivalence is about knowing that you have certain feelings for a person and the opposite feelings for that person, and they coexist. And not that you have to get rid of one side of it. So it's a very interesting new thinking about then what is development? What is psychological development look like at this moment?
Starting point is 00:18:24 Well, technology certainly does not feel like it is driving. us towards greater development, but more anti-development towards infantilization, instant gratification, lowered expectations, inability to deal with that complexity. That's one of the experiences that take place. At the same time, there's a lot of other uses from, look, to be really clear, this is not a critique of AI in and of itself, or even of the virtualization of things. I work, we are here, you and I, because of it. This is the conversation we're able to have.
Starting point is 00:18:58 we were able to get through the pandemic thanks to it. You know, people listen to my podcast. Where should we begin, thanks to it? I mean, I mean it. Let's not vilify the virtual. You know, this is not that. But there is something about understanding what it means to do it responsibly. So, yes, there is the help that I can get from it as a therapist in terms of progress
Starting point is 00:19:21 notes, in terms of tracking questionnaires, in terms of symptom compliance, in terms of even just having an assistance. you know, a virtual assistant, but more than a person. All of these things are really useful. And I think that the bot can ask you to do your breathing exercises, to help you track your thoughts, to help you map your feelings. But there is something else that has to do with differential diagnosis, that has to do with treatment conceptualization,
Starting point is 00:19:50 that has to do with the big picture insights that one has when you work, and that has to do with the highly relational aspect of our work. And that are the pieces that we are not there yet and that need to be done very incrementally and very gradually. That's it. It's not about can we replace people or not replace people. It's that we need to find waste for the experience of mutuality and reciprocity that is part of an intimate experience.
Starting point is 00:20:20 In reality, I mean, one of the main things we teach children is that when you are little, it's all about you. But part of growing up is becoming aware of the presence of others, the needs of others, the humanity of others, the impact of your behavior on others. It's that you become a relational creature. If it's about just having my needs satisfied and I am being tracked and you are always there for me
Starting point is 00:20:45 and you don't have any bad days and you don't forget anything, what exactly are we saying about relationships, you know, that's the piece is that it only will make us more anxious and more unable to deal with the unexpected
Starting point is 00:21:03 and the unpredictabilities and the messiness of life. Part of why we are experiencing more and more anxiety is because we are having less and less spaces in which to experience the friction, the obstacles, the failures that help us know who we are
Starting point is 00:21:20 and who we are not. You know, why are people having less sex. You were saying before, you know, as a sex therapist, I could say, you know, friction is a major part of sexuality, isn't it? The erotic relies on friction and attraction plus obstacles equals excitement, says Jack Morin. So this idea that you smooth it all out, in effect, makes us less prepared for true relationships. What is the consequence? You retreat even more into a place where you don't have to deal with any of these challenges. It's the mass atrophying of relationality, our ability to deal with that complexity.
Starting point is 00:21:58 So it's just the weakening of all those muscles. I mean, from the perspective that you just laid out, technology is infantilizing our experience of being with that complexity because the whole premise of it is I can get what I want right now. If I'm not interested in the thing that someone in front of me is saying, I can just check out and check my email and get a couple things done right now. I can cater to my own individual needs. If I don't know what to say to that complicated text message that the person I just went on a date with sent to me, I can just, instead of responding to it and being forced to be there in a
Starting point is 00:22:23 synchronous present state where they're staring at my eyes and, you know, sharing something personal and vulnerable, I can just say, oh, that's an uncomfortable text message. I'm just going to go watch a couple YouTube videos right now. And so it's the mass... Or I can ghost you all together. Or ghost you all together, which is, I think, one of the things we talked about when I was in your apartment in 2013 and this, the phenomenon of ghosting going up and up. And then there's those who justify that and saying, well, we have more relationships and more connections than ever. And so are we ghosting or are we just dealing with the complexity of how many relationships we have to manage? That was actually one of my first interests in the attention economy.
Starting point is 00:22:57 Before there was even social media unraveling democracy and kind of causing all these big geopolitical trends. I was just interested in how the rewiring of attention and the relationships and just feeling overwhelmed by how many relationships we'd have to keep up, how hard that was. Like that was something I was worried about just in fraying the social fabric. Even before the engagement-based AIs that are pointed at our brains, strip-mining our attention, even before that, just managing all these relationships is something that's very complex and was kind of weakening the social fabric. But, you know, modern loneliness often masks itself as hyper-connectivity.
Starting point is 00:23:35 That's the thing. You can have a thousand virtual friends, but nobody to come and feed your cat. You know, who is going to actually go and get your prescription at the pharmacy when you need to? So to call these relationships or to call these friends, that was such an amazing, you know, coup. That's a word that has been hijacked. Right.
Starting point is 00:23:56 You know, and totally changed its meaning. And there's a whole trend like that at this moment of words that are being used in the vernacular with a complete redefinition. Right. Followers, community. What community? On what basis? You know, what are the rules of engagement?
Starting point is 00:24:15 engagement. The community gives you a sense of belonging, but also a sense of obligation and duty and norms to follow. Not just I'm part of something. What is that something? Who is that something? So I use the word community sometimes and I think, oof, I fell in the trap. I'm participating in this thing. What am I? Which community? Followers? What am I? A saint? A prophet, a messiah, the religiosity of it as well. You know, the kind of unconscious replacement of religion. But the system is the same. Right. And this is how technology, back to the first point of our conversation, the mediation of everything, the changing of the meaning and the way that we relate to, what is a community, what is a friend, what is it to be, quote unquote, liked as
Starting point is 00:25:03 represented by an arbitrary button with a thumbs up that was designed and placed in certain places because it was good for a business model of driving up engagement and attention. And it's rewriting the meaning of all these core things that we're not questioning anymore, which is, again, I think with your work, it's about how do we see through the language that has kind of hijacked the meaning of what a relationship is, what a community is, what belonging or connection are. So one of the trends that we're also seeing is people sanctioning a replica of themselves, Rather than someone not asking you before they create an artificial, Astaire virtual therapy chatbot, there's a 23-year-old Snapchat influencer with more than 1.8 million followers.
Starting point is 00:25:48 Her name is Karen Marjorie. She created a replica of herself, which she calls a virtual girlfriend, and she will rent that replica out to other people for a dollar per minute. And she has more than a thousand, quote, boyfriends. She says her goal is to cure loneliness. I offer emotional and physical experiences just like a human does, but delivered digitally. Let's get to know each other better. And, you know, this strikes me as the kind of thing that we're talking about,
Starting point is 00:26:13 except this time this person is sanctioning a replica of themselves. You can imagine a future world where everyone has a replica. And actually, Digital Minister of Taiwan, Audrey Tang, has found, you know, she has so many requests for interviews that she actually did create a large language model of herself so that when the press asks her for interviews and she can't do it physically, she'll hand them to her AI replica. And I'm just curious what you see in the benefits of this approach, and then also what are going to be the costs?
Starting point is 00:26:40 I'm not a futurist, so I'm going to just respond in the moment. I mean, you know, I think children often have an imaginary friend, and they feel very intimate with that friend. That is totally a figment of what they produce in their head. Sometimes the friend is represented in a stuffed animal. So there is something, you could call it infantilization, or you could say it actually is a recall of something, something very deep that we experience in childhood,
Starting point is 00:27:11 and that actually profoundly accompanies us and does help us with loneliness. When your little one is walking around, they're talking out loud to their imaginary friend that may not even be represented or has a form of representation, they are really in a state of relationship. They are related.
Starting point is 00:27:31 When they take the doll and they put them to sleep or when they give them a bath and they talk out loud like children do you think who you're talking to and just having a whole world you know so I actually think yes there's something about that we have the capacity you know our brain is all about
Starting point is 00:27:49 fantasy and imagination and it can imagine entire worlds and the systems of relationships in which we have a role in which somebody expects us etc etc what does it do to human relationships is a different question does it help us
Starting point is 00:28:04 like Lars and the doll you know Remember that movie, Lars and the Real Doll? It's one of the first ones where, like her, one of these visionary movies where he basically grieves and mourns at loss through his relationship with this real-life doll. She won't wake up. Bianca! I think she's unconscious.
Starting point is 00:28:25 Hey. Are you sure? Look at me. Bianca. Hey. Hey, hey, hey, hey. Hey. Bianca's unconscious.
Starting point is 00:28:32 Do you hear me? Bianca. The doll gradually becomes a transitional, object. And that's the thing that happens in childhood, is that our tools, our toys, become transitional objects that gradually allow us to learn the ins and outs of having real relationships. But by transitional object, you mean something that's a temporary developmental container that allows you to practice something. On the way to something else. On the way to something else. But what would be regressive is to get stuck with that imaginary friend psychology. And
Starting point is 00:29:04 that's the infantilization that we are in. It's not to build. the imaginary process or the fantasy, but it's understanding what is the right relationship or where is that right developmental stage for that kind of way of relating. So if I wake up and I feel is life worth living, which is one of the questions that we ask when we feel deeply alone,
Starting point is 00:29:23 who would notice if I wasn't there? Who cares about me? You know, what's the point? Why am I trying so hard? Maybe I'm not made for this world. Maybe I'm not made for living. You know, will the bot give me a reason to go on? There's a deeper layer to just having somebody to talk to like I talk to Alexa, you know.
Starting point is 00:29:45 And there's something when people describe the conversation that is so superficial to the kinds of aches and pains that we really live with. I don't see this ever being, the word suffering never appears in any of these languages. Yet, you know, what we live with, you know, it's easy to say, the price. and anxiety. You kind of give it a label, you diagnose it. And then it's bad, yeah. You know, it comes with deep grief, longing, loss, suffering, pain. And those things are embodied experiences.
Starting point is 00:30:20 And those things keep you up at night. And those things make you weep. And those things make you frozen. And, you know, when I watch this, 40 years of practice in my office, you see these things. You don't just label these things. You live them. You know, you sit with some.
Starting point is 00:30:35 somebody and your tears come down as theirs come down because you're in touch with the human experience, not with the machine. That's the difference. I mean, there's sort of the transhumanists who say that suffering is a mistake, an evolutionary mistake, and we should engineer suffering out of human existence. And that's kind of the trajectory of technology is to make it easier to just erase suffering, to erase loneliness, give you an instant ability to run away from yourself, get virtual connection, virtual help. What do you think of that mindset, which I think a lot of people in Silicon Valley believe is the next step? Let's just erase these harmful experiences so we can end up with this high fructose corn syrup version.
Starting point is 00:31:14 Of course, they don't see it as high fructose corn syrup. They see it as a bootloader to a deeper kind of way of relating and better than the alternative, which does not have anything at all. It's very hard for me to relate to this because a therapist, we don't usually sit in the room with people who are happy and who feel fulfilled in life. at all times. I mean, we sit with people with loss. And the same people who say we shouldn't have any suffering also tell you that you can live forever and never be sick. So they have a system. It's a very hermetic system. You will never lose anybody because nobody will be unhealthy, because we will, you know, super young at every stage and defended against any of the diseases. But that's not the reality, at least not in my world at this moment. I live with people who lose
Starting point is 00:32:03 their children, who lose their parents, who lose their limbs, who lose their livelihood, who lose their countries, who are rejected, who are betrayed, whose trust has been violated. I mean, that's suffering. What are we trying to do, flatten the entire human experience? I also see people who come out of it, and so the first time see the light again and rejoice and are able to connect again and to love again and to feel alive again. again. And what I miss when you describe this thing to me is I don't hear the word alive. I hear something that's basically flat. That's very different. Alive is the erotic energy. It's a sense
Starting point is 00:32:49 of vitality, vibrancy, radiance, it's energetic. There is zero energy in anything that list you just made for me. That is not living. There's a different story here. I think we should get into solutions. I mean, both I think our audiences care and are interested in what we think a more ideal world should look like. And how do we have like the Toyota Prius, the hybrid version of our virtualized experiences with our real world experiences? And how do we have technology that actually invites us to live with some portfolio of virtualized mixed with the physical, the convenient mixed with the complex and the friction? and I actually remember being a design ethicist at Google
Starting point is 00:33:36 when I was thinking about how to redesign literally the operating system of Android or iPhones to be privileging this kind of view. Imagine if the people who are designing technology as a first order basis, they think about the design in terms of this portfolio, this healthy portfolio. I don't like using the FDA, Nutrition Facts Pyramid
Starting point is 00:33:53 sort of version of a healthy media diets or like that formulation. But you could imagine kind of a reality formulation of what is a portfolio that blends the physical with the virtual, the embodied, with the disembodied. I would love to hear, maybe it's a catalyzing effect, where you would go. I mean, interestingly, I'm tempted to answer this more anecdotally and even personally, because it's actually a subject that touches me a lot
Starting point is 00:34:21 and that I am very involved with. So I think when I remember the first beginnings of Facebook when my kids would use it, and it was a way to actually gather people to know where's the party tonight. It was used as a way to lead you in an in-person meeting. It didn't take it as a replacement. The point is that the improvisation, the serendipity, the spontaneity, that the whole element that makes us feel alive. It's that meeting with the unknown and the unpredictable.
Starting point is 00:34:54 That needed to be done in person. I mean, just to link this example with what we're doing right now, this is a podcast. You and I both do podcasts. I'm sure you also experience this. When we do episodes, right now, as we're recording this, I'm just seeing you on my screen. I don't have a felt sense of the million people who are going to listen to this. And I don't feel whether this is going to matter, frankly. I mean, we do this because we want to change how the tech industry and society harmonize so that we actually have a more humane future. And that's why we're doing this. We put a lot of thought and effort into it. But I don't have any felt sense of whether this
Starting point is 00:35:28 podcast is making that impact, except when I maybe run into people out in the world, and I do get to hear that. But let's imagine, again, to translate this back into design, because the whole point is technology can be designed to reinforce the maximum virtualization and efficiency of podcast production, or it could be designed in a way, where imagine Apple Podcasts says, hey, there's these community events that for everybody who loves your undivided attention or loves your podcast, Astaire, here are events that are happening in your area that are coming up on these dates, or here ways that community can demand and here are the topics people want to know about and here's a way that we can bring people physically together if not with the hosts of the podcast at least with
Starting point is 00:36:05 each other like meetups but integrated directly into the way the podcast imagine that that was how Apple did this thing right imagine that's how Spotify did this thing instead of maximizing the virtual they really did try to create this hybrid version of an embodied and virtualized experience so I try to do it on a personal level all the time I think that it's a beautiful beautiful tool, and, not but and. I think you can put it in the design. I think you can be on a dating app and the dating app is also inviting you as soon as possible to meet. I think that everyone today that wants to have in-life events is talking about how hard it is to get people to come outside. People are complaining that they feel lonely and disconnected and flat, but they're
Starting point is 00:36:52 not able to actually go outside and re-engage. And I think that if we try, to promise them that they can have the same experience by staying at home as they would have if they left, do the streaming of the concert, do the streaming of this and that, then we are basically going to transform human nature into something else. We are changing human beings. As we have always changed human beings, it's not the first time, but we are changing. Something very profound is switching inside of us. And what happens is that when you and I are concerned about this, we are seen by the techno-optimists as kind of retrograde. You know, we're from the last century. They were nostalgic for a golden era for what it meant to be human 50 years
Starting point is 00:37:38 ago. Who's to say that what it means to be human where everyone's plugged into VR, the metaverse and virtualized relationships and synthetic boyfriends and girlfriends with replicas and chatbots? That why, you know, if people are quote unquote happy in that, who are you, Tristan and Esther, to say that that's bad for people or that's not real? And this is where I think the adult developmental psychology in someone like you who's studied in a deeper way, when you get subterranean, when you get underneath the hood of what's going on for people, there's a lot more loneliness or anxiety that actually, I think those signals of loneliness and anxiety being on the other side of that virtualized environment show us that it's not fulfilling. And there are more
Starting point is 00:38:14 embodied experiences and more complexity and more friction that we know that people who are living in those ways maybe have a lot less of that loneliness and anxiety. But we need to be able to establish more normative facts about human well-being and flourishing so that we can say that, yes, it would be a good thing for technology companies to design and privilege, more friction, more complexity, more in-person. We did the same thing with junk food, right? We first had a promise of repeatable, always available, shelf, stable food, and it was irresistible. And only now are we coming to terms with the fact that we traded away nutrition, right?
Starting point is 00:38:51 in the same way that we're going to trade away real human connection and that the consequences were not very good for our physical health and well it won't be any different when we talk about our relational or our mental health. Okay. Esther, do you have any parting thoughts for our listeners
Starting point is 00:39:05 and closing out this conversation today? I think you're doing something so important, you know, because what you're really asking for is responsibility. You're asking for people who are, the forefront of the changes in our society, in our economy, in our political system, in our humanity, to think responsibly about what they are doing. If when it comes to mental health, it means that we're doing it with caution and with care, and we don't just look at money.
Starting point is 00:39:40 You know, psychiatry is a real complex history of cultures, and you don't just erase all of this and basically just say, oh, come on, if I can promise you a life without suffering, wouldn't you want it? No, actually, because the first philosophical questions that come up is how do you know you're happy if you've never been unhappy?
Starting point is 00:40:02 I mean, we live as dialectic creatures. So I'm in awe of what you do and the relentlessness with which you have been at it since the first time we met, you know, and I've learned a ton from you because I think of it really much more in my own sphere because that's the one I know something about. I take full advantage of many technologies,
Starting point is 00:40:22 but I still very much want to talk to you in person and I want to reach you in your relationship life in person, you know, whereas you are really working with larger systems and the design of these systems. And that is the piece where I don't go because it's not my world. What stands out for me, and this will be one of my parting words actually,
Starting point is 00:40:44 is that I go from one technology conference, to another. I can tell you the word relationship is never mentioned. Never mentioned. There are moonshots about everything,
Starting point is 00:40:56 environment, education, health, transportation, you name it. Nobody ever is looking at a moonshot for relationships. And yet, all these technologies are profoundly affecting how we relate to each other,
Starting point is 00:41:09 how we relate to ourselves, how we form relationships. And I think that it's irresponsible to not pay attention to that part. That's the main piece here that I would like to highlight is it is the challenge that everybody is avoiding, partly because relationships are complex systems
Starting point is 00:41:29 with a lot of contradictions inherently, and they don't just suit one person, but they are at the root of societies. If you don't manage relationships, you don't manage social and political systems either. You wanted to talk to me about politics, I mean, it is about how do you hold two parts, you know, and when you are alone, you usually have to live with those different parts inside of you that pull in different directions.
Starting point is 00:41:59 When you are in a relationship, you often outsource onto the other person, the part of the equation that you repudiate. And when you do it in a society, you do the exact same thing, and it becomes, I am pro-life and you're what, pro-death? I mean, what exactly are we saying here, you know? So if I accepted to do this conversation, it's in large part because I want those of you who are at the forefront of this technological revolution to not forget relationships. It's not enough to just say that GPT4 can be as smart as me and be as creative as me.
Starting point is 00:42:33 It's great. I mean, I'll use any tool that can enhance me. But you have to be responsible and look at the social consequences and the relationship. consequences of what we are creating. Well, I couldn't have said it better. Thank you so much, Esther, for coming on Your Undivided Attention. It's an honor to be your friend, to know you, and to really just admire all the work that you're doing
Starting point is 00:42:58 to raise awareness about the complexity of how we hold relationships and have a more healthy society. So thank you so much. Thank you. Your Undivided Attention is produced by the Center for Humane Technology, a non-profit working to catalyze a humane future. Our senior producer is Julia Scott. Kirsten McMurray and Sarah McRae are our associate producers.
Starting point is 00:43:21 Sasha Fegan is our managing editor. Mia Lobel is our consulting producer. Mixing on this episode by Jeff Sudaken, original music and sound design by Ryan and Hayes Holiday. And a special thanks to the whole Center for Humane Technology team for making this podcast possible. Do you have questions for us? You can always drop us a voice note at humanetech.com slash ask us.
Starting point is 00:43:40 and we just might answer them in an upcoming episode. A very special thanks to our generous supporters who make this entire podcast possible, and if you would like to join them, you can visit humanetech.com slash donate. You can find show notes, transcripts, and much more at humanetech.com. And if you made it all the way here,
Starting point is 00:43:59 let me give one more thank you to you for giving us your undivided attention.

There aren't comments yet for this episode. Click on any sentence in the transcript to leave a comment.