The Truth About Trust | David DeSteno | Talks at Google

Video Statistics and Information

Video
Captions Word Cloud
Reddit Comments
Captions
MING: Good afternoon, my friends. My mane is Ming, and I'm delighted today to introduce my friend the renowned social psychologist David DeSteno and the author of this book, "The Truth About Trust." I first got to know David-- thank you, Dana. I first got to know David through his highly innovative work in studying the science of compassion, which is a topic I'm very passionate about. David and his lab, they are renowned for devising very creative methods in studying how emotional states affect behavior. And they are also known for studying moral behavior in real time. Real time, not fake time like the other labs, real time. David is most interested to figure out how to foster prosocial behavior all around the world. And he told me that in summary his work can be summarised in three words, vice and virtue. Vice and virtue, there's one that I prefer over the other. And with that, my friends, let's please welcome my friend David DeSteno. DAVID DESTENO: And thank you for having me. It's a real honor to be with you here and share this work with you. And as you can probably guess by the title today, I'm going to talk about trust. I think it probably comes as no surprise to you that issues and dilemmas of trust pervade our lives. Trust determines who we want to work with, who we love and who would marry, who we trust to learn from, who we'll go to for support. Now, we all can remember the big stuff, the times trust really matters. Is a new business partner going to be trustworthy, or is he going to skim profits? Is a spouse being faithful or unfaithful? Is a child using drugs even when she swears, trust me. Trust me. Trust me. I'm not. But issues of trust aren't just about those potentially momentous situations. Issues of trust pervade our common daily life. Will my neighbor remember and I trust him to really feed my dog while I'm away, or am I going to come home and the dog's hungry? The mechanic, is he really being honest when he says my car need a new transmission? Was the salesperson when I bought this suit really honest when he told me it makes me look thin? You can tell me. I don't know. I won't comment on that. But whether it's big or whether it's small, what all of these issues have in common is a simple dynamic. They really depend on trust. And we know the more we trust individuals, we can gain a lot more by working together and cooperating. But in reality, as you probably can guess, trust is a double-edged sword. Yes, we can gain more by working together. That's why we have trust in the first place. But trusting somebody also makes us vulnerable to that person. It means that our outcomes are dependent on them being competent, on them having integrity, on them working with us. And so given that trust is so central to human life, you would hope, you would like to think that we really understand how it works, that we can make really good decisions about who we should trust or whether we're going to be trustworthy ourselves. But I'm here to tell you we're not really good about that. And until recently, the science underlying that hasn't been really good about it. And so in some ways, that's what led me to write this book. As a scientist, I really wanted to work on correcting a lot of the misconceptions that are out there to empower people to make better decisions but also so that we can work together to nudge us all to nudge society to become more trustworthy and cooperative overall. So to do that, you have to start like you would with anything else. You have to get rid of the misconceptions and figure out how trust really works. And so that's what I want to talk about today in general. In the book, I talk about lots of issues that I'm not going to talk about today. We talk about issues of how trust affects learning and academic success. One of the best predictors of a child's academic success isn't how much they like their teacher. It's how much they trust their teacher. And they trust that the teacher is competent in telling them and giving them information. We talk about how trust affects our relationships and especially romantic ones and how it can function to smooth out the bumps in those in ways that operate even below our conscious awareness to keep harmony with those we love. In the book, I talk about how trust is affected by power and money. There's great work out there showing that people's trustworthiness tracks socioeconomic status. This is work by my friend Paul Piff. He's a psychologist at Berkeley where he shows that higher SES correlates to increased untrustworthiness. But really it's not about being in the 1%. It's not a birthright of the 1% that makes you untrustworthy. It's simply about money and power relative to those around you. And so any of you, if we put you in a position even for 10 minutes where you feel elevated sense of power, it becomes a lot more difficult in some ways to actually be trustworthy. And also how and when can you trust yourself? It's already February. A lot of New Year's resolutions have gone by the wayside, so if it's a good thing to know. But today what I want to talk about is three broader themes. And the first is what does it mean to be trustworthy and how can we understand how trust operates within ourselves and our own trustworthiness? The second is can we actually detect whether somebody else is going to be trustworthy? In some ways, this has been the holy grail of governmental research and security research. And we've been pretty bad at it. But I have some new data I want to share with you that suggests we can do it. And then finally, the question that's probably closest to my heart in the work that I normally do, which is, how can we increase trustworthiness and thereby increase our own and each other's resilience around us? So let's start with the first question. Most of us, when we think about trust, we think about it as this stable trait. A person's trustworthy or they're not. But I want to convince you that that's probably not the best way to think about it. That's not how it really works. Growing up, we have this idea that it's a typical motif, right? You see it in cartoons all the time. There's an angel on one shoulder and a devil on the other, and they whisper into your ears. And if you grow up listening to the angel, well then you're going to be a good person. You're going to be trustworthy. Everybody is going to love you. Everything's going to be good. There's just one problem with that. And that is if you actually look at the scientific data, it doesn't really hold up. What we've learned over the past decade especially in psychological science is that people's moral behavior is a lot more variable than any of us would have expected. And it's a lot more influenced by the situation. And so if you want to control your own behavior and predict the behavior of those around you, you need to realize that it's not a stable trait. You need to understand how it's affected by the situation. And so my model for understanding trustworthiness, it's better to think of it as a scale, the old school type of type with the plates that go up and down as opposed to a digital one. In any one moment, your mind, whether you know it or not, is weighing two types of cost. It's weighing costs and benefits in the short term versus costs and benefits in the long term. And those usually correlate with what's good for me in an expedient fashion right now versus what's good for me to do even if it costs me in the moment to built a reputation and to build social bonds in the long term. And depending upon the situation, which decision you choose can change from moment to moment. You can think about it. If my friend Ming loans me money, in the moment if I don't pay him back, well, I'm ahead. I've profited in the short term. But long term, it's probably a poor decision because he's not going to give me money again. I'm going to get a reputation as being a cheater. But if I can get away with it, my mind, unbeknownst to me and my own moral codes that I endorse, will try to push me to be a bit untrustworthy. And so I want to suggest to all of you who think this can't happen to me and that you are completely honest and trustworthy and wonderful, it can happen to any of us. And let me show you an example of how it happens and also why you probably don't think it's true of you even though it is. So the first issue is how do you study trustworthiness? I can't really walk around with a clipboard and say, Cindy, are you a trustworthy person? Because what people will probably say is they'll do one of two things. Either they know they're not and they'll, yes I am. Because who wants to say I'm not? But what happens more frequently is they think they are, and they predict they will be, but when push comes to shove, time and again our behavior isn't what we expect. And so the way that we have to study trustworthiness is not by asking people or looking at their past reputations but by staging events in real time as opposed to fake time where we can actually see when push comes to shove, what will people actually do when real rewards are on the line? So let me give you an example of how we do this. So we set up an experiment in our lab to look at this, and it's rather simple. We bring people into the lab. These are normal community members or even undergraduates from the Boston community all known to be trustworthy people. We bring them in and say, look, we've got two tasks that need to be done. One is really long and onerous and it's these terrible logic problems, and circling letters E, and random digit strings, and all the things that you would feel like is a big waste of your time. Or you can do a fun photo hunt on the computer. Here's a coin. I want you to flip the coin, and whatever one you get, it will determine whether you do the photo hunt or the logic problems. And whichever one you don't do, the person sitting in the next room is going to get. And we're going to trust you to do this the right way. Is that OK? They say sure. And then we let them go. What do you think happens? A lot of people just assign themselves to the good task. Any guesses for how many? 80%? Close, 90%. We've done this many, many times. So it's not a fluke finding. We've done it in our lab. Other people have copied the methodology. 90% of people-- well they do one other thing. Some of them don't flip the coin and just say, oh, I got the good task when they come out. Or some of them, because we have them on hidden video, flip the coin repeatedly until they get the answer they want, which is the same as not flipping at all. But they feel better about themselves. And so these are people who we asked them before, if you don't flip the coin, is that untrustworthy? They said, oh, it would be terribly untrustworthy. But they do it. And if you ask them when they come out, we have them rate on a computer how trustworthily they just acted. So here higher numbers mean higher trust on a one to seven scale. So when they're judging themselves doing this, they're above the midpoint. So they say, yeah, it was OK. I was trustworthy. If you take those same people and you now have them watch somebody else do this, they condemn the person for it. That person was not trustworthy. When I did it, it was OK. When that person did it, they're definitely not trustworthy. Now, the interesting thing about this is these were people who are normal people. And so when we see people like Lance Armstrong or Bernie Madoff, you think, oh, it's something wrong. Those people are morally corrupt and untrustworthy. No, well, yes what they did was untrustworthy. But the same process-- on a smaller scale, of course, and we can only study on a smaller scale in the lab-- happens with us. It happens with any of you. Now, the question is, well, why don't we realize this? Why don't we learn to stop trusting ourselves? Well, the reason why is our mind whitewashes our own behavior. So if you ask these subjects, why did you not flip the coin? They'll say things. They'll create stores like, well, yeah, I should have, but today I was late for an appointment. And if I'm not there, somebody depending on me. And so it was OK. So they'll create all kinds of justifications for why it was OK for them in the situation and how it doesn't reflect on the fact that they're an untrustworthy person but they can be untrustworthy. Now, in some way that's a good thing. It has to be adaptive, because if any of us felt like we couldn't trust ourselves, that alternative is much worse. Because it means we're not going to save money for the future because we know future us is going to go blow it a casino. We're not going to diet and take care of our health because we assume three days from now I'm going to gorge on ice cream or chocolate cake. We're stuck with ourselves. If somebody else is untrustworthy, we can stop interacting with them. We can't stop interacting with ourself, and so we need to trust ourselves even when we make mistakes. So that's OK, but what I'm here to do is to help you try and learn that so that you can decrease the probability that you're actually going to make those mistakes. But what I haven't told you yet is that there's any evidence that people actually recognize what they did was wrong. So let me give you an example. So in psychology we had this method which is called cognitive load. And it's a way to kind of tie up people's mental resources so they can engage in rationalization. And the way it works is we give them random digit strings of numbers, say like seven digits. And you have to remember these digits. So what we're doing is you'll get a string of numbers, and you'll have to say 7-6-5-4-1-0, 7-6-5-4-1-0, and then you'll have to answer a question, how trustworthily did you just act? And you have to remember these numbers because I'm going to have you enter them in a minute, and you've got to get them right. And so what this does is it ties up your mind. It prevents your mind from engaging in rationalization. So when we did this experiment again, and we have 90% of people who did cheat even though they said they wouldn't, what you find is on the white bars on the bottom, those who were under cognitive load, there's no difference in how you judge yourself or how you judge others. And those are significantly less. You see yourself as less trustworthy than when you have the time to rationalize. So the second, the moment that you're committing the transgression, your mind knows it. You feel in your gut. You feel that pang of guilt. But what happens is you don't want to think of yourself as untrustworthy. And so your mind engages in this rationalization. The good you tamps away, tamps down the guilt so that it can create a view of you. Well, I had a reason, and it's OK. And I am trustworthy. So the point is to remember that all of us, even if we think of ourselves as trustworthy, I'm sure most of you in general are trustworthy, but your mind is making these calculations. Here when we gave them anonymity-- or at least they thought they were anonymous. They didn't know we have them on hidden video-- their mind's impulses for short-term gain created a story. It pushed them to say, well, I can get away with it now. Even not consciously, it just pushes them to make this decision as an impulsive way. And then they justify it because the long-term consequences they believe are not there, because they believe they're anonymous. Let's turn to the second question. The second question is, can I trust you? How do you figure, how do you determine that question about somebody? Now, as we all know, human society flourishes when we cooperate with each other and when we trust each other. The problem is if one person doesn't uphold his or her end of the bargain, that person can gain at the other's expense. And so what you have is a very dynamic yet delicate balance that we every day have to navigate through and optimize our outcomes. If we make the wrong decision over and over again, we're going to have a problem. So here what we try to do is we try to use people's reputation. Now, as I just told you, reputation isn't a great predictor, and so often we're wrong. But the problem that confronts us other times is sometimes we have to decide if we're going to trust somebody new who we don't know anything about. And we don't know their reputation, yet we're negotiating with them. What do you do there? You have the opportunity for establishing a long-term relationship or you have the opportunity for being screwed over in a way that you couldn't predict. And if you're wrong, well, time and time again that's going to cause you a lot of problems. It's a very non-optimal outcome to be wrong. So given all that, it would be nice if we could actually detect if somebody else was going to be trustworthy. Now, as I said at the beginning of this talk, people have been looking for the Holy Grail of what signifies deception or untrustworthiness for a long time. Is it a true smile? Does that mean I can trust you? Is it shifty eyes? Does that mean I can't trust you? And the TSA spent $40 million on this program to look for these single microexpressions that in GAO testimony before Congress has been shown to be utterly useless. And the problem is I think the reason why we haven't found how we can detect trustworthiness is we've been going about it in really the wrong way. There is not going to be one marker. There is not going to be one golden cue. Cues to trustworthiness are going to be subtle and dynamic. Why is that the case? Well, it's very adaptive if I'm standing here and you're looking and me and all of a sudden I see a major threat behind you to show fear. Because that lets you know even without turning around very quickly, there's something dangerous there. But trust isn't something that you want to communicate very easily or untrustworthiness. Why? I mean, imagine if you're trustworthy person and you had a clear tell. It's like walking around with a big T on your forehead that says, I'm trustworthy. What would happen? Everybody would want to cooperate with you, more of them so they could probably take advantage of you because they know they could. Or if you were untrustworthy and you walked around with a big U on your forehead, well, everybody would ignore you. And nobody would cooperate with you. And your outcomes would be poor. And so trust signals have to be played close to the vest. We have to interact with each other. I can get a feeling for you. You can get a feeling for me. And then we can decide and reveal our cards very slowly. So they're going to be subtle and dynamic. They're also going to be the context dependent. What signals trust in any one specific culture may vary. What signals trust in any one situation may vary. Think about it. There's different kinds of trust. There's integrity. So can I trust that you're going to do the best job you can to help me? Are you meaning well toward me? That's different than trusting your competence. If you don't have the competence to help me, all the intention in the world is going to be useless. And so the cues I look for for competence versus integrity may be very different, and we have to think about that. But the main reason why I think we haven't found the cues to trust is that they're going to occur in sets. I mean, think about it, right? If touching my face means I'm going to be untrustworthy, if I do this, am I doing that because I've got an itch or because I'm going to cheat you? Don't know by one thing. You can't tell. The only way you can begin to read cues to trustworthiness is to look for them occurring in sets so you can disambiguate the meaning of any single one. And that's what the field typically doesn't do. And so I'm going to quickly tell you about two experiments that we did to show how trust can be read. The first one is kind of exploratory. We threw out everything that we had known before, and we simply started to try and identify what cues actually predict real monetary trustworthy behavior and to demonstrate that they do this in an accurate way. And the second part was designed to actually confirm in a very tightly controlled, highly precise way that these are the cues that matter. And I'll show you what I mean by that in a second. We have an exploratory phase and a confirmatory phase. So how did we do this I will start with the exploratory phase. What are candidates for signals related to trust? Well, we brought 86 people into the lab and we put them into dyads, which are groups of two. The only requirement is you couldn't know the person with whom you were now going to interact. We gave them five minutes to have a get-to-know-you conversation. You could talk about anything you want. We gave them a list of topics, but they could talk about anything they wanted. And you're going to play a game for real money, a game that pits self-interest versus being trustworthy, communal interest. And I'll show you how the game works in a second. And then we gave them topics to start, but they could talk about anything that they wanted. So we brought them in. They simply sat across from each other at a table, half the subjects. And we had three cameras on them that were time locked so we could record every single gesture, every single cue they made. Now, we also had another group of subjects who conversed in their get to know you in separate rooms over Google Chat or Gchat-- any type of internet chat. And the logic for this is the same amount of information is being exchanged in the conversation, but in one condition you have access to the person's nonverbal cues. In the other you don't. And then we brought them into separate rooms if they weren't in separate rooms already. And we said, you're going to play this game. We gave each of them four tokens. And the tokens are worth $1 to each of them but $2 to their partner. And so this game is called the give some game. And it's a nice analog for self-interest versus communal interest. Because if you want to be selfish, you can try and get the other person to give you all of his and give nothing. And that means you'll have $12 and he'll have nothing. But the most trustworthy thing to do if you really implicitly trust each other and want to benefit each other is to exchange all you have at the same time, because then you all started with four and now you have eight. And so we had people making real decisions and we paid them accordingly. And we also had them tell us what they thought their partner was going to do. Now, the nice thing about it was whether or not you talked to your partner over an internet chat or face to face, the amount of trustworthy behavior didn't change, which is nice. I think it's because people are now becoming very used to communicating over internet mediated platforms. And so it's not like being face to face made people more trustworthy. There was people who were cheating and being cooperative at equal levels in both cases. But here the axis is error, the amount that you were off. And so lower bars mean accuracy in terms of absolute value. If you were in the presence of the other person, your guess for how much that person was going to be trustworthy or cheat you in absolute dollars was significantly greater. So what this tells us is that people are picking up on a cue. There is some information there that your mind is gleaning from body language, whether you know it or not. And so what we did next was we ran models of all these possible combinations of cues to see what would matter. And the model that predicted untrustworthiness the best consisted of four cues, touching your hands, touching your face, crossing your arms, and leaning away. If you think about it, what does this really mean? Well, we know from the nonverbal literature that fidgeting with your hands and touching your face repeatedly is usually a marker of anxiety and not feeling comfortable. Crossing your arms and leaning away is a marker of I don't want to affiliate with you. Put them together, what does it mean? It means, I don't really want to be with you. I don't like you. And I'm nervous because I'm going to screw you over in a minute. And so none of these cues predict it on their own, but together they did. So the more often you saw a partner show this set of cues, the smaller number of tokens you expected that person to share with you, which meant the more selfish you thought that person was going to be. And the more often you yourself or any subject emitted this four set of cues, the less trustworthy you actually were, the more tokens you kept and tried to get from the other person without sharing. And so in some sense, what we're showing is ground truth here. These cues are predicting actual financial cheating versus cooperative behavior. Now, the most interesting part about it was if you asked our subjects, so what did you use? They had no idea. Or they would suggest it was other cues that didn't predict anything. Yeah, I showed you they were more accurate when they saw the person. And they adjusted their numbers and guesses accordingly. So what this means is your mind is using these cues even though you're not aware of what they are. It's still building intuitions with them. But how do you know those are the right cues? People are doing lots of things. How do I know that when I'm crossing my arms, it's not that my left pupil is dilating and that's the magic cue. Well, being scientists, we needed to have precise control. No matter what actor I had, I couldn't get them to have exceedingly precise control of every expression they're emitting. So what do you do? You need a robot. So this robot, her name is Nexi. She was designed and created by my collaborator Cynthia Breazeal at MIT's Media Lab. And so we simply used Nexi. And I'll show you a video of it in a moment. But the experiment was simple. We repeated the same thing we did before, except we replaced one of the people with the robot. So now you're talking to this robot. And the robot will emit the cues and express the cues that we think signify untrustworthiness or not in a very, very precisely controllable way. The robot was controlled by two people, one who was the voice of the robot, the other who would control whether or not she made the cues. Because you don't want the same person doing it, because they might give cues in their vocal tone. And because I know a lot of you are engineers, I'll give you a quick idea of how this worked. The one person here you can see who is sitting in front of the computer, there's a webcam on her face. As she moves her head, it's gotten by the webcam. The robot's head moves in real time. She's wearing a mic here, so as she speaks, it picks up the phonemes and the robot's mouth moves in real time. The next person controls whether the robot gives these untrustworthy cues or other similar cues. And the third person is our robot mischief person who basically controls and monitors the system. Because every once in awhile, it would go haywire and the robot would like it's doing something crazy, like it's possessed or something. It would break down. But normally it worked wonderfully and just fine. And so we had participants, 65 participants from the community. We brought in 31 of them. We showed that we had seen the cues that Nexi meant untrustworthy. The others didn't. Here's a picture of it crossing its arms. Here's a picture of it touching its face. I'll show you a video in a second. So the first part is people have to get used to the fact that they're talking to a robot. So we had this kind of part where they just acclimated to it. And here's what that looks like. [VIDEO PLAYBACK] NEXI: So my name's Nexi. What's your name? KIM: My name's Kim. NEXI: Kim? It's very nice to meet you. KIM: You too. NEXI: To get started today, why don't I tell you a little bit about myself? KIM: OK. NEXI: I was born and built at the MIT Media Lab two years ago. So I guess in human years, I'm pretty young. But in robot years, that's more like being 20. KIM: (NERVOUS LAUGH). DAVID DESTENO: So you can see she's a little uncomfortable. In fact, we had to put that black barrier on the bottom because people were afraid it was going to go Terminator on them and kill them. So we needed that little barrier. But they quickly acclimated to this, as you'll see. [VIDEO PLAYBACK] KIM: That's basically all I do for fun though. I don't have a lot of time. NEXI: Did you grow up in Upstate New York? KIM: Yeah, I did until I was 18, when I moved out here. NEXI: It seems like that must have been a big transition for you. KIM: It was. It was a really big transition. But I kind of decided that it wasn't the life that I wanted. DAVID DESTENO: So they would self disclose. We heard about pets dying and all these things. One person kept asking the robot if it believed in God. That person was hard. But for the most part, people behave. And just so you can tell what it looks like face on, I'll just show you a 10-second clip. [VIDEO PLAYBACK] NEXI: We all share a big, open room. There are lots of cords and gadgets. So it's probably not like your house. But it's home for me. Why don't you tell me about where you're from? MAN: Well, I was born in Lawrence, Massachusetts, and I had a residency in Somerville right now. I've been doing residential the past four months. DAVID DESTENO: And so we then had them play this game with the robot. We told them, look, the robot's got an artificial intelligence algorithm that it's going to decide how much money it wants to give you and how much money it thinks you're going to give it based on how the interaction went. It didn't, but that's what we told them. And then we asked them questions about how much they trusted the robot, et cetera. So what happened? So to make a long story short, what happened is these are, for those of you who are mathematically inclined, these are standardized regression coefficients. When Nexi made the cues that signaled untrustworthiness in the human to human interactions, people reported trusting it less. Now, the important thing is they didn't report liking it less, because I was worried, oh, they just might think it's doing something weird. No, they liked it equally, but they trusted it less. Now, that's important, because to me that makes it real. Because we all have friends that we like who we wouldn't trust with our money. And so OK, and the less they trusted it, the fewer tokens they predicted Nexi would give them, basically meaning they thought Nexi was going to be selfish and cheat them, and the smaller amount of money in that game they actually gave it. And so what this tells us is that we know these are the cues because we manipulated them with exact precision here while nothing else was happening or things were happening that we could control. And so cues to trustworthiness can be imperfectly assessed, but better than chance. And so the TSA starts to need to look for cues in sets in a context-dependent way. But in some ways, the more interesting part of this is that what it suggests is that technology is now good enough that the mind will now use these cues to ascribe moral intent to robots, or to avatars, or to virtual agents. So while you may not get it from R2-D2, you will probably get it from Wall-E. See, I'll hear aww. Wall-E is not human in the least, but he has enough human characteristics in the eyes and in the hands that he can move them in a way that pings our mind's mental machinery to make us feel trust, or warmth, or compassion toward him. So what does this mean? It's a whole Pandora's box, because in some ways it's good. So for people like Cynthia who want to design these robots and she's working on the smaller ones so that they can actually accompany kids for medical treatment where parents can't go. Think radiation treatments for kids with cancer. They can go with the children. They'll seem more trustworthy, more comforting. But like any other science, it's not good or bad. It depends on the uses of the people who want them. We all know trust sells, so if I'm a marketer, what does this mean? It means that I have the perfect trustworthy or untrustworthy person that I can show you. Because in a human, stuff leaks. No matter how much we're going to try and control it, which is why we could pick up on untrustworthiness. There is no leaking here. We can control everything. And so as we're conversing more and more with automated agents and avatars, our trust is going to be manipulated in ways that we could never have known before or that our mind is not ready to defend against. OK, finally, the last part of the talk, how do we go about enhancing trustworthiness and enhancing the compassion and resilience of each other? To let you know just how powerful this can be and how quickly trust can change and compassion can change, let me give you of my favorite examples. Some of you may know this story. It's called the Christmas Eve truce of World War I. So it was 1914, and the British were fighting the Germans outside of Ypres, Belgium. And it had been a long and a bloody battle. And they were each in their trenches separated by the no-man's land in between. And on Christmas Eve, as the Brits looked across the no-man's land, they started to see lights appear. And then they started to hear songs. And at first, they didn't know what they were because they were in German, and they didn't speak German. But then they soon recognized the melodies. And what they were were Christmas carols. And what happened next was amazing. The men came out of their trenches, and they started celebrating together. They started exchanging trinkets. They started talking about their families, showing pictures, celebrating. Now, these were men who hours ago were trying to kill each other. And no one would have ever trusted if I walked out, was an open shot, I couldn't trust that you weren't going to shoot me. They always had shot each other before, but here they were celebrating with each other in a very communal way, by their own words, very amazing. Here we were laughing and chatting to men who only a few hours before we were trying to kill. Now, if that's not a big change in how trustworthy somebody can be, I don't know what is. So the question is, how do we display such trust and compassion in one moment and such cruelty the next? Because if we can understand that, then we can do something about it. But to answer that question, what we have to realize first is how do we address a different one? How do we identify who is worthy to help? The world is full of more people than we could possibly help. Not that we don't want to help them, but it could be overwhelming. And there's this phenomenon that we know of in psychology called compassion fatigue, which is when you're confronted with people over and over and over again who need help, you begin to dial it down. I have this experience that I'm not proud of when I go with my daughter to New York, and we're walking by, and there was a homeless person. She was like, daddy, help this person. And then I realized that in that moment, I'm completely ignoring this person, because it's a common thing that I face all the time. And if I stopped to try and help every person, it would be overwhelming. And so we have to understand how our mind goes about deciding whose pain is worthy to feel, who it's worth to help, and who it's worth be trustworthy toward. And once we understand that, then we can figure out, OK, how do we increase the number of people to whom we should feel that? Well, one way that I think our mind does it is it uses a simple metric. And that metric is similarity. So it comes back to this is Robert Trivers, who was the discoverer of reciprocal altruism, which is basically the idea, why do we help people in the biological sense? It's I scratch your back today. You'll scratch mine tomorrow. In some ways, that's what similarity is. When there's a lot of people who need my help, going back to that equation of short-term versus long-term gain, who should I help? Who is it worth it for me to help? What your mind does shaped by evolution is it decides the person who is more similar to me is the person that it's worth helping, because that's more likely the person who's going to pay me back and be around later. At least initially that's how it works. And so what we wanted to do was to see how deeply embedded this bias is. If I said to you, if an American soldier's on the battlefield and he comes across an American soldier and a member of the Taliban and both of them are suffering the same wounds, who is he or she going to feel more compassion for? And if I said the American soldier, you might not find that surprising. But what I want to argue is that it's not dependent on longstanding conflict. It's this unconscious computation that your mind makes. And so we tried to strip that down to as basic a level as we could. And we did that by using something called motor synchrony, which is people basically moving together in time. You see it in the military. You see it in conga lines. You see it lots of places. You see it in lots of rituals. And the idea is that if two people are moving together, that's a marker that for here and now, their outcomes, their purposes, their goals are joined. And so we wanted to see if we could actually show this effect at that level. So we brought people into a lab, and they thought it was a music perception study. So they sat across from each other. And there was sensors on the table, and they had earphones on. They didn't talk. All they had to do was as you hear the tones in your earphone, tap the sensor. And so it was constructed so that the two people would either be tapping in time or completely randomly and out of time. That was it. Then what happens is they see the person they were tapping with get cheated. This party is staged, but they don't know it. They believe it's real. They see this person get cheated in a way that makes that person have to do a lot of extra work that they shouldn't have had to do. And then what happens is we give them a chance to decide if they want to go and help that person and relieve that person's burden. And that's what we look at. So what happens? We asked the people, how similar were you to that person in the experiment? The simple act of tapping your hands-- they didn't talk. They didn't do anything-- made them feel that they were more similar to the other person. Now, if you ask them why, they'll create a story. They'll say, oh, I think we were in the same class. Or I think I've met this person somewhere or we share the same goals. They don't know. They never talked to this person before. None of that was true. But because they had this intuition that they felt more similar, they had to create a story for it. How much compassion did you feel for this person when they got cheated and got stuck doing this onerous work that they weren't supposed to do? Remember, in both cases, the amount of suffering is exactly the same. Yet they feel more compassion for this person if they were just tapping in time with them. How many wanted to go help this person? This I found truly amazing. 6 out of 34 people would say, oh, I'll go help that person who was harmed and cheated versus 17 of 35. We had a threefold difference. When you tapped your hands in time with this person, 50% of them said, I want to go help this person who was wronged. That's a huge effect if it's scalable. How much time did they spend helping? These are seconds. So if you tapped your hands with this person, you spent a lot more time knowing that everything you did would relieve that person's burden. And if you look at it-- again these are regression coefficients-- if you tapped your hand in time with this person, yes, you felt more similar to them. And yes you like them more. But what actually predicted the compassion you feel? Not how much you like them but how similar you felt to them If you tapped with them, you felt more similar. That predicted how much compassion you felt toward them even though the level of suffering was the same objectively. And the amount of compassion you felt for them directly predicted how much time, how much effort you put into relieving their pain. Now, what this suggests is that compassion and trustworthiness are flexible. Because if you're going to be trustworthy to me, that means you're going to sacrifice your own immediate outcomes to benefit me like these people did here. Can I trust you to help me? Can trust you not to shoot me back with the Brits and the Germans. Where I live, what this means is trying to solve some of the more contentious things we have in Boston, which is Yankees versus Red Sox. But what that means basically is not thinking about your new neighbor as the guy who hates the dreaded Yankees. Think about him as the guy who likes Starbucks as much as you do. If you can actually retrain your mind to find similarities that you have with people, it will increase your trustworthiness toward them and the compassion that you feel toward them. When you think about social media, there are tremendous ways to do this. We can use the computational power of social media in ways to connect people that have never been connected before. Think about things like profiles on Facebook or other things. We have vast knowledge of what people like and don't like, what they've done or haven't done. Perhaps what you can do is find what people in conflict have in common very rapidly in the background and surface that information to them. And if you do, then it should function in just the same way as tapping your hands. There's nothing magic about tapping your hands. We've done it with wearing the same wristband colors, et cetera. Anything that you can do to highlight similarity with someone will make your goals seems more joined, which will increase the compassion you feel to them if they're suffering, which will increase how trustworthy you are toward them even in ways that don't involve compassion. So at Google, I'm really interested and open to talking with any of you about if you have ideas about using the computational power that you all have to kind of nudge trustworthiness and compassion in the world. But that's all kind of a top-down way. This is a way that we have to remind ourselves, OK, think about this person as similar to me or not. It will be nice if we had a way that could make it automatic, a way that works from the bottom up so that we don't have to stop and remind ourselves. And one way that we can do this-- and I know it's an idea close to Ming's heart, and I'm really honored to be here to be able to talk with him about it-- is mindfulness. If you read the paper or know anything about mindfulness, what you'll know is that it is enjoying a renaissance. And we know it does all kinds of wonderful things. And probably many of you have more experience with it here thanks to Ming's course. It will increase your creativity. It will increase your productivity. It's good for your health. It'll lower your blood pressure. It'll even increase your scores on standardized tests. These are all good things. But if you think about it, it's not what it was originally designed for. If you look at what Buddha said or many of the other ancient meditation teachers-- well this is a quote by Buddha. I teach one thing and one only. That is suffering and the end of suffering. There weren't LSATs and GMATs back then. So all these other things that meditation does are wonderful, and they're great. But one of the main purposes was to foster compassion and end suffering and to increase our being good to each other and being trustworthy to each other. And so what we decided to do was actually put that idea to a test. And so we brought people into the lab. These were people who had never meditated before. They were members of the Boston community. And they were all equally interested in meditation, doing a class for eight weeks. We assigned half of them to actually take a mindfulness class led by a Buddhist lama. And they also would go home during the week with MP3s created by the lama that they would practice. The other half were put on a wait list. So this way we had groups that were equally interested in medication, because you might imagine that if we just recruited people who wanted to meditate, they might have been different types of people in the first place. So both groups were equally interested, but only half of them actually got the course. The other half got it after we did the measure. After eight weeks, we brought them back to the lab. Now, they thought they were coming to have their memory, and their executive control, and all these cognitive measures tested, which we did. But before we did those, what we really were interested in is what was going to happen in the waiting room. And so in our waiting room, we had three chairs. Two were filled by actors and the third was for the subject. And so when the subject arrived, what did the subject do? Well, all them except one sat down. We couldn't get that other guy to sit down no matter what. But most of them sat down. And then a third actor would enter the room. This person was on crutches, had one of those foot boots you wear when your ankle is broken. And as she'd walk down the hall entering the room, she would kind of wince in pain and looked noticeably uncomfortable. And she'd enter the room, and there weren't any chairs. And so she'd lean against the wall. And the question was, what would the person do? The actors were told to busy themselves in their iPhone and to not pay attention. Now, in psychology we call this a bystander effect where this really limits helping. If you're in a situation where you see somebody in pain and other people aren't helping, that tends to decrease anybody's odds of helping, because you say, oh, it's not a big deal or maybe I shouldn't help. And so this situation is one that makes being counted on that you're trustworthy to come and help the lowest it possibly can. So what happens? So people who were in the control group, only very small percentage of them helped, like 16%. Among those who meditated, 50% of them helped. That's a threefold increase. And that's a threefold increase in the situation that is designed to actually work the most against your willing to help. Now, if that can happen after eight weeks and if that is scalable, that is a huge, huge effect that you can count on other people. You can trust them that they're going to help you. Now, why does it work that way? It works that way because one part of mindfulness is this idea of equanimity. And that means realizing that I am similar to you, and you are similar to me. Friends are enemies, are enemies can become friends. And what that does is it trains the mind to see us all as valuable and interlinked. And it breaks down the categories that we put on each other of we're different in religion. We're different in sports teams you like, et cetera. And I think that's why it works. And then it becomes automatic. It does the same thing that my little tapping example was doing. And so when it comes down to it, really what I want to say is that in the end, it's trust or it's dust. And what I mean by that is without trust, our ability to be resilient as a society is exceedingly low. And so how can we build it up? Anything we can do to nudge it up is important to being resilient. In the fall of 2012, I don't know how many of you remember out here, but on the East Coast, super storm Sandy hit New York. And it was a devastating storm. And there are neighborhoods that still aren't recovered. But the AP did a great study. Controlling for the amount of damage that occurred, they looked at what was the single most important predictor of a neighborhood's resilience. The single most important predictor was how much neighbors trusted each other. How much they know that the other person they could count on them, that that person was going to have compassion for them, that they were going to work together. The neighborhoods that were higher in trust were the neighborhoods that got up and running in terms of commerce, and support, and social services faster than anything. And that's why I say in the end, it really is trust or dust. If we don't trust, we're harming everybody. But of course there were people in neighborhoods who price gouged and who did things they shouldn't. And so really my message in the book is, yes, trusting is good. We should all trust. But trusting wisely is better. And so it's my hope that any of you who read this or come into contact with this work, it will empower you to think about the way trust actually works and the forces that impinge on it to make better decisions about who you can trust but also how to foster trustworthiness in yourself. And I thank you so much for listening to me. MING: Thank you, my friend. We have time for questions. Anybody have any questions? AUDIENCE: Two fairly related questions, in the earlier test about looking for cues, so I'm just wondering how you came to the domain of different things you were looking for that could conceivably be a cue in your analysis. Because you could say whether the pinkie is touching the hand and the hand's touching is or not. And then along with that, with whether you were going into it starting off focused on physical cues or if you were also considering the types of issues that were brought up in conversation, which could then play into the talking with a robot or talking over the internet. DAVID DESTENO: So let me do the second one first. We were primarily interested in physical cues. There's lots of work out there as well on linguistics and the type of phrasing that people use as well as vocal tone. We weren't looking for those. It doesn't mean that those don't matter. And so I'm not saying these are the only cues that matter, but these are sufficient to predict. The more that we know about, our accuracy will go. But we were interested in the actual physical, biological motion cues. How did we get them? We simply started with the brute force method, was looking for individual cues that had some predictive ability at all. Because even if they're not predictive greatly on their own, they have to have some predictive power on their own. And then we would begin to do is to assemble different subsets just trying to maximize the amount of accuracy that we can predict. So it was a very bottom-up approach. And then once we have those four sets, those predicted the greatest amount of variance in people's selfish monetary behavior. Which is then why, again, it was really important to use the robot. Because you're right. This was a correlational method. Who knows what we could be picking up. Maybe on every time I cross my arm, it was my pinkie. And so we could actually manipulate it with precision with the robot to validate it. MING: If you could put a pinkie down here, it's not trustworthy. That's what I learned from a movie. AUDIENCE: Let's see if I can word this the right way. It seems like the general conclusion from the research or your conclusion is you're saying we should be more trusting of others, like sort of the hope for the better world. And the question is, is there also then some drive for people themselves to be trustworthy? Like in the example you gave in the beginning about people in a marriage, one cheated on the other, is it to say, put that aside. Trust that person. Or is there some other conclusion in that sense? DAVID DESTENO: It's a good question. What we know from all the-- so people like Martin Nowak at Harvard is a straight mathematician, evolutionary biologist. And so they run these fantastic models. And what we know is that if you are untrustworthy in the short run, you will profit immensely. But over time, that profit then starts to go down. And so in the long run, people who are trustworthy profit the most in terms of everything and even as a group. And so we know that's the better outcome. But if you can be untrustworthy and not get caught, you're going to profit. So how do we try to balance those? And so what we're trying to do is to make everybody want to be more trustworthy but at the same time also make better decisions. It will be impossible to have a world where everybody is trustworthy. Because if everybody's trustworthy, you stop even looking and caring. You just automatically say, yes, I'll trust you. And then if there's a mutation or whatever that causes people to be more untrustworthy, they're going to profit like crazy. Until then everybody starts caring, and so it's always going to be in an equilibrium. The question is, can we increase the set point for trustworthiness to a higher level? And so it's about increasing your own trustworthiness but about deciding if you can trust somebody else wisely. So yes, if you know absolutely nothing, it's better to trust than not trust in the long run in terms of quantifying the benefits that can happen. But it's certainly not as good as making an informed correct decision. And so my hope is to try and open people's eyes to how trust really works so that you can make better decisions. AUDIENCE: So similar to that question, trust over time, have you done any research into how analysis of trust has to change, how much it needs to be dependent on data changing? Like the first study was on the initial get to know you, how much I trust you as a person. And then the question is later on, something happens. How much should future events be added into that evaluation? DAVID DESTENO: You mean at what point will I change my judgment of whether I can trust you? It's dependent on a lot of things. It's often dependent on the magnitude of what you have held up your end for what you haven't. But I guess my argument is that you need to look at each situation if it's important and new. Because even somebody who has been always trustworthy, if the costs and benefits change-- the reason they're always trustworthy is the cost and benefits are rather stable. Take that person and change the cost and benefits either by dangling a reward that is immense in front of them or giving them anonymity so they won't get caught, like our people, and they'll change. So I think there's not a clear time frame. I think we all adjust at different rates depending upon the magnitude. But my message is no matter what you think, consider the situation. If it's somebody you always trusted, consider has their power changed? Has anything else changed? Because they may not want to be untrustworthy just like our subjects did, but they will be and construct a story for why. AUDIENCE: Hi and thank you for coming. I have a question actually kind of related to that. So have you done any study in how people's relationship long term potentially changes how sensitive or not sensitive they are to social cues? I can imagine someone who perhaps does all of the social cues that you mentioned in terms of lying, but perhaps like a brother and sister, for example. They've gotten numb to it over time, and maybe they can't pick up on it anymore. Or do you have a sort of sense of how long-term relationships can change how people pick up on that? DAVID DESTENO: Two things on that, we haven't done that work, but what we know from the nonverbal literature in general is that people have a what are often termed accents. So there's a kind of panhuman way of doing it. But then different cultures or even different families or individuals will have modifications of that. And so the longer you are with someone, you can learn that for this person, this is that person's tell in some ways. And it will be some combination of these. But other things added to that will increase your accuracy for that person. But another thing in terms of long term, what does trust do, it's beneficial. So there's great work done by Sandra Murray. She's psychologist who studies romantic relationships. And one thing that the trust does in our relationships is it smooths out the bumps, as I said. So we've probably all had times when our significant other does something and it makes us go hmm, whether it's you think the person's flirting with someone or they're working late or what is it. Well, if you inherently trust this person, at a very nonconscious level, that trust erases that hmm. It just gives you an intuition that everything's fine. And if you trust that intuition, that's good. Because lots of times we'll do something that we're not trying to be untrustworthy. It's just an inadvertent thing. But if a person interprets that as, oh, you're untrustworthy, it can start you into kind of a death spiral. And so the good thing about trusting someone over the long time in a relationship is that it helps smooth out those bumps so there aren't mistakes made. So that one person doesn't interpret the other person as flirting or doing something with somebody else that they shouldn't have. Now, if they keep doing it, well, then you're going to know it's real. But that's a benefit of trust long term. MING: Thank you, my friend. So the book is "The Truth About Trust" available where books are sold, also available at the back of this room. And David will be around to sign books. And my friends, David DeSteno. DAVID DESTENO: Thank you all.
Info
Channel: Talks at Google
Views: 19,790
Rating: undefined out of 5
Keywords: talks at google, ted talks, inspirational talks, educational talks, The Truth About Trust, David DeSteno, how does trust work, trust fundamentals, trust in the modern era, trust basics
Id: b_AtRE0hDbw
Channel Id: undefined
Length: 54min 37sec (3277 seconds)
Published: Thu Feb 27 2014
Related Videos
Note
Please note that this website is currently a work in progress! Lots of interesting data and statistics to come.