Hidden Brain - Ep. 64: I'm Right, You're Wrong
Episode Date: March 14, 2017There are some topics about which it seems no amount of data will change people's minds: things like climate change, or restrictions on gun ownership. Neuroscientist Tali Sharot says that's actually f...or good reason. As a general rule, she says, it's better to stick to your beliefs and disregard new information that contradicts them. But this also means it's very difficult to change false beliefs. This week, we look at how we process information, and why it's so hard to change our views.
Transcript
Discussion (0)
This is Hidden Brain, I'm Shankar Vedenthu. There are certain issues where we just don't
see eye to eye with people on the other side of the political aisle. Often it feels like
we don't just disagree over policy, we can't even agree on the facts.
I think people are mostly just setting their ways.
I would say I'm right objectively, The facts prove that there is global warming.
I don't know how you can dispute it.
But everybody thinks they're right.
I think they're wrong.
I think they're making a huge mistake because security has to be a very big issue.
My guest today has spent years studying the way we process information and why we often reach biased conclusions.
She says it's surprisingly difficult for us to change
one another's minds no matter how much data we present. But just a little bit of emotion,
that can go a long way.
Tali Sharot is a cognitive neuroscientist at University College London.
In her upcoming book, The Influential Mind, she explores how our minds work, why we ignore
the facts, and how you can get people to actually listen to the truth.
Tally, welcome to Hidden Brain.
Thank you, great to be here.
Some months ago, you were listening to a Republican presidential debate, and candidate Donald
Trump was asked a question
about the safety of childhood vaccines.
Here's what he said.
You take this little beautiful baby and you pump.
I mean, it looks just like it's meant for a horse, not for a child.
And we've had so many instances, people that work for me just the other day.
Two years old, two and a half years old, the child, the beautiful child went to have the vaccine
and came back and a week later got a tremendous fever, got very, very sick, now is autistic.
Tali, you're a mom. You have two small children. At the time, Trump said this, one of your kids was two
years old, the other was seven weeks old. Describe your emotional reaction to what he said.
weeks old, describe your emotional reaction to what he said. Yes, so when I was listening to Trump at that debate, I was already quite concerned because
I was a mother of a small child, so you're always concerned about the health and safety
of your kids, especially when they're only a few weeks old.
And so it really tapped into this fear that I had and the anxiety that I already had.
And when he talked about this huge syringe, a horse-sized syringe that was going to go into
the baby in my mind, I could imagine this syringe inserted into my small little child and
all the bad things that could happen.
And this was a very irrational reaction on my end because I know that there's not an actual link between
autism and vaccines, but it's not enough to have the data. Ben Carson, Dr. Ben Carson,
was on the other end. Well, let me put it this way. There has been numerous studies and
they have not demonstrated that there's any correlation between vaccinations and autism. But that wasn't enough because the data is not enough.
And even if the data is based on very good science,
it has to be communicated in a way that will really
tap into people's needs, their desires.
If people are afraid, we should address that.
I'm curious when you sort of contrasted, you know, the weight of the evidence on the
one hand and this very powerful image of the horse syringe and your seven-week old
baby on the other hand, how did you talk yourself into trusting the data over that emotional
image?
What really helped is that I understood what was happening to me. You know, the first
instinct was like a stress and anxiety, but because this is what I study, I knew what
my reaction was. I knew where it was coming from. I knew how it was going to affect me,
and I think that awareness helped me to put it aside and say, okay, I know that I am anxious for the wrong
reasons, and this is the action that I should take. It's a little bit when you're on a plane,
and there's turbulence, and you get scared. But telling yourself, I know that turbulence
is not actually anything that's dangerous. I know the statistics on safety and planes and so on, it helps.
It helps people reduce their anxiety.
The facts don't always relieve our anxieties, though. Sometimes, they only harden our views.
Some time ago, Tally did a study where she presented information to people who believe that climate
changes real and to people who are skeptics.
She found, for both groups, people strengthened their pre-existing beliefs when new information
confirmed what they thought, but both groups ignored information when it challenged their
views.
I asked Ali about this.
Our psychological biases are the same across individuals on average.
We all have what's known in a confirmation bias.
A confirmation bias is our tendency to take in any kind of data
that confirms our prior convictions
and to disregard data that does not conform
to what we already believe.
And when we see data that doesn't conform to what we believe,
what we do is we try to distance ourselves
from it. We say, well, that data is not credible, right? It's not good evidence for what
it's saying. So we're trying to reframe it to discredit it.
I want to sort of push back at you just on one point, which is sometimes it seems to me
that it actually might be rational to reject
data that comes in.
You know, if I've seen gravity work all my life, it makes objects fall toward the ground.
And you take me into a room where an object seems to be levitating.
It seems to me the appropriate reaction is not okay, everything I knew about gravity
was wrong, but let me try and understand this anomaly.
Let me try and understand why this object is not doing what it's supposed to do. It seems to me, and I'm clearly betraying
my own beliefs here, given the overwhelming weight of evidence that climate change is
real, showing me data that one winter is colder than normal shouldn't make me change my overall
view. Should it? You're absolutely right. So the way that people tend to update their beliefs is that they use the
new information in light of what they already believe because that on average is the rational way
to go. Exactly as you said, when we encounter some kind of information that really contradicts what
we believe very, very strongly on average that information is wrong. So I give an example in my book where if someone comes in and says,
I just saw pink elephants flying in the sky, and I have a very strong
belief obviously that no pink elephants fly in the sky.
I would then think that they're either delusional or they're lying, and there's good reason
for me to believe that. So it's actually the correct approach
to assess data in light of what you believe.
There's four factors that determine whether we're going to change our beliefs.
Our old belief are confidence in that old belief, the new piece of data, and our confidence
in that piece of data.
And the further away the piece of data is from what you already believe, the less likely
it is to change your belief, and on average, as you go about the world, that is not a
bad approach. However, it also means that it's really hard to change false beliefs.
So if someone holds a belief very strongly, but it is a false belief, it's very hard to
change it with data.
When we come back, we're going to talk about some of the ways we can reshape the beliefs
of other people or our beliefs and try and answer the question, if throwing good information
at people doesn't drive misinformation out of circulation, how do you get people to
buy the truth?
Stay with us. In often fields as though fear is used to motivate us to act.
The vast majority of Americans today do not feel safe.
On Sunday, Americans woke up to a nightmare that's become mind-numbingly familiar.
This could be the great Trojan horse of all time.
Politicians use fear to get us to vote.
TV programs use fear to get us to keep watching.
Public health officials use fear to get us to quit smoking.
I asked Tally whether fear might be an effective way
to persuade people to change their minds
and maybe even their behavior.
Fear works in two situations.
It works when people are ready, stressed out, and it also works when what you're trying
to do is get someone not to do something, an inaction.
For example, if you try to get someone not to vaccinate their kids, fear may work.
If there's, you know, an apple that looks bad, I don't eat it.
Fear is actually not such a good motivator
for inducing action, while hope is a better motivator
on average for motivating action.
You talk about one study in your book
where a hospital managed to get its workers
to practice hand hygiene, to get its workers to practice hand hygiene
to get staff members to wash their hands regularly.
But it turned out the most effective thing was in frightening the staff about the risks
of transmitting infections.
It was something else.
So in a hospital on the east coast, a camera was installed to see how often medical staff
actually sanitize their hand before and after entering a patient's room.
And the medical staff knew that the camera was installed, and yet only one in ten medical staff sanitized their hands before and after entering a patient's room.
But then an intervention was introduced, an electronic board that was put above each door, and it gave the medical staff in real-time positive feedback.
It showed them the percentage of medical staff that washed their hands in the current shift and the weekly rate as well.
So, any time a medical staff will wash their hands, the numbers will immediately go up and there will be a positive feedback saying, you know, good job. And that affected the likelihood of people washing their hands significantly.
It went up from 10% to 90% and it stayed there.
Instead of using the normal approach, instead of saying, you know, you have to wash their
hands because otherwise you'll spread the disease, basically instead of warning them of
all the bad things that can happen in the future,
which actually results in inaction, they gave them positive feedback.
One important idea that Tally has explored is something known as the Equality Heuristic.
It's a mental shortcut, and it's really quite simple.
We tend to assign equal weight to everyone's opinion.
But sometimes, when there's an expert in the room,
this mental shortcut can lead us astray.
Different people have different expertise,
and it's better to put more weight on people
who are more knowledgeable or have more expertise
in the domain that we're making the decision in.
And there's been a study showing that this equality
heuristic is something that people do around the world. So it's been a study showing that this equality heuristic is
something that people do around the world. So it's not something that people do
only in democratic governments, but studies have been conducted in other
countries such as China and Iran, and there too people go according to the
equality heuristic. If they need to make a decision, they will get the opinions
of quite a few individuals
and then tell you them up and that's how they make the decision.
Instead of actually using the person in the room who is more knowledgeable and has more expertise.
But how can you know who the expert is?
When your question is about cancer, you can turn to a non-collegeist.
But for other questions, many people might claim
to be knowledgeable. It turns out there's a clever technique to separate the experts from
the pretenders. It's called the surprisingly popular vote. Let me show you how it works
with an example. I have two questions for you. The first is, what's the capital of Brazil?
Okay, here's the second question. What do you think most people will say is the capital of Brazil? Okay, here's the second question.
What do you think most people will say is the capital of Brazil?
Maybe you thought the capital was Rio de Janeiro, but maybe you knew the right answer, Brazilian.
By looking at the results of those two quick polls, I can tell you with no prior information that the capital of Brazil is Brazilian.
How would I know that?
Because if I think Rio de Janeiro is the capital of Brazil,
I will also think that most people think that.
However, if I think Brazil is the capital of Brazil,
I will still think the Rio is the answer
that most people will give.
So what this ends up being is that Brazilia will be an answer that's more popular than people expect.
I have to say it took me a minute to understand how this idea works. It's complicated, but it really
is very clever. Here's another example. Let's look at a math problem. In a lake, there's a patch of
lily pads. Each day, the patch doubles in size. If it takes 48 days for the patch
to cover the entire lake, how long would it take for the patch to cover half the lake?
The correct answer is 47 days, because the lilies double in area every day. If they fill the lake
on day 48, they will fill half the lake on day 47.
But the intuitive answer is the 24th day,
because 24 is half a 48.
If you ask people what the answer is,
most people will say 24, a few people will say 47.
When you ask people what they think others will say,
nearly everyone will predict that others will say 24.
Since more people say the answer is 47, then expect others to say the answer is 47.
47 turns out to be the surprisingly popular vote answer and the correct one.
It's basically a clever way to figure out who has expert or insider information.
Betting markets work on a similar idea.
If there's someone in a group who has knowledge about something and you give them a financial incentive to come forward
with that information, you can bring insider information to the surface.
So the surprising popular vote tends to predict the correct answer in many
different domains. It can be factual, it can be math, it can be even estimating the price of art.
So they did a little study where they asked MIT students to
estimate the price of art. Now, MIT students, for the most
case, they don't really have a lot of knowledge about art.
And yet, when they asked them to estimate the price of art and
also to say what other people will think about that, and they
took the surprising popular vote, the surprising popular vote, the spot on.
I wrapped up my conversation with Tali by exploring one last idea about how we might convince
others to listen to the truth.
It had to do with a study of Princeton students who got their brain scanned while they listened
to a story of a young woman named Annabel.
Producer René Clarre read an excerpt from that story.
I know everyone has some crazy prom stories, but well, just wait.
I was a freshman in high school in Miami, Florida, and I'm new to the freshman scene.
I'm new to the high school scene, I should say.
And it's almost December, so I've been in high school for about three months, and this boy
Charles asks me out. He's British, he's a junior, so I've been in high school for about three months, and this boy Charles asks me out.
He's British, he's a junior, and he's really cute.
But sort of shy, but just, well, it doesn't matter.
So I say yes, I'm excited.
The story goes on. It includes, as Tally writes,
love, rejection, blood, alcohol, and a couple of policemen, all the requirements of a bestseller.
I asked Tally to tell me what happened in the minds of the Princeton students as they listened to that
story. So this is a study that was conducted by Professor Hassan's group at Princeton.
And what they showed is that when one person listens to another person's story, the activity in the two brains synchronize.
So if you're listening to me, then it is possible that the activity of your
brain looks a lot like the activity in my brain. And at the beginning, the
activity in your brain, listening to my brain, it will be delayed in a little bit,
right? So I'm thinking about something, I'm saying the words, you're
perceiving the words, and then the activity in your brain will follow the activity in
mind. However, they found in that study that after a while, the brain of the listener started
proceeding the brain of the storyteller. Because the brain of the listener was predicting
what the storyteller will say.
So if you looked at the listener brain, you could predict what the activity in
Annabelle's brain would look like. But the idea of synchronization seemed to be
important for influence. Why? In another study that the same group did, they found
that when people were listening to very strong speeches, what they found was
the brains of the different people listening to those speeches started synchronizing.
So if we all listen, for example, to Kennedy's famous moon speech,
her brains would likely look very much alike.
Those who came before us made certain that this country rolled the first waves of
the industrial revolution, the first waves of modern invention, and the first wave of nuclear power.
And this generation does not intend to founder in the backwash of the coming age of space.
We mean to be a part of it. We mean to lead it."
And this is not only in regions that are important for language and hearing. It's
also in regions that are important for emotion, in regions that are important for
what's known as fear of mind, our ability to think about what other people are
thinking, in regions that are important for associations. And you try to think,
well, what's common to all these influential
speeches that can cause so many people's activity to synchronize? And one of the most important things
is emotion. If the storyteller or the person giving the speech is able to elicit emotion
in the other person, then he's actually having somewhat of a control on that person's state
of mind.
So, think about it like this.
If you're very sad and I'm telling you a joke, well, you're sad, so you're not going
to perceive the joke as I perceive it when I'm happy.
But if I'm able to first make you happy and then tell you the joke, well, then you perceive
it more from my point of view.
So by eliciting emotion, what you're able to do is change the perception of everything
that comes after.
To perceive information as the person who's giving the speech wants you to perceive it.
So you can see how this coupling, this idea that the audience is mind and the speaker's
mind are in some ways coupled together.
You can see how this could potentially be used to spread good information.
You have a great teacher in high school and you're captivated by the teacher and you're
being pulled along by the story the teacher is telling you, maybe about history or maybe
about geography.
But you can also see equally how the same thing can work in the opposite direction.
That you could be listening to a demagogue or you could be listening to somebody who has a very sort of captivating rhetorical style.
And this person could also lead you astray in just the same way that the whether we will be influenced by one person
or ignore another person are the same whether the person has good intentions or bad intentions.
The factors that affect whether your influential can be, can you enlist it in motion in the
other person? Can you tell a story? Are you taking into account the state of mind
of the person that's in front of you?
Are you giving them data that confirms
to their preconceived notions?
All those factors that make one speech more influential
than the other or more likely to create an impact
can be used for good and can be used for bad.
Tali Sharad, I wanna thank you for joining me on Hidden Brain today.
Thank you so much for having me.
This week's episode was produced by Maggie Penman and Raina Cohen and edited by Tara Boyle.
A special shout-out to our former intern Chloe Connelly,
who interviewed the Liberals and Conservatives we heard at the start of this episode.
Our staff includes Jenny Schmidt, Parts Shah and Renee Clar.
This week our unsung hero is Brian Malford of National Public Media.
NPM is the group that sells our sponsorship messages.
Brian is a great mix of liberal and conservative.
He's always liberal in his encouragement and conservative in his promises.
As the saying goes, he under promises and over-delivers.
Thanks Brian. I'm Shankar Vidantum and this is NPR.