No Stupid Questions - 104. How Simple Is Too Simple?
Episode Date: June 19, 2022Why are humans so eager for magic-bullet solutions? Can you explain how a pen works? And how does Angela feel about being forever branded “the grit lady”? ...
Transcript
Discussion (0)
Whoa, that's a lot to spend for mechanical pencils.
I'm Angela Duckworth.
I'm Stephen Dubner.
And you're listening to No Stupid Questions.
Today on the show, why do humans tend to choose simplicity over complexity?
I kind of feel like we need the opposite of Occam's razor.
Maybe we could call it the Dubner-Duckworth razor.
Occam's razor. Maybe we could call it the Dubner Duckworth razor.
Angie, I have a question for you that is both a straightforward question and, I suspect,
a bit of a trick question. Are you intrigued?
I am. What is it? Is grit the key to success? Or maybe I should say, is grit the key to success? How do you answer that?
pointed out that's overselling and unlikely to be true. I do not think, Stephen, that grit is the key to success in the sense that I don't think grit is the only determinant of success.
I didn't choose the name of my TED Talk. People may not know this, but when you write an op-ed,
the name of my TED Talk. People may not know this, but when you write an op-ed,
say, for example, the New York Times, you don't get to pick the title.
If you're lucky, they'll run it past you and say, can you live with this? But likewise, I just gave a talk and they named it The Key to Success. I emailed thepowersthatbe
at TED and I said, titling this The key to success is really not an accurate title. And
they just changed it to my definition of grit, passion and perseverance for long-term goals.
And I can tell you the epilogue of what happened after that interaction. I've been thinking about
how the human mind likes for there to be the key to weight loss, romantic satisfaction, a happy life, fill in the blank. We love magic bullets.
So I think the idea of overselling is a really interesting one from a psychological point of
view. Like, why do we oversell things? Why are we so eager to be oversold?
That's what I really wanted to talk about, this embrace of the magic bullet way of thinking about the world, because I see it kind of everywhere I look.
A lot of the work that I do with Freakonomics Radio, or at least what I try to do, is to cut against that, to show that while there is a virtue to simplicity, the world is complex.
Moreover, even if a simple solution works, it will probably fail to work forever.
Or for everyone.
I was just reading Max Bazerman's new book.
I don't think it's out yet, but you know Max Bazerman, yes?
I do.
He's a friend of ours.
He's a professor at Harvard Business School.
A friend of ours, not in the Cosa Nostra way that a friend of ours might imply.
Wait, what is Cosa Nostra?
Is that like an Italian restaurant?
That's kind of the mafia.
Oh, is that what it really is?
In the lingo, at least in the mob movies that I've seen, because I've never been in the
mob.
When you say he's a friend of ours, that means he's one of us.
Well, I can't think of somebody who's less likely to be in the mob.
Can you think of anyone who's less likely to be in organized crime than Max Bazerman?
I will say this, though.
His name is the perfect name for the Jewish guy that's in the mob,
because there's usually one or two Jewish guys like Hesh from The Sopranos.
So Max Bazerman definitely sounds like a badass.
I say that in an admiring way.
But anyway, you were saying this non-badass Max Bazerman.
Professor Max Bazerman at Harvard Business School.
Again, professor is the perfect nickname for a mobster.
They call him the professor.
And he has this new book called Complicit,
How We Enable the Unethical and How to Stop.
But there was this one
passage that's so related to this magic bullet psychology. I'm going to read you what he said
about teaching a group of executives. He says, I was teaching this group of executives online.
I asked the class, what caused the massive fraud at Theranos? Each executive was asked to enter
their answer in the chat function.
The time allowed was generous. Clearly, there were multiple causes of fraud at Theranos.
But when I looked at the class's answers, I found they tended to be simple and singular.
62 of the 70 responses offered a single cause. 56 of those 62 were a simple description of Elizabeth Holmes, such as her
ego or lack of integrity. We should say Theranos is this company that turned out to be mostly
fraudulent that was supposed to do really quick and easy and cheap blood testing to give all
kinds of diagnostics. Like a single drop of blood and we could tell you everything you need to know.
The point here is really less about Theranos and more about the psychology of these executives who thought of a cause, a reason, not multiple
reasons. And here is his conclusion. In other words, the vast majority of the sophisticated
executives in my class who knew all aspects of the story blamed a single source. And then Max goes on to talk about the
single cause bias. We have a propensity to stop thinking after we see or understand one cause
of an effect when if you just pause to think about it, really almost everything, like why did Angela just get a little mini can of Diet Coke
out of the refrigerator? Is there one cause or is it multiple causes? And it's hard to even think
of a single phenomenon or human behavior that doesn't have many causes. It's so interesting
because what's pushing against his word of caution is a long history of embracing simplicity as a means to figure out a
very complex world. Occam's razor is a famous example. When there's a complicated explanation
and a simple explanation, let's assume the simple explanation is right. Aristotle and Kant and many,
many, many people have praised the virtue of simplicity. So how would you suggest that we think about
the relationship between simplicity and truth? It's an interesting question of where Occam's
razor comes from. This might be apocryphal, but wasn't Occam a friar or a monk, right? I think it
was Brother Occam. Don't know. Parsimony was supposed to be not only preferred aesthetically, but more likely to be true.
That was the idea that nature is simple, which again, if you ponder almost any aspect of human
behavior, you're like, wait a second, the complex explanation ought to be right. But there really is
a bias towards simplicity.
ought to be right, but there really is a bias towards simplicity.
I think what this is really about is just understanding how cause and effect works.
Maybe it's our fault for using that phrase cause and effect as opposed to causes and effects.
Both plural.
I'll give you an example of where I first really began to wrestle with this. So years and years and years ago, when I first met Steve Levitt, he had written this paper along with the legal scholar and economist John Donahue, which argued
that surprisingly, one of the biggest drivers of the drop in crime in the US starting in the 1990s
was the legalization of abortion by the US Supreme Court. Supreme Court, Roe v. Wade, in 1973.
And their argument, which is probably familiar to many people who listen to this show,
was that what the legalization of abortion did was provided an avenue for parents,
particularly women, who decided that now was not a good time to have a baby.
And it turns out that if you don't want a baby at a given time,
that's a pretty good signal that baby will not have the greatest odds for success in life.
They basically argued that unwantedness is an undesirable component for a child. Fewer unwanted children meant that the small share of those children who might have become criminals
children meant that the small share of those children who might have become criminals was declining even more.
And so that's what led to a drop in crime.
Now, let's put aside the fact that it's a fairly nuanced argument.
It's a fairly controversial argument.
When Leva and I wrote about that argument in Freakonomics, we were very careful to write
it in a way that tried to make clear that this was
not an argument about the upsides or the downsides of abortion per se.
We tried to write it as an argument about how it can be that policy leads to behaviors
that lead to downstream unintended consequences.
We also went out of our way to say, look, if your first instinct is to
lower crime, that abortion is a terrible way to do it. Putting aside any moral or ethical questions
about it, it's extraordinarily inefficient. So this was not a utilitarian argument.
You're not making a policy recommendation. You're trying to explain
what accounted for a pretty remarkable drop in
crime. We wrote about this in Freakonomics in the context of all the other factors that either did
or did not decrease crime. And in fact, several years after the abortion paper,
Levitt wrote another paper called Understanding Why Crime Fell in the 1990s, four factors that explain the decline and six that do not.
This is one of the reasons I really fell in love with economic research and with Levitz in particular.
It really acknowledges like those are 10 factors right there.
Were there three other than abortion?
There were three others in abortion.
That's really my point here.
Wait, I need to know what they are before we continue this conversation. I can't not know. One factor that did cause a drop in crime was
a big increase in imprisonment. In other words, if you put a lot of people in prison, which we did
in the 80s and then 90s, that is definitely going to decrease the crime rate because this is a basic
tenet of criminology. It's a relatively small group of
people who do the majority of crimes. So if you're very, very tough on them, you convict them and you
put them in prison for quite some time, then crime will go down. And I think I found the other two.
One I want to say is more policing. Yes. Increases in the number of police. I'm not a genius. I'm
just using Google. All right. Tell me what's the fourth that matters. The fourth one is the waning of the crack epidemic. Crack was a really interesting
drug and market. Crack created a lot of violence. It was so profitable that the gangsters who were
selling it fought over the rights and they killed a lot of people. The really interesting question
is why did crack mostly go away? It's hard to argue this empirically, but one good answer seems to be that crack is such
a terrible drug.
It's so damaging to a person, both physically and mentally.
If my uncle is a crackhead or my older brother is a crackhead and I see what's going on with
them, I'm like, no way.
I might use heroin.
I might smoke weed.
I might drink a lot.
I'm not doing crack.
Even if someone reads Freakonomics, where we actually walk through this paper of Levitt's
and say, here is evidence that there were four pretty major contributors to the drop in crime
and six contributors that you might think had contributed. Those include a stronger economy,
innovative policing methods, changing demographics, gun control laws,
carrying of concealed weapons, the use of capital punishment. Those were some that Levitt
empirically argued didn't decrease crime for a variety of reasons. It is astonishing to me how even someone who's
read that fairly carefully seems to gravitate toward the magic bullet or single cause explanation
and say, oh, it was abortion. By the time they get to their next dinner party, it's the only
thing that's left. They're like, oh, did you know that crime's down because of abortion?
Exactly. and they ran this tournament giving this like huge team of scientists almost all of the data
that was there but they withheld some of the outcome data from age 15 and the question was
how well could they predict those outcomes from everything else in the data set and you might
think pretty well we have really powerful computers and they have all these theories
and they understand human nature. But the conclusion
of the paper, and I'll just read you verbatim because it's so compelling, despite using a rich
data set and applying machine learning methods optimized for prediction, the best predictions
were not very accurate. Overall, these results suggest practical limits to the predictability of life outcomes.
And I think that is very similar.
I wrote a paper predicting outcomes at West Point.
In the study, I do show you can predict West Point graduation from grit.
But here's the last line of my paper.
There may be severe limits to how well any set of personal attributes can forecast an individual's destiny.
People change, context change, life trajectories are shaped by the whims of chance and path dependency.
In other words, it's complicated.
Occam's razor was not right.
I kind of feel like we need the opposite of Occam's razor. Maybe we could call it the Dubner-Duckworth razor, that given
a simple explanation and a complex one, that you should choose the complex one.
But the fact of the matter is that sometimes simplicity is really valuable. Einstein was
really, really, really good at taking an extraordinarily complex situation and understanding it so that the important pieces
were reduced to a simplicity.
It's the mark of a good mind to acknowledge the power of simple cause and effect while
also acknowledging that complexity is real.
It makes me think, did you ever hear the story of NASA trying to come up with a pen that would write in space?
No, I don't think so.
I am imagining myself in a gift shop in some science museum buying a pen that writes upside down or something.
I've bought that pen too.
Right next to the astronaut ice cream.
Oh, yeah, that's the worst food ever, but it's really fun to eat.
Horrible.
So here's the way the story goes.
And I will emphasize the way the story goes, because I don't want you to take it as fact.
Got it.
So as the story goes, as NASA was booting up and starting to do more and more ambitious
things, and obviously we were competing with the USSR. There was a space race. There was the realization
that a pen that works well here wouldn't work in zero gravity.
Because you need gravity to pull the ink down.
It's interesting. That's one of the exercises used in the illusion of explanatory depth research by
Steve Sloman, right? Which is explain how a pen works. And
most people would tell you, oh yeah, I know how a pen works. I'm just saying right here,
I have no idea how a pen actually works. You're owning it. I am owning it. So NASA
was theoretically, according to the story, spending, let's call it billions of dollars
along with all the other rocket stuff and spacesuit stuff and training
stuff. They were also having to come up with a pen that would work in space. And the way the
story goes is, well, what about the Soviets? What did they do? Did they have billions of dollars
to spend to develop a gravity proof pen? And the answer was no, they used a pencil.
This is a story that was told for years and years, which turns out to be, I'm pretty sure, almost entirely false. Really? It's such a good
story, though. I think the true story was that they had these kind of fancy mechanical pencils
that cost much more than a regular pencil. And when the budget item was noticed in the NASA budget,
someone flipped out and said, whoa, that's a lot to
spend for mechanical pencils.
And so that gave rise to this story about creating an even more expensive pen.
But the moral of the story is the Soviets were smarter in this case.
They came up with the simple solution, a pencil much simpler.
But to me, there's a second layer of lesson here, which is that a pencil seems to be of
simple technology,
but in fact, there's a very famous essay called I Pencil. It's a first-person essay written
theoretically by a pencil, wasn't actually written by a pencil, explaining that no one person knows
how to make a pencil, that it's actually an incredibly complicated process that's been perfected by a lot of different people in a lot of different fields and industries over many, many, many
years.
You have to figure out how to get the lead inside, how that's mined, how it's crafted,
all these different supply chains and things like that.
So even the simple thing turns out to not be very simple.
And so simplicity is awesome.
And when it's very simple for me to use a good or a service or an interaction, I love
it.
But it's a mistake to think that it didn't take many, many, many, many hours and many,
many, many dollars and a whole lot of brain cells for a lot of people to actually come
up with something so simple
and that, in fact, complexity is real and that we shouldn't poo-poo it.
So that's my takeaway from the story of the pencil.
That is a great takeaway.
I think it is a good thing to think about when we try to say,
you know, I know what's going to happen to this person.
Like, pencils are complicated and so is human life.
And if that is true, there are many, many reasons why we're going to do something like take the Diet Coke out of the fridge or stand up or raise our hand.
You really want that Diet Coke, don't you?
I already took it out of the fridge.
Now you have to guess the 95 reasons why I did that.
Still to come on No Stupid Questions, Stephen and Angela explore the
human desire for magic bullet treatments and reflect on how it played out during the pandemic.
These recommendations were so complicated and I kept thinking like, well, can't you make it simpler?
Just give me the pill.
now back to steven and angela's conversation about the human tendency to oversimplify so here's my question for you i understand that it's convenient to come up with a single
cause explanation i understand that it's easy. I understand that our brains
aren't really great at memory. Generally, most of us, like if I give you a list of 10 items
to remember till the next day, if you have two or three still, that's probably pretty good.
Or even the next minute. I'm pretty sure I would not be able to remember 10 things.
That's true. So I get that there are cognitive limitations.
I see why it's attractive, but it's often just a tip of the iceberg.
So why do you think that we so want
to believe the simplest versions?
You really do have to ask yourself why,
because evolution should work
against systematic mistake-making.
Why have we evolved to look for single causes
as opposed to multiple causes
if the world really does work
in a multiple cause way. And I am thinking about Get Out Your Shot class, Danny Kahneman.
His research on system one thinking versus system two thinking, these very quick heuristics versus
more deliberate weighing all the evidence. One of the things I have learned from Danny is that the human mind likes to substitute for a complex problem a simpler one. And it's true that two
competing explanations that are equally descriptive of what's going on, all things being equal,
you should probably choose the simpler one. It's certainly going to be more actionable. It's
taking a lot less of your cognitive capacity to understand. But I think the thing that we're saying is that things are not
equal. The simple explanation is not always as true as the complex one. And so the reason why
the human mind gravitates towards simple explanations is that when we have that simple
explanation, we're not feeling at the
moment that we're incomplete. We're feeling like we chose the one that is more tractable.
Also, just the fact that we can more fully understand the simple explanation draws us to it.
Like grit, it's the secret to success. Oh, I got it. I picked that as opposed to what I would say,
which is like grit is one
thing among many. And then, of course, there is such a thing as the rate at which you learn.
And by the way, they're very complex situational factors that interact in nonlinear ways with all
of the above. Nobody wants to hear that. But then I guess in your case and the story of grit,
you become known as the grit lady. And so all that anybody wants to ask you about.
I am still known as the grit lady.
This happened to me yesterday.
I was asking somebody about something entirely different and they kept thinking that I was
asking them about grit.
It was a little bit frustrating.
That's the point I was about to make, which is once an idea is known, that's the focus
of the follow up inquiries.
And it sort of encourages you to give a response
that's a little bit un-self-interested,
which is to say, yeah, I wrote a book called Grit.
I've studied Grit, but let's talk about
all the other things that actually go into this.
You could argue that your incentives to do that
are not very strong.
I mean, I really don't think that Grit
is the most important thing.
Like, you know, in Disney movies, when there are fairies that grant wishes,
if you only had one, would you make your child gritty?
Like, oh my gosh, wait a second before you make that choice.
What about kindness and integrity?
What about happiness?
Height.
Height, obviously. The incentive for me is
enormous to introduce to the conversation the many things that any parent would want their
kids to grow up to be. But I think the force that pushes against the complexifying of anything is
this desire for single causes. Like, what was the reason why
Theranos collapsed? What is the secret to success? What is the one food that I need to eat to stay
healthy? So I'm not the only one. That means that anybody who's trying to talk about things in their
full complexity is pushing against a very strong force for simplicity.
And I think people who work in all different realms have come up against this. Think about
cancer research. Over the past 40 or 50 years, we've heard about all these different sorts of
magic bullet treatments, which have turned out to be somewhat successful on some cancers,
but mostly not successful on many. You know, the phrase magic bullet goes back to Paul Ehrlich from, gosh, probably over
100 years ago.
He was talking at the time about a drug that theoretically could solve the problem in the
human body without doing anything bad to the human body.
The magic bullet would like hit its target, but not hit anything else.
Exactly.
I mean, we're still wrestling with this in medicine in many areas, not just oncology,
but think about antibiotics.
You want antibiotics that target the bacteria that are doing the bad stuff, but don't do
anything to anything else in the body.
And I mean, if you want to talk about a complex dynamic system, it's the human body.
In a way, the human body can make public society seem kind of simple.
Look at all those different interacting systems.
So I think that the more sophisticated a medical researcher gets or researchers in many, many,
many realms, the more you realize, yes, simple ideas can have massive leverage, but you have to appreciate them in the context of the complexity.
I do wonder, when it comes to psychology, does this notion, the single cause bias,
as you've called it, does this intersect at all with what you psychologists call the
fundamental attribution error? The fundamental attribution error is one of the big ideas to come out of social psychology. These are the psychologists who and their influence on their own behavior and
underweighting, ignoring even the situational pressures. So, for example, a kid comes into
class, doesn't have their homework. The fundamental attribution error is the teacher thinking like
little Steven without his homework doesn't care. Lazy, stupid kid. And who knows?
Maybe the dog literally did eat the kid's homework.
Or maybe my mom is undergoing cancer treatments.
You can think of many situational factors. was shown to exist is that you would tell people very explicitly that the person who did this thing
was under situational pressure to do it. Like the reason why they wrote this pro-Castro essay is
that they were told to write the pro-Castro essay. And then you as a participant, you're asked,
how much do you think this person really is pro-Castro? And people would still attribute some motive to that person's own beliefs,
even knowing that they were just compelled by some other person to write it.
So that's what the fundamental attribution error refers to.
It's been around for so long that it's part of intro psych for students everywhere.
The more recent argument that I have been reading is like, look, in real life, almost everything we do quite obviously is a product of the situation and the person,
when we are in a lab study, that just kind of comes out. We're carrying over common sense.
In general, anybody who would write an essay is probably doing it in part because of what
they have to do, given the situation, and also what they really feel.
You know, there's one more example I'm thinking as you're speaking about the single-cause
bias or the magic bullet idea and a way in which it's really damaging.
If you look at the way the world, but especially Americans, responded to the COVID-19 pandemic,
every single element of our response became a binary choice about this is the thing or
this is not the thing.
Masks will save us all.
Masks do nothing.
I think it reached its apotheosis with the vaccine.
In the beginning, there was this conversation about, wow, here's a terrible virus.
It's killing people.
We don't really know how to treat it.
We don't know the long-term effects.
So the only really responsible thing to do it. We don't know the long-term effects. So the only really
responsible thing to do is to come up with a vaccine. Let's ask some of the best scientists
in the world to do it. And then they did it. They came up with a vaccine that works,
and that's somewhere between, let's call it 60 and 93% effective. Well, that's not the magic
bullet. You know what? There are hundreds of thousands, maybe millions of people alive because of that.
And yet we still have a really hard time putting our minds around the fact that, oh, someone
who took the vaccine still got really sick.
Or, you know, there might be some side effects of this vaccine.
Or you might need to do the vaccine and certain health precautions.
Or you might need to do the vaccine and certain health precautions. I remember during the very opening months of the pandemic, these recommendations were so complicated. And I kept thinking like, well, can you make it simpler?
Just give me the pill. bias. And I'll say this, Stephen, it's not just the pandemic and it's not just trying to forecast
what's going to happen with the person that you just hired or your own life or your kids. I think,
for example, when you say, hey, how are we going to fix the achievement gap that opened up like
the Grand Canyon during the pandemic for rich and poor kids? If we could avoid single cause bias,
we will be on much firmer ground. We both read a paper recently about how there was a study of cognitive therapy, and it was being compared. And what I remember taking from that paper was that they were smart to
actually not only ask the question, which is better, but they actually had a group that got
both. So they could look to see the effect of having a combination of therapy and a small cash
grant. And I believe that was the group that actually in the long run did the best. And so it's not either or, you know,
my favorite go-to is Both And. And Both And is not single cause, right? Both And says there's
a lot going on here. The paradox of this conversation is that you and I, but also I
think everyone who would listen to a show like this for the most part is already on the side
of Both And. So in a way, we're
preaching to the choir, but that's okay because, you know, a choir can then go out and preach to
people who aren't in the choir. Yeah, we can go sing. I guess my takeaway is that nothing
is 100% all the time for everyone except maybe death and the passage of time in this podcast.
There are no stupid questions, and there really are no simple answers.
No Stupid Questions is produced by me, Rebecca Lee Douglas.
And now, here's a fact check of today's conversation.
In the first half of the show, Angela says that Ted changed the title of her talk
from The Key to
Success to Passion and Perseverance for Long-Term Goals. She was slightly off here. The Ted talk was
actually renamed Grit, The Power of Passion and Perseverance, the same title as her best-selling
book. Later, Angela wonders about the identity of Occam of Occam's Razor. William of Occam was a 14th century English Franciscan friar.
His razor, or heuristic, comes from his 1323 publication,
Summa Logicae, where he wrote,
It is futile to do with more what one can do with fewer.
Finally, Stephen shares what he is pretty sure is a mostly fictional story
about how Americans spent billions developing a space pen while Soviets simply relied on pencils.
He was right to doubt the veracity of this narrative. In the 1960s, Paul C. Fisher of the
Fisher Pen Company developed a pen that worked in space, underwater, and in extreme
temperatures. This unique writing utensil was used by both American astronauts and Soviet cosmonauts.
Wooden pencils were considered a fire hazard in spaceships, where the atmosphere was 100% oxygen.
And as Stephen mentioned, NASA was briefly embroiled in a controversy about how much they were spending on mechanical pencils.
That may have been where the false narrative originated.
That's it for the Fact Check.
Coming up next week on No Stupid Questions,
Stephen and Angela answer a listener's question
about teaching children financial responsibility.
I do think that kids working is a good thing.
I made slash asked my two daughters to work as soon as it was legal in the state of Pennsylvania.
That's next week on No Stupid Questions.
For that episode, we want to hear about your very first job.
Were you a cashier? A babysitter?
Did you run a killer lemonade stand?
What did you take away from that experience?
To share your thoughts, send a voice memo to nsq at Freakonomics.com
with the subject line, First Job.
Make sure to record in a quiet indoor space with your mouth close to the phone.
And please keep your thoughts to under a minute.
No Stupid Questions is part of the Freakonomics Radio Network,
which also includes Freakonomics Radio,
People I Mostly Admire,
Freakonomics MD,
and Off Leash.
All our shows are produced by Stitcher and Renbud Radio.
This show was mixed by Eleanor Osborne.
We had help on this episode from Lyric
Bowditch and Jacob Clemente. Our staff also includes Neil Carruth, Gabriel Roth, Greg Rippin,
Morgan Levy, Zach Lipinski, Julie Canfor, Ryan Kelly, Jasmine Klinger, Emma Terrell,
and Alina Coleman. Our theme song is And She Was by Talking Heads. Special thanks to David Byrne
and Warner Chapelappell Music.
If you'd like to listen to the show ad-free,
subscribe to Stitcher Premium.
You can follow us on Twitter
at NSQ underscore show
and on Facebook at NSQ show.
If you have a question
for a future episode,
please email it to NSQ
at Freakonomics.com.
To learn more
or to read episode transcripts, visit Freakonomics.com. To learn more or to read episode transcripts,
visit Freakonomics.com slash NSQ. Thanks for listening.
Let me just keep it simple for you, Stephen.
Hey, can I just say I'm not that much of a simpleton.
the Freakonomics Radio Network the hidden side of everything Stitcher