Podchaser Logo
Home
When Uncertainty Hides in the Blindspot of Overconfidence

When Uncertainty Hides in the Blindspot of Overconfidence

Released Wednesday, 17th April 2024
Good episode? Give it some love!
When Uncertainty Hides in the Blindspot of Overconfidence

When Uncertainty Hides in the Blindspot of Overconfidence

When Uncertainty Hides in the Blindspot of Overconfidence

When Uncertainty Hides in the Blindspot of Overconfidence

Wednesday, 17th April 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:02

You should not, right? You are

0:05

no scientist. You are a foolish

0:07

person. Your

0:11

article could cause new cancer deaths.

0:16

Get. Beyond your pet that a

0:18

left wing ain't east over

0:20

the empire fascist lies. What

0:22

are your qualifications? I. Looked

0:24

up your by own was totally unimpressed.

0:29

You call yourself a science writer.

0:31

Your article was all lies. I

0:35

have no clue where you got your info

0:37

on vitamins. There is no doubt in my

0:39

mind they helped me in a multitude of

0:42

ways, bashing them for no good reason, Christie

0:45

ash want and should not be

0:47

allowed to write for any magazine.

0:50

Maybe she has no friends? Maybe that's

0:53

the real problem here. I'm.

0:59

Christy Ashooh undone and this

1:01

is uncertain from Scientific American.

1:10

Today. We've gotta show and

1:12

to axe exploring how studies can

1:14

make us so over confident that

1:16

we lose sight of the underlying

1:18

uncertainty. Act

1:20

One That backfire effect. Back.

1:24

And twenty eleven. I gave a short

1:26

talk at a journalism conference about what

1:28

ahead learn from hate mail. The bottom

1:30

line I said was that if you

1:32

tell people that they're wrong about something

1:34

they hold dear, they will immediately get

1:36

defensive. And if you're a journalist like

1:38

me, they will send you nasty letters

1:40

like the ones you heard at the

1:42

top of the show. I

1:45

based my argument or my

1:47

own personal experience, having written

1:49

about things like climate change,

1:51

cancer screening, and most controversially

1:53

the Useless. Of multi vitamins

1:55

but I also backed it up

1:57

with science specifically some research. by

2:00

Brendan Nihan, a political scientist

2:02

at Dartmouth College. Here's

2:05

how I explained it in the talk. Nihan

2:08

studied what happens when you tell people that

2:10

they're wrong, and what happens is

2:12

you go back and think about all the reasons

2:14

you were right in the first place, and

2:17

in the process of doing this, you become

2:19

more sure of yourself. This

2:21

became known as the backfire effect. Here's

2:26

Brendan Nihan. After

2:28

I graduated from college, I had worked as

2:31

one of the founders of a non-partisan fact-checking

2:34

website called Spinsanity that was one of

2:36

the original online fact checkers. We

2:38

ran from 2001 to 2004, and that ended up putting us in

2:44

the middle of the war in Iraq and how

2:46

that was presented, the aftermath of 9-11

2:48

and the conspiracy theories around that. Oh

2:51

yes, the 9-11 conspiracies, there are a

2:53

lot of those. Then you fast

2:55

forward to the end of the Bush administration Barack

2:58

Obama is running for president, the

3:00

Obama-Muslim myth starts spreading. That

3:03

becomes a very salient misperception.

3:06

The first two years of the Obama

3:08

administration, we have the Affordable Care Act,

3:10

so the death panel myth is a

3:12

hugely influential policy misperception, and

3:15

the Obama-Muslim myth morphs into the birther

3:17

myth. So people like you

3:19

are trying to correct these misperceptions and not

3:21

getting much traction. People are trying

3:23

to understand why are these so

3:25

prominent? Why are our efforts to fight these, not

3:28

working? Your studies used undergraduates

3:30

as test subjects. What was

3:32

the setup? Each of the

3:35

studies involved a mock news

3:37

article about a salient political

3:39

misperception. The experiment was whether that

3:41

mock news article randomly included corrective

3:43

information debunking the claim made

3:45

by the featured political elite

3:47

or not. What kinds of

3:50

misperceptions were you testing? So

3:52

for instance, there's an example in

3:54

the article is George W. Bush

3:56

suggesting that his tax cut will

3:58

reduce the deficit. rather than increase

4:00

it. This past week I have been making the

4:03

case for tax reductions. I've

4:05

asked Congress to act quickly on my tax relief

4:07

plan so that Americans can

4:09

face these uncertain economic times with more

4:11

of their own money. Which is

4:13

a finding that even his own

4:16

economists held that his claims were

4:18

not accurate. Ah, so one group

4:20

would be told that, well, actually the Bush

4:22

administration's own economists say the tax cut will

4:24

increase the deficit by many millions of dollars.

4:27

That was an experimental variation. Are

4:30

you being exposed to that corrective

4:32

information or not? And you found

4:34

that people exposed to the corrective information, what

4:37

we might now call a fact check,

4:39

actually ended up becoming more sure of

4:41

the misperceptions rather than correcting it. The

4:44

finding was that two of the five

4:46

studies we conducted showed evidence of this

4:48

backfire effect. Our interpretation when asked was

4:50

that people might be thinking about reasons

4:53

they held a belief in the first place and

4:55

coming to double down on them as a result

4:57

of being corrected. I loved

4:59

the backfire effect. It felt so

5:01

true and resonated so much with

5:03

my personal experience. I was

5:06

positive that this finding explained everything in the

5:08

same way that I was sure that

5:10

the dress from our previous episode was white

5:12

and gold. That finding turned out to

5:14

be a real brain worm for all

5:17

of us. It

5:23

resonated with people's experiences like

5:25

yours in reminding

5:27

people of the times when they struggled

5:29

to convince someone who seemed even more

5:32

sure of their original belief. As

5:34

a result, it just resonated. And

5:37

as we told and retold the story in the

5:39

way humans do, the scientific

5:41

nuance was lost. It felt so

5:44

true. Like it didn't need nuance. I

5:46

think you're an extremely careful scientific journalist. It's not

5:48

a criticism of you in any way. Well, thank

5:50

you. That is very kind. But it's like

5:53

as I tell the story and then you tell the

5:55

story and as it Propagates out there

5:57

into the world.. And Then of course,

5:59

we told the story. we lost control

6:01

of how it was described overtime and

6:03

it's just. Spread. Everywhere and

6:06

you know good at the time the most other

6:08

wasn't better. at the time there really wasn't better

6:10

evidence or this was the study I conduct an

6:12

airport or what I found and I try my

6:14

best interpretation of it. From. A scientific

6:17

perspective. Let's take a moment to step back

6:19

and take a closer look at what's happening here.

6:22

Friend. And the group got some really interesting

6:24

data. But. Those are just raw

6:26

results. The. Human brain doesn't think and

6:28

data. We think and. Stories.

6:31

So. He and his colleagues do at

6:33

all scientists do. They created a story

6:35

to explain their data. To

6:37

do that, they had to do some interpretation

6:39

and bring human judgment to bear. Let's

6:42

listen again to at Brendan said a moment

6:44

ago. Our interpretation when ass

6:46

was that. People. Might be thinking

6:49

about reasons they held belief in the first

6:51

place in coming to double down on them

6:53

as a result of being corrected. Noticed. That

6:55

she said people might be thinking about

6:57

reasons they held the police. Because.

7:00

He's a careful scientists bread and miss

7:02

classes here to say that this is

7:04

a possible explanation. But. It's really

7:06

just a guess and is pretty careful not

7:08

to overstate it. Whether or

7:10

not the result south holds up, this

7:13

explanation of it is a hypothesis. We.

7:15

Can think of it as a just so story. Richard

7:19

Kipling classic collection of just so

7:21

stories for little children provides fanciful

7:23

explanations for things like how leopards

7:26

at their spots and how camels

7:28

got their humps. Just. So

7:30

stories or tails invented to explain

7:32

observed phenomena just like the explanation

7:35

scientists present for their scientific results

7:37

are invented to explain the data,

7:40

And. So they're tested. These stories are

7:42

just best. guess is that because he's

7:44

explanations are created a perfectly fit the

7:46

data, it's easy to jump on the

7:49

story and forget all about the uncertainty

7:51

that surrounds it. As. It

7:53

turns out the back fire fact is a

7:55

classic case of a very exciting first finding

7:58

that we are some fun and then. Prove.

8:01

It disappears and later studies. Of

8:03

course, it's important to note that only

8:05

two of the five original experiments found

8:07

any evidence of the backfire effect, so

8:09

there was a lot of uncertainty about

8:11

it from the very first study. Since.

8:14

Then. Many other studies

8:17

have been done by myself and my

8:19

coauthors by many other scholars, Finding the

8:21

backfire effect seem to be exceptionally

8:23

rare. That I know the genie is

8:25

out of the bottle. This followed that is never

8:27

good as much media attention is the first time,

8:29

right? Those and forces don't get the

8:32

so seem kind of. Play.

8:35

I feel like this is a little bit not all

8:37

right. You're presented with evidence

8:39

that the original finding on the back

8:41

fire fact might not hold up, and

8:43

instead of doubling down, you actually opened

8:45

up and said wait a second. Let's

8:48

revisit this. Let's put all the evidence

8:50

together. You are and are you know?

8:52

I'm I'm grateful to all the. People.

8:54

Who interacted with me in a good faith

8:57

manner and I think that helps us all

8:59

be better people in in and having these

9:01

kinds of exchanges right? It never became personally,

9:03

never became. Vitriolic. And

9:06

that helped citing keep. The

9:08

focus on the science in a way that

9:10

help me of we he responded i'm in

9:12

a more in an appropriate way. It's.

9:14

Kind of the idealized version of has

9:17

Sciences Sit behaves. I can easily imagine

9:19

a different universe where you double down

9:21

and you may be give a Ted

9:23

talk and right a pop science book

9:25

about the power of the backfire fact.

9:28

Instead, you published a paper titled why

9:30

that backfire effect does not explain the

9:32

durability of political misperceptions. And it's

9:34

a great reminder that we don't have all

9:36

the answers. I think it's really change how

9:39

I think about science and how I do

9:41

science. Has it changed? the way you operate?

9:44

I hope it's a help me a process

9:46

question. With. More humility

9:48

because. I had many preregistered

9:50

studies at that point. Oh right

9:53

Femur Dysart studies are experiments where you

9:55

lay out and commit to specific plan

9:57

you can. it's how you a period.

10:00

The study and analyze the results before

10:02

you begin. This. Is meant

10:04

to reduce the temptation to fiddle around with

10:06

the analysis once the results start coming in.

10:09

So. What a few studies found. The.

10:11

Findings were clear that. Correct.

10:13

Information do tend to reduce

10:15

proceed accuracy of false claims

10:18

Someone. right? And I had

10:20

plenty of preregistered studies in which I had found

10:22

exactly that. So. Does the

10:24

backfire effect really exist? We.

10:26

Know I wasn't. Finding. Evidence

10:28

of. Pervasive. Are substantial

10:31

back firefox either. Since

10:33

that talk thirteen years ago at that. a lot

10:35

about the back side effect and what it says

10:37

about science and how we know what's trill. One.

10:40

Thing that's become clear to me is that it's really

10:43

easy to get a result. He. Did a

10:45

study and he found the saying. It's

10:47

a lot harder to get an answer to

10:49

the underlying question. When I

10:51

was getting that talk about the backfire fact,

10:53

I was trying to explain why people resist

10:55

changing their beliefs when they. Are offered a

10:58

correction. I really wanted a

11:00

universal answer. But. That's

11:02

asking for a lot. For. We expecting too

11:04

much for the backfire fact. Became.

11:07

To stand in for why the

11:09

social phenomena exists, rights and no

11:11

scientific research can bear the weight

11:14

of these. There's a demand for

11:16

model causal explanations for complex social

11:18

problems, right? And this one certainly

11:21

couldn't. We. Want a single

11:23

answer that explains everything? But maybe

11:25

there isn't one. Alex itself

11:27

that the most scientific answer to

11:29

any question is it depends. I.

11:32

Think as a learning opportunity for everyone's really no

11:34

warning A pretty for me. As a scientist, I've

11:36

done my very best in my own research to

11:38

follow up on the study to not let it

11:40

be the last word to. Publicize.

11:43

The emerging body of evidence that better

11:45

represents the scientific consensus as as time

11:47

has gone on and I think. Maybe.

11:50

Most importantly, I've tried to not be

11:52

defensive about my own study, because out.

11:55

In. A What Kind of a? Reservist.

11:57

Information would I be if I wasn't willing to change my mind

11:59

when the. Change. That's really noble

12:01

of you, but we're all human. It's easier

12:03

to sell the sense of how do you

12:05

live away from the desire to be right,

12:07

toward the goal of finding the right answer?

12:10

I've. Gone on to work on a number

12:12

of studies with or eat Importer and Tom

12:14

would have done some of the most important

12:17

follow up research. In. This area and

12:19

we work together collaboratively instead of having

12:21

one of these adversarial. Scientific.

12:23

Disputes work he side says the other

12:25

is wrong we've for to pass a

12:28

try to move that research literature in

12:30

it. More. Informative direction and

12:32

I'm I'm really excited about. What?

12:34

Lessons can. We take the Backfire saga.

12:37

The. Single splashy study ones itself

12:39

to media coverage. It's

12:41

what the big general interest science journals look

12:44

for. By. Sometimes.

12:46

Studies are surprising precisely because

12:49

they don't correspond said prior

12:51

knowledge. City. Be more skeptical

12:53

of really surprising findings. We should

12:55

be cautious that they may be anomalous rate

12:57

and our our study turned out to be

12:59

anomalous. I don't know of an air we

13:01

made, I of course it's science. It's not

13:04

our fight or findings. A provisional and subject

13:06

to being overturned. This Harkins to

13:08

the hidden moderators we talked about in

13:10

our previous episode, does unremarkable decisions that

13:12

researchers make that may end up having

13:15

a big influence on the results they

13:17

get. Has. This experience change

13:19

the way you approach study design. There

13:22

are a number of factors that I would you very

13:24

differently now. With. You know the knowledge

13:27

of. Be a more

13:29

than fifteen years and switch back to

13:31

those studies and I guess what's important

13:33

to take away his. It. Doesn't have

13:35

to be. Wrong in the

13:37

sense of it was it was constructed

13:39

fraudulently or correctly. He can just be.

13:42

One data point among many. So

13:44

lesson here is to be careful about putting

13:46

too much weight on a single study. It's.

13:48

Easy to focus on the exciting or

13:50

surprising finding and forget all about the

13:52

other ones that may be less interesting.

13:55

Here. a lot of people including me

13:57

overlook the fact that from the very At

14:00

the start, the Backfire Effect was an outlier.

14:02

It only showed up in two of the

14:04

five experiments. Brendan, how has

14:06

this experience influenced the way you evaluate

14:08

studies? We

14:10

interrogate the science that we discuss

14:12

in my classes. I encourage the

14:15

students to take apart

14:17

the methodology of the research,

14:19

the assumptions that the authors

14:21

make, the flaws in their

14:23

experimental designs. I try

14:26

to foster the kind

14:28

of critical scientific inquiry that

14:31

I try to practice as a scientist in my

14:33

teaching in the classroom. You're a

14:35

political scientist. How does an understanding of

14:37

uncertainty factor into our political system?

14:39

We are coming into contact with

14:42

lots of studies that are being offered

14:44

as evidence for various kinds of policies and

14:46

interventions. And I think learning

14:49

about the uncertainty in science

14:51

and the value of

14:54

high quality science as well as the uncertainty inherent in

14:56

the process. People hold those both in our

14:58

heads at the same time is a really

15:01

important part of

15:03

hopefully being a leader in

15:05

this country in the 21st century. So

15:07

putting this all together, what does this tell

15:09

us about science? It's an

15:11

imperfect human process with many flaws,

15:15

but it's an incredibly powerful

15:17

one for learning about the world. It's

15:19

the best way we know how to learn

15:21

about the world. And I

15:24

think we can keep what's

15:27

great about science while

15:29

also remembering its highly

15:31

provisional nature. And if

15:34

we can strike that balance, it

15:37

can do great things. Oh, I love your

15:39

optimism. Today's

15:45

episode of Uncertain is about the ways

15:47

that studies can leave us overconfident. We

15:50

just heard about how Just So Stories can

15:52

make us feel overly certain about results that

15:54

are still a work in progress. And

15:57

now we've arrived at Act Two.

16:06

Sometimes studies get misleading results because

16:09

of random error, a weird samples

16:11

or study design, But sometimes science

16:13

gets things wrong because it's done

16:15

by humans. and humans are fallible

16:18

and imperfect. No matter

16:20

how much we like to think

16:22

that science is impartial, the truth

16:24

is human preferences and values influence

16:26

how science his son and that

16:28

means cultural factors and embedded prejudices

16:31

can sway the results that studies

16:33

produce. Someone who knows a

16:35

lot about this is Brandon of Bunim. Yeah.

16:38

I know this episode had a Brendan

16:40

and our first act and now we've

16:43

got a Brandon. Don't let it confusion

16:45

Brendan as the political Scientists Brandon is

16:47

a biologist. Here's

16:50

brand. I'm I'm an assistant professor that

16:53

Obama because you the Evolutionary Biology at

16:55

Yale University, in an external professor at

16:57

the Santa Fe Institute. Branded.

16:59

As given a lot of thought to scientific

17:01

authority, he grew up with a lot of

17:03

respect for scientists. I was raised

17:06

with this appreciation for scientific quote,

17:08

unquote ideas and certainly people who

17:10

did it. and it's A A

17:12

D. That is because of the

17:14

great thing said. Sides accomplish home,

17:16

you know, going to the moon

17:18

and. Drugs. For diseases

17:21

in the engineering breakthroughs. Ah, the

17:23

last century in so you know,

17:26

As a product of the things that sizes

17:28

created you know I had this view of

17:30

in this very positive view of sight. Out

17:32

of his plans, your views on uncertainty. Was

17:34

there a moment in your career or your

17:36

education where you fully came to grips with

17:39

the importance of uncertainties. A science. While.

17:41

I'm a i'm sitting a little bit here

17:43

but I think you know it. It looked

17:45

like every time I saw in the media

17:47

has Obama mother be as just Cc. the

17:50

sees the center of gravity fall of me.

17:52

All. Of these ideas white

17:54

certainty or merits. Or.

17:57

All these institutions. Send. Me:

18:00

The. Never. Made sense. right?

18:02

It's because. I. Was

18:05

raised by an African American woman

18:07

who's smarter they any body in

18:09

my department now. She.

18:11

Just didn't have the opportunity. To.

18:14

The notion that life gives people a chance based

18:16

on merit and you just work hard, work hard

18:18

and you'll get with were that's never been through.

18:20

Ever. Good. I saw somebody

18:22

brilliant and hard working. Did

18:25

nothing but barriers for that. The

18:27

her life she knew she was some up

18:30

more than most people and they get. Challenged.

18:33

Sounds sounds sad, sad it's I've

18:35

seen her be right a hundred

18:37

times and still why school administrators

18:39

or whoever the landlords of level

18:41

we were dealing with the top

18:43

be wrong and they when. Pressing.

18:47

Had it that influence your views of

18:49

scientific authority. Soviet That notion that

18:51

like that that the world's most of

18:53

work a certain way was never right

18:55

with me I think for me and

18:58

that sides the African American experience. And.

19:00

Marriage rates. At all

19:02

of these voices of authority. Told

19:04

us where we were supposed to be in

19:06

society and what we were supposed to be

19:08

doing and how we couldn't go to school.

19:11

It's so that's never worked. As I said

19:13

it's except that that is that The Gospel

19:15

of Certainties is another kind of authority. right?

19:18

And I think so I'd My point is,

19:20

I've never really entered sides that way. So.

19:22

How do you think a science now? I.

19:24

Think as I've made sure that

19:26

book to as a person N

19:29

B has had become meet you

19:31

know a member of the sides

19:33

workforce. And. The both my

19:36

appreciation of for science has

19:38

grown a and become much

19:40

more nuanced. And and I

19:42

mean my love of It's In the

19:44

Bay. Positive sense. And.

19:47

My understanding of it's

19:49

flaws in it's major

19:51

problems and pitfalls and

19:53

challenges. So he of us

19:56

science despite it's shortcomings. I

19:59

appreciate. many more ways the power

20:01

of it and what makes it

20:03

beautiful and I

20:06

understand the challenges and the damaging

20:08

aspects of you know the damage

20:10

that it can cause. What

20:12

do you mean when you say the damage it can cause?

20:15

On one end we mean that

20:17

the products of science can be

20:19

damaging right so you know we

20:21

talk about the bomb, we're talking

20:24

about biological weapons. Right

20:26

the products of science can be very

20:28

harmful and in these two cases they

20:30

were actually created specifically to cause damage.

20:33

The much more pernicious but

20:37

simultaneously more violent

20:40

way is the ideas. So

20:42

you mean some of the actual ideas and theories that

20:44

science puts out? I'm thinking here

20:46

of eugenics. You can't look

20:49

at like the eugenics movement and say oh

20:51

well that was just a fringe thing. Like

20:53

no it wasn't. That

20:55

was the field. That

20:58

was the way the mainstream thoughts.

21:01

For instance, Thomas Jefferson was the

21:04

preeminent polymath of his

21:07

time. He was a technical

21:10

and evidence-based and profoundly

21:14

scientific thinker for his time.

21:17

What is the lesson we should take from this? When

21:20

the process for all of

21:22

the reasons doesn't have

21:24

pay attention to the

21:26

way social influences or

21:29

biases or power

21:31

are influencing the types of things

21:33

we're asking and how we're asking them. Science

21:38

can get things shockingly

21:40

wrong. It can lead

21:42

us down really really bad incorrect

21:45

paths that end up causing harm.

21:48

How does science get these bad ideas to begin

21:50

with? It can be because other

21:53

interests, biases that you

21:55

may not see your own and

21:58

societies infiltrate process

22:00

and steer the

22:03

process in order

22:05

to try to get to a

22:07

product that we oftentimes find out

22:10

was based on a broken idea. We

22:13

think that science is objective, but you're

22:16

arguing that, in fact, we're bringing our

22:18

human biases to bear here. Even

22:21

the nature of the questions we're asking is not

22:23

objective, right? The term bias

22:26

doesn't even necessarily mean bad

22:28

or harmful. It

22:32

means that you are a person, sentient

22:36

individual who has experienced things

22:38

on the earth. Invariably,

22:40

there is no way that

22:43

is not going to influence the

22:45

manner that you do your work. From

22:48

me writing my manuscripts in English.

22:53

Let's just start there. Not

22:55

only does that influence to

22:57

whom I'm speaking, it

22:59

influences the shape of the manner that

23:01

I describe certain ideas, because

23:04

languages have different means of describing

23:06

things. So you're saying that scientists

23:08

need to look for and acknowledge the limitations of

23:10

their work? I report

23:13

the limits of my work for me.

23:16

I think that is the responsible way to

23:18

do science. But that stands out to

23:20

people, and they're like, wait a minute, what

23:22

are you doing? Why would you be pointing out the flaws in

23:25

your own work, Brandon? Is that

23:27

imposter syndrome? No, it's not imposter syndrome.

23:29

It's me being a responsible scientist. I've

23:32

heard a lot of talk within the scientific

23:34

community in recent years about making it easier

23:36

for scientists to correct course or

23:38

issue retractions without undue penalty.

23:41

Should we think of being wrong as

23:43

the occasional price of being ambitious and

23:45

creative? Science is

23:47

not about the proportion of times

23:49

that you're right. It's what you are

23:51

right about. In fact, many

23:55

of the people who are right

23:57

about important things were all about

24:00

many many many many many many many many many many

24:02

many many many many many many many many many other

24:04

things. Okay, so you're saying there's

24:06

a lot of cases where scientists have been

24:08

wrong and we tend to focus on the

24:10

correct ideas and forget the ones that we

24:12

abandoned. I have this ambition

24:15

to give a seminar where I talk exclusively

24:17

about things I was wrong about. Okay?

24:20

And when I do that,

24:22

frankly, that will be a

24:25

sign of my eminence. And

24:28

you can boldly profess your wrongness without losing

24:30

your credibility. Admitting you're

24:33

wrong in explaining what you

24:35

got wrong is actually a

24:38

demonstration of your power,

24:41

knowledge and authority. And

24:44

demonstrating the wrongness and what you've

24:47

learned is precisely what makes science

24:49

the greatest knowledge creation instrument in

24:51

the history of the universe. How

24:54

do you balance wrongness with authority? Look,

24:57

I'm an arrogant professor like everybody else

24:59

is. I'm not modest. I

25:01

think I'm right about things. I'm not

25:04

saying you should expect to be wrong

25:06

or try to be wrong. But if

25:08

you are, just say so and explain

25:10

to us why you're wrong and what

25:12

we can learn because being wrong is

25:15

a critical part to building the understanding.

25:17

Does the public need a better understanding of the

25:19

provisional nature of science and the way that being

25:22

wrong is a natural part of the process? I

25:24

believe it's the only way science is going to survive as

25:27

an institution is if we start

25:30

to actually lay bare and make transparent

25:32

all of these dimensions of the way

25:34

that it works. Uncertainty

25:36

is not a sign

25:39

of a problem. Uncertainty

25:43

is oftentimes a

25:46

feature of the way we're studying the

25:48

natural world. Uncertainty

25:51

will be a part of virtually

25:53

every natural system in the universe.

25:56

The question is, How

25:58

do we change that? uncertainty?? How

26:00

do we inform the parameters of that

26:02

uncertainty? What's. Your party

26:04

take away. what should the public

26:06

know about uncertainty and science? So.

26:09

The idea is we need to disabuse ourselves,

26:11

have served the even be the goal. The.

26:14

Goal is understanding the world

26:17

and understanding chad exist with

26:19

accumulated uncertainties That sides is

26:21

kind of informing the. And

26:24

to me with our data. Let

26:29

go of the need for certainty that

26:31

seems like that. Life advice. Till. We.

26:45

Spent much of today's episode exploring how

26:47

easy it is. Pressed the fall a

26:49

little to in love with ideas which

26:51

is problematic when they turn out to

26:53

be wrong. And

27:00

the history of science is full of

27:02

ideas that were not quite right or

27:04

even dead wrong. That science has also

27:06

eliminated many useful truth about the universe,

27:09

and there's some things that scientists can

27:11

do to point themselves in that direction.

27:13

The way that we phrase it

27:15

is that the priority for researchers

27:17

should be on getting of bread

27:19

not be reached. The notion that

27:21

I have to be right means

27:23

that I need to take a

27:25

defense attack. When you find a

27:27

flaw, that prioritization of getting it

27:29

right seems that if you say

27:31

I think I see something wrong

27:33

with your research, I say oh

27:35

really, what is it helped me

27:37

cause my goal is to make

27:39

sure that it's right on an

27:41

episode of Uncertain Meters. for hooking.

27:44

Be. In

27:51

his produce a me is the Us food and and

27:53

sat on the. Know how serious heart

27:55

is there any he says he to

27:58

have I. I'm use as them. That

28:00

an. Ending at

28:02

a series was provided a U C.

28:04

Berkeley is greater good science and for

28:06

his part of the expanding awareness. Of

28:08

the science of intellectual A humility initiative.

28:11

Which is supported by the time simple can

28:13

envision. This is uncertain

28:15

of such as from Scientific

28:17

American and hopeful.

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features