Podchaser Logo
Home
How the Right Won the Internet and How the Left is Fighting Back

How the Right Won the Internet and How the Left is Fighting Back

Released Sunday, 9th June 2024
 1 person rated this episode
How the Right Won the Internet and How the Left is Fighting Back

How the Right Won the Internet and How the Left is Fighting Back

How the Right Won the Internet and How the Left is Fighting Back

How the Right Won the Internet and How the Left is Fighting Back

Sunday, 9th June 2024
 1 person rated this episode
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

This episode is brought to

0:02

you by PayCor, the HR

0:04

and payroll software made for

0:07

leaders. It's never been harder

0:09

to recruit, hire, and engage

0:11

workers. That's why HR leaders

0:13

and frontline managers depend on

0:16

PayCor for all things people

0:18

management, from onboarding and performance

0:20

reviews to compensation and benefits.

0:22

Learn more at paycor.com/leaders. That's

0:24

paycor.com/leaders. The

0:27

MAGA world has much stronger ties among

0:29

its supporters and I think it's because

0:31

it really does empower them. What Trump

0:33

did, I think, was he helped cultivate

0:35

a community where he sent a signal

0:37

by his own behavior because at 3

0:39

a.m. he was tweeting about whatever.

0:42

Crazy shit. Crazy shit came to mind or whatever he

0:44

saw that he thought was cool or funny. And

0:46

he sent a signal to

0:49

his supporters that it's

0:51

good. It is useful to my campaign for you to just

0:53

post whatever you want about

0:55

me. Hey everyone, you

0:57

just heard from today's guest, journalist and author,

1:00

Sasha Eisenberg. So

1:02

way back in 2012, when I was working on my last campaign, Sasha

1:07

wrote a book called The Victory Lab about how the

1:09

Obama reelect and other Democratic campaigns had

1:12

used what was then revolutionary technology to

1:15

reach voters and increase turnout. It

1:17

was an optimistic take on the role of tech

1:19

and politics based on the impact of

1:22

the pandemic and the impact of the pandemic on the country. It

1:26

was based on the idea that voter targeting and

1:28

digital fundraising could lower the barriers to

1:31

political participation and organizing. Then

1:33

the 2016 election gave us Trump, fake news, alternative

1:37

facts, and a mess of bad actors who

1:39

exploited technology and online platforms to

1:42

spread disinformation and propaganda that

1:45

made voters more confused and

1:47

more radicalized than ever. It

1:49

came from the Russian government, but also from

1:51

the Trump campaign, right-wing media, and

1:53

from his trolls and kooks who remain difficult

1:56

to track. All of this

1:58

led Sasha to go back and reevaluate the... offline.

4:00

That was a very analog book.

4:02

I hardly mentioned the internet in it.

4:04

And that's because all these innovations were

4:07

tethered to people's real identities at their

4:09

home address and campaigns getting better at

4:11

door knocking and direct mail

4:13

boring stuff. And their

4:16

use of experiments to figure out what worked. And the fact is a

4:18

lot of that was being done

4:20

in this country by people

4:23

who their funding structure and their ambitions

4:25

were tied to getting people more involved

4:27

in the process. So they were these

4:30

innovations were used for campaigns to

4:32

get more efficient at registering voters,

4:34

turning them out. You know, these micro

4:37

targeting, it's not to say that

4:39

every statistical model was used to give people

4:41

good information, but it's good to give people

4:44

more precisely targeted information, which

4:46

is generally in the service of informing

4:48

voters and educating them. And

4:51

I think what we've seen is that we have all sorts of people

4:54

in politics now who are

4:58

willing to do things in the context of a campaign

5:00

that we're outside of what political consultants

5:02

sort of thought was fair game before. And so

5:04

what we see is that all a lot of

5:06

the technical innovations can be turned

5:09

on their head and used for ill. So

5:11

what made you want to write this latest book? So, you

5:13

know, I've been having a conversation with a book editor

5:16

that book came out in the fall of 2012, right

5:18

before the reelection. And a book

5:20

editor had been coming to me for for years saying I'd

5:22

love to do a follow up to the victory lab and

5:24

all the stuff in that area seemed very incremental to me.

5:27

And a few years ago, it really struck me that the most

5:30

interesting sort of frontier

5:34

of innovation in campaigns was thinking about this

5:36

new asymmetry that had been created. We're

5:38

now in an environment where legitimate

5:42

political actors, candidates,

5:44

party committees, super

5:46

PACs, labor unions, people who, you know, have

5:49

to report to the FEC or the IRS

5:51

and be regulated and held

5:53

publicly accountable, are now

5:55

dealing with a situation where their

5:57

opposition isn't necessarily isn't their opponent. It's not.

6:00

another candidate. It's not

6:02

another party committee. It's not somebody

6:04

who's playing by the same rules, might not even be in the

6:06

same country as we are. And the

6:09

whole structure of all

6:12

the logic of political communication, the playbook

6:15

that people like you and your co-hosts

6:17

here grew up with was,

6:19

I think, tethered to a bunch

6:21

of 20th century institutions and assumptions

6:23

about communication. And what

6:26

we sort of call disinformation, which I think is a sort

6:29

of useful catch-all term for a lot of crazy stuff

6:31

that happens online now, I

6:33

think has made clear that all

6:36

of those assumptions are basically outdated.

6:38

And this question of

6:40

how and when you respond to

6:43

an attack, the old playbook

6:45

just doesn't make sense anymore online. And so this

6:47

was a few years ago really struck me that

6:49

coming to discover what the smartest people in

6:52

this area of politics were thinking about

6:55

how you make those determinations and navigate

6:57

the online landscape was

6:59

really interesting to me. I want

7:01

to get into sort of how

7:03

the playbook has changed and how

7:05

political actors are sort of using

7:07

technology, not just for good, but for

7:09

ill now. It also seems to

7:11

me like some of

7:14

this change happened because

7:16

the internet itself changed and

7:18

technology changed. Like Obama

7:21

talked about this last time we interviewed him on

7:23

Pod Save America. And he

7:25

said that the key difference was

7:28

that our campaign saw the internet and

7:30

social media as a tool to organize

7:33

people offline and

7:36

in-person meetups. He was talking about

7:38

that, which also allows

7:40

for like more nuanced, thoughtful

7:43

conversations between supporters of

7:45

a potential candidate who may not

7:47

agree on everything. And now

7:50

so much of political organizing

7:52

and campaign work itself happens

7:54

only online. And

7:56

obviously the internet and social media, especially not

14:00

modern campaigns on the right and saying, is there any place you

14:02

should be talking to? And

14:04

what you recognize is

14:08

that you cannot find people in

14:11

the Republican party, and this bridges the sort

14:13

of MAGA world and the establishment professional world,

14:15

who will even accept the premise of

14:18

that quote unquote disinformation is like a

14:20

meaningful category of speech. And

14:23

basically I think the consensus view, and

14:25

this has now become a core, a

14:30

core precept of the Republican party

14:32

now is that disinformation is a

14:34

concept that academic

14:39

researchers conjured so

14:41

that they could work

14:44

with federal agencies to

14:46

exert pressure on

14:49

social media platforms to censor conservative speech

14:52

in the service of Democratic candidates and

14:54

campaigns. And that's

14:56

what it is, and that is a

14:58

very hard conspiracy theory to break. And

15:01

now we actually see what that looks like when

15:03

it's turned into action, which is the Jim Jordan

15:06

Weaponization Committee in the House,

15:09

Republican attorneys general who are, have

15:12

been suing platforms about this. But this,

15:14

you know, it's really notable, Donald Trump

15:17

announces he's running for president in 2024,

15:19

literally the first policy pronouncement he makes. He puts

15:22

out a video, it's not about trade, it's not

15:24

about immigration, it's not about any of the things

15:26

that we think Donald Trump cares most about. It

15:28

is saying I will get rid of the people

15:30

at FBI who

15:33

are doing counter disinformation work and my

15:35

government will not work with

15:37

social media platforms. This is the first pledge that

15:39

a candidate for president makes. And I think

15:41

it's because it is a sort of central organizing principle

15:43

of the Republican party now. Do you have a

15:45

theory on why there

15:48

is that difference in how each side

15:51

views disinformation? And

15:54

is it based in ideological

15:56

differences? Is it political? Is

15:58

it situational? and that I

16:01

think Republicans have sort of historically,

16:04

especially in the Trump era, sort of been

16:06

on the offense in trying to

16:08

spread or use some of these

16:10

tactics to either suppress voter turnout or

16:12

confuse people, stuff like that, and maybe

16:14

Democrats have not done as much of

16:17

that. Yeah, so I think

16:19

that there's a bit to all this. I

16:21

think it's situational to some extent. I think

16:23

that, you know, in

16:25

the sort of thrashing about for explanations for

16:27

why Hillary Clinton lost in 2016, disinformation

16:31

in some form or another popped

16:33

up a lot. And so I

16:35

think that there's a sincere view

16:37

among non-crazy Republicans that this

16:40

has become a crutch that Democrats

16:42

use to sort of evade

16:45

responsibility for

16:48

elections that they lost. It's sort

16:50

of the new favorite like deus ex

16:52

machina for Democrats who just want to sort of, you

16:54

know, don't want to contend with the fact

16:56

that either their policies are unappealing

16:58

or their candidates suck or whatever it is.

17:00

So I think there's something to that sort

17:02

of turn on on Silicon Valley from the

17:05

right that has sort of become this supposed

17:07

free speech campaign is

17:11

pushing back against the fact that content

17:14

moderation, the political side of content

17:16

moderation policies are seem

17:19

to do more damage to conservative communication

17:21

online than in the US than

17:23

liberal progressive communication online. And we should know like

17:25

most content moderation stuff that they're doing is about

17:27

like child porn and people

17:30

selling drugs and like all sorts of stuff

17:32

that does not have a conventional sort of

17:34

partisan attachment to it.

17:36

But the fact

17:38

is that, you know, the online

17:40

right figured

17:44

out how to use social media more

17:46

effectively for certain types of communication than the left

17:48

did. And, you know, there

17:50

were all of those, you know, when

17:53

Kevin Roos from the New York Times started

17:55

publishing the top 10 most

17:58

traffic things on Facebook every day. from

18:00

CrowdTangle, you know, it was always Ben Shapiro,

18:02

and it was like seven or eight of

18:04

the ten were like overtly conservative day after

18:06

day, right? And when people

18:09

started mapping how information

18:11

moved online, there was an

18:13

asymmetry to it, which is

18:15

that these right-leaning

18:19

sites, many but not

18:21

all of whom trafficked in

18:23

stuff that was, you know,

18:26

inaccurate or disreputable, had

18:28

a much stronger hold and

18:31

sort of network analysis on traffic

18:33

than the, you know, mostly

18:36

mainstreaming news organizations that were providing

18:38

stuff to folks on the left.

18:40

And so I

18:43

think that folks on the right basically

18:46

think that all of this is a correction

18:48

to weaken the hold on a

18:50

sphere of American life that they had sort of figured

18:52

out how to master. And

18:55

now I think the idea

18:57

is that the American left is

18:59

sort of not playing fairly and

19:01

trying to weaken their hold or silence

19:03

them. This

19:12

episode is brought to you

19:14

by PayCorp, the HR and

19:16

payroll software made for leaders.

19:18

It's never been harder to

19:21

recruit, hire, and engage workers.

19:23

That's why HR leaders and

19:25

frontline managers depend on PayCorp

19:27

for all things people management,

19:29

from onboarding and performance reviews

19:32

to compensation and benefits. Learn

19:34

more at paycorp.com/leaders. That's p-a-y-c-o-r.com/leaders.

19:38

Escape to Ocean City, Maryland and

19:41

discover a place that just feels lighter, where

19:44

every day feels like Saturday and

19:46

french fries are a food group, where

19:48

flip-flops are always in fashion and

19:51

seafood is always in season, where

19:53

the boardwalk is bustling and the

19:55

beach is right outside your door, where

19:58

you can rise with the tide. and feel

20:00

like a kid again. Ocean

20:02

City, Maryland. Somewhere to smile

20:04

about. Book your trip at

20:07

oceaocean.com. When

20:12

booking with other vacation rental apps

20:14

sounds like this. This

20:16

place doesn't look like the pictures. Come

20:18

on, the doors are on back. Is

20:23

there a door behind all those spiders? It's

20:29

time to try one that sounds more like a

20:31

vacation. Look

20:33

at how many spiders there aren't. Where

20:36

should we lie down for eight

20:38

consecutive hours first? Relax, you booked

20:40

a Vrbo. Post-2016,

20:47

I remember sitting in a focus group for

20:50

this other podcast I do, The Wilderness, and

20:53

I was asking people why, this was

20:55

like third-party voters, people

20:57

who decided not to vote for Trump or Clinton, and I said, why

20:59

didn't you vote for Hillary, you know, and they said, oh, she's

21:02

just awful, and I was like, what was it, was

21:04

it, and I'm thinking about all the different, was it

21:06

the email, was it this, and they're like, oh, well,

21:08

she killed all those people. I

21:11

was like, I'm sorry, what? So it's like, now

21:14

here is someone who clearly has

21:17

bought into a piece of disinformation,

21:19

right? But absent

21:22

that disinformation, is that the

21:24

kind of person who probably would not have voted for

21:27

Hillary Clinton anyway? And that, I think, is what's hard

21:29

to tell. I also think

21:31

the real, the challenge for the left is, and

21:33

this may be a bit of a straw man, but

21:36

I have noticed over the last

21:38

several years, particularly after 2016, some on the left

21:40

seeming to

21:42

believe that we can

21:45

fact-check our way out of

21:47

disinformation, and I wonder

21:49

what your reporting and research says about that.

21:52

Yeah, so I think there are a lot of reasons to think

21:54

that kind of the conventional way of pushing

21:56

back on some of the things that are

21:58

happening, and I think that stuff just is

22:01

ill suited for this, right? So a few reasons.

22:03

One, there

22:07

is the, what

22:09

people call the Streisand effect, right? So

22:11

you take something that's maybe not circling

22:13

at a high level or reaching a lot of people, and

22:16

you come out in the Biden campaign. We could talk a

22:18

little bit about the Obama experience because you guys had a

22:20

very different posture towards this in 2008. But

22:22

the idea that you'd come out and say,

22:25

or if you're Hillary's campaign, and you would

22:27

say, she has not killed anybody, might send

22:29

people looking for it. Wait, what? Two,

22:32

there's actually a lot of cognitive science that suggests if

22:34

you go out and say, I did not kill anybody,

22:36

and you don't really frame it the right way, you

22:38

actually reinforce the suspicion among people who

22:40

might have thought possibly you killed somebody,

22:42

that there's something there. The

22:45

whole structure of the internet, of the

22:47

algorithmic internet, social media in particular, is

22:49

that in trying to

22:52

fact check or dunk on or

22:54

sass somebody who posts something wrong,

22:57

you can end up, by

23:00

engaging with it, sending more traffic to

23:02

it. So it is directly

23:04

counterproductive. And then I think the bigger thing is

23:07

a general sort of communications playbook thing, which is

23:09

if you spend all day

23:11

just answering what other people say about

23:13

you, you have only a limited capacity

23:15

to communicate on the things you care about.

23:19

To your persuadable voters,

23:21

to your donors, to your supporters,

23:24

to the media. And

23:26

if you spend all day responding to crazy things

23:28

that people say about you online, you will run

23:30

out of the ability to talk about the things

23:32

you want to talk about and you'll be. And

23:35

so, and then

23:38

I think there's the basic thing, just common sense,

23:40

which is part of the

23:42

reason we're in this situation is that

23:44

people are fundamentally non-trusting of a large

23:47

significant portion of the electorate is not

23:50

trusting of traditional news sources or other academic

23:52

institutions, the people who produce fact checks. And

23:54

so I think it, and

23:56

that those organizations have trouble getting in front of the

23:58

people who. who are in

24:01

information silos or filters, whatever metaphor we

24:03

use. And so the idea that if

24:05

only they saw this thing from PolitiFact

24:07

or the AP, it would change their

24:10

mind. No, the reason that they are

24:12

open to that is related

24:14

to the fact that they are fundamentally

24:16

distrustful of the Associated Press. And so

24:19

we're just not addressing the issue.

24:21

This to me is the sort

24:23

of fundamental asymmetry between left and

24:25

right in trying to fight

24:28

disinformation is because the

24:30

right, especially in the Trump era now, is

24:32

a coalition of low trust

24:35

voters. And that's

24:38

okay because their project

24:41

under Donald Trump is to continue

24:43

to erode faith in institutions, to in

24:45

some cases tear down institutions. And so

24:47

that works for them if people are

24:50

pissed at institutions, even if their

24:53

people are running the institutions. It doesn't

24:55

matter. They're just trying to confuse people.

24:58

The left has now

25:00

been put in the position of,

25:03

especially Joe Biden, being the defender

25:05

of institutions, the liberal

25:07

democratic order, but of course all

25:10

the other institutions that come along with that. And

25:12

so we're trying to like, it's

25:15

an uphill battle trying to both

25:17

fight disinformation and trying to

25:20

defend a set of institutions

25:22

that people have lost faith in. Yes. And it's

25:24

not the job of any political campaign to protect

25:28

institutions or educate voters for its

25:30

own sake. And so I think

25:32

that there is this sense that

25:35

the folks I write around this book are pushing

25:38

back on, which is we need to, and

25:41

so much of the disinformation, the

25:44

Hillary killed people woman.

25:46

That question, which

25:49

is, is this somebody who theoretically was persuadable,

25:51

got this bit of information and it pushed

25:53

them away from Hillary Clinton? Plausible.

25:56

Yeah. Is this somebody who, Clinton

26:01

was receptive to this ridiculous

26:04

claim and then used it

26:07

to justify what she had already decided to do

26:09

or wore it as a token of identity. So

26:12

much of on

26:15

any given day of, if you're on

26:18

the Biden campaign, so much of the stuff you

26:21

would see on the internet attacking Joe

26:23

Biden is being created by Trump supporters

26:25

to amuse and impress other Trump supporters.

26:29

It's like an Etsy

26:31

DIY marketplace, but with just memes

26:33

and gifs and what Trump has

26:35

done that's

26:38

sort of remarkable, I think, is that

26:40

he, by retweeting and sharing, endorsing, whatever

26:44

stuff that people create, has helped

26:47

to turn this

26:49

cargo cults into really

26:52

formalize it. So

26:54

now you're not just creating lies and conspiracy

26:56

theories and stuff to impress your friends and

26:58

build the community, but if

27:01

the leader takes it and holds it up

27:03

for everybody, you've succeeded. And so

27:05

the difficulty for the Biden campaign, I write a lot

27:07

of how they sort of navigated this in 2020, and

27:09

it's similar to what they're thinking about today, is

27:12

how do we measure, and this is

27:14

just a, how

27:16

do we figure out the small share? If you spend

27:19

all day, if you're a public

27:21

figure or institution, people

27:23

are going to be lying about you and the stuff you care

27:25

about, issues you care about every day online, probably

27:28

for the reasons we just talked about, 99% of

27:31

it you should ignore. It would be counterproductive to

27:34

engage with. The 1% could

27:37

be a real problem. And

27:39

so figuring out what is the 99% that

27:42

is not actually an electoral problem, it's not getting

27:44

in front of the small share of voters who

27:47

are persuadable on topics that they

27:49

care enough about to change their vote choice, or

27:52

in fact, their likelihood of turning out, their

27:54

voter's likelihood of turning out. That's

27:59

the challenge. is

28:01

isolating, getting

28:04

data to make an intelligent determination about

28:06

when is this going to make the

28:08

jump from the sort of insular community

28:10

of people who use disinformation not to

28:13

persuade other people, but I think to just sort

28:15

of affirm their community and

28:17

get in front of a share of people

28:20

or people who by definition are lower information,

28:22

later to tune in and how

28:27

do you respond to it then without

28:32

reinforcing it or appealing

28:34

to forms of authority that those

28:36

voters are not likely to respect

28:38

in the first place. So talk

28:40

a little bit about how the

28:43

Biden campaign did this

28:45

in 2020 because, you know,

28:47

Rob Flaherty is one

28:49

of the big characters in your book. He was just

28:51

done offline a couple of weeks ago. He's now the

28:53

Biden's deputy campaign manager, but he was the digital director

28:55

on the 2020 campaign, and

28:57

they sort of had a new strategy

29:00

around how to deal with myth out

29:02

of both track and then fight disinformation.

29:04

Yeah. So Flaherty and the

29:06

Biden digital operation were in the summer of 2020

29:08

blessed with these two

29:11

things you don't get a lot. They had a fair bit

29:13

of time by campaign standards and more money than they

29:15

actually knew how to spend. And

29:17

so Flaherty had been thinking a lot. He'd worked on

29:20

Clinton's campaign in 2016 and so had sort of been there at

29:22

the origins of

29:25

like the panic around disinformation and had

29:27

sort of regretted that they did not

29:29

take it seriously early enough to do

29:31

anything about it. And so he'd always

29:33

wanted to do something. And so they

29:35

end up in the summer of the

29:37

campaign and there's this money and time

29:39

that comes their way. And

29:41

he decides that they

29:44

really want to dig into identifying

29:46

what he calls distinguishing this market

29:49

moving information. He says, you know,

29:51

we have one goal. It's to get to 50 plus

29:53

one. We do not care about

29:55

anything that's online unless it's helping us get there. We

29:59

need to be really. ruthless about not getting

30:01

distracted by disinformation that isn't actually an

30:03

electoral problem for us. And

30:07

we need to, as a

30:10

result, focus just on that, which is going

30:12

to get in front of those persuadable voters in the, you know,

30:14

whatever it was, seven or eight states that

30:16

they care about on the issues that will

30:20

affect their vote. And

30:23

what they did was they worked with this

30:25

firm called Bully Pulpit Interactive, which was one

30:28

of the big democratic digital firms. They

30:32

did, they tested a number

30:36

of basically viral narratives that were

30:38

circulating about Biden online. They called

30:40

them all disinformation. It's not

30:42

a great term for this because some of them are based

30:45

in a fair bit of truth and some of

30:47

them are not. And

30:49

they polled a universe of persuadable

30:51

voters and they basically care

30:54

to ask two things. They

30:56

said, would you be familiar with this message? Would it

30:58

affect your, with the storyline

31:00

rather, and would it affect

31:02

your vote choice between Trump and Biden? And

31:04

they mapped this on an axis. So one

31:07

axis was the reach,

31:09

how many people were familiar

31:12

with the storyline. And the other axis

31:14

was the impact. Would

31:16

it affect their vote? And

31:19

so they found, for

31:21

example, you know, and the idea was only

31:23

stuff in the upper right hand quadrant where

31:25

it had both reach and impact was worrisome.

31:27

Anything in the bottom, let's keep an eye

31:29

on it. Maybe it can make

31:31

a jump. Maybe voters opinions can change, but we don't

31:34

need to do anything on it. What they found, for

31:36

example, was that the Hunter Biden storyline had

31:39

a lot of reach. People were very familiar with it. Trump

31:41

had already been impeached over this, but

31:44

they- They had a lot of reach

31:46

even though it was apparently censored. It was, yes. Even

31:49

though the laptop was censored. We did not get to see

31:51

the dick pics, but people were still aware. Yeah, we're still

31:53

aware, right? And

31:55

but it didn't have a lot of impact. Voters said it wouldn't

31:57

affect their choice, and they did focus groups, and it revealed that

31:59

in the- In this case, voters

32:02

did not see, this fairly

32:04

small share of persuadable voters did not see

32:07

Biden as fundamentally motivated by personal gain. So

32:09

even though they were familiar with all the Hunter

32:12

Biden stuff, they did not actually change their impression

32:14

of Biden on something that mattered. The

32:16

age stuff, what they called the Sleepy Joe

32:19

storyline had both reached, people obviously knew that

32:21

he was old, and

32:24

it had impact. Voters said it could affect

32:26

their choice. And the focus groups revealed it

32:28

wasn't actually because they saw him as, that

32:31

they were worried about physical fitness. And

32:34

the thinking in the campaign and the sort

32:36

of traditional communications team for a while had

32:38

been, okay, we need to deal

32:40

with this. We're going to have photo ops of him

32:42

on a bicycle. We're going to have him jog up

32:44

the steps of the plane. And

32:46

the focus groups revealed that these

32:49

voters were not, they weren't worried that like he wouldn't get

32:51

his steps in in the White House, right? They

32:53

saw him as a fundamentally weak political figure. I think

32:55

a lot of this has to do with being defined

32:57

as vice president. He got through

32:59

that big primary field and was like never the main character.

33:02

And voters didn't really know what he

33:06

cared about and what he would do. And so they

33:08

maybe had suspicions that he wouldn't be calling the shots

33:10

or that he would be following somebody else's agenda. And

33:13

that manifest itself in a

33:15

receptivity to claims

33:17

that he was frail and

33:19

old. But

33:21

this research revealed that wasn't the underlying

33:24

issue. And so I quote

33:26

Danny Franklin, who's a pollster, didn't work

33:28

for the campaign. He says, you know, fundamentally

33:31

they weren't worried about his fitness. They were worried

33:33

he was not going to be the author of

33:35

his presidency as the line. And so

33:38

the way that they pushed back against that was

33:42

not to do things that look like

33:44

they were responding to age. But

33:47

if you found yourself Googling

33:50

Biden and senile, for example, in the fall of

33:52

2000, you might

33:54

get a cookie dropped in your browser so

33:56

that when you go to YouTube the next

33:58

time, the pre-roll. video you see

34:00

was like 15 seconds of Biden

34:02

speaking to camera about how

34:05

growing up in Scranton gave him this

34:07

view of economic policy. And

34:09

this was the stuff that tested best because

34:11

the voters who were receptive to this weakness

34:14

were reassured by saying, oh, this guy can

34:16

actually articulate what he cares about and why.

34:19

And so how

34:21

Flaherty defined this shift in thinking

34:24

was to go from having a

34:27

what he called a supply side approach

34:30

to disinformation, which I think is

34:32

generally the way that the media,

34:34

a lot of academic researchers, certainly

34:37

the default assumption of people in politics has

34:40

been, which is, okay, where's this coming from?

34:43

What platform is it on? Did people use

34:45

AI to make it? How's

34:47

it spreading? Are they making money off of it? And

34:52

instead what he called a demand side

34:54

approach, which is, what are

34:57

the underlying anxieties or concerns of

34:59

the voters who might be open

35:02

to a given storyline and how can

35:04

we address those without

35:06

playing whack-a-mole with whatever the new

35:09

trending thing is today? And

35:11

it is, you know, I think

35:13

both a real tactical shift in thinking

35:16

about how you do this, but it also is

35:18

a sort of conceptual shift that I think many

35:20

of us who are worried about this as a

35:22

problem should think about. I think way too many

35:24

of our conversations and news coverage are sort of

35:26

about what plot, you know,

35:28

who's doing it, how it's moving, what tools

35:30

they're using, and not enough about like the

35:32

fundamental question, which is basically

35:34

why such

35:36

a significant portion of our country right now, gullible,

35:39

so gullible and susceptible to certain storylines over

35:41

others, right? There are tons of conspiracy theories

35:44

that are going to travel today that are

35:46

not going to catch on, right? That somebody's

35:48

going to create on a 4chan board and

35:50

is going to go nowhere. QAnon,

35:52

start on a 4chan board and seven

35:55

years later is driving aspects

35:58

of our politics. trying

36:00

to think about and research why some of

36:03

these, you know, why is vaccine skepticism around

36:05

COVID so persistent,

36:07

but, you know, people

36:09

stopped worrying about fluoride in their water, right,

36:11

why, and if you care about

36:13

pushing back on them to

36:16

figure out what the underlying issue is there instead of

36:18

just, you know, chasing whatever

36:21

today's meme or gif is. I

36:23

will say from a, from

36:25

my experience as a campaign hack,

36:28

it feels very intuitive, this

36:30

whole, and because if you're looking at

36:32

it from a campaign perspective, you of course want to

36:35

focus on the demand side and not the supply side.

36:38

And even hearing like the,

36:41

how the Biden age concern is

36:43

really a concern about weakness,

36:46

that is also something that candidates

36:49

have always had to deal with. Like we had

36:51

to deal with that. Obama wasn't old. It was

36:53

actually the opposite problem. Obama was too young, so

36:55

he was going to be weak. Obama was too

36:57

inexperienced, so he was going to be weak and

36:59

not the author of his own story. Then once

37:01

he was the president, then

37:03

it was he's too aloof,

37:06

he's too nice, he's

37:08

too professorial, but they all, the

37:10

underlying challenge was that people were

37:12

worried he was not going to

37:15

be as in control of events.

37:18

And, you know, we did the same thing in, in 08. We

37:21

tested some of the

37:23

scariest attacks that

37:25

we thought would be most damaging to Obama. We

37:28

tested like, I think at some point, I

37:30

think Axorad and company like created fake ads

37:32

to test with focus groups, right? Of like,

37:34

how was the Bill Ayers thing going to

37:37

play? How was Reverend Wright going to play?

37:40

And the challenge now

37:42

is, all right, so you identify

37:44

the most damaging narratives based on

37:46

the information environment in which you're

37:48

operating, conspiracies, QAnon, whatever it may

37:50

be. Then the question is,

37:53

how do you determine the

37:56

best way to counter those narratives? And

37:58

then how do you... disseminate

38:00

that message to the right people. Yeah.

38:03

And I should say, like, I think a lot

38:05

of this shift, a lot of this does seem

38:07

intuitive to people who were

38:09

thinking about message strategy informed

38:11

by opinion research, but

38:14

a lot of the decisions about responding

38:16

to this were being done by digital

38:18

operatives who... And I think this is... That's

38:20

interesting. There's a big... There has been in

38:22

democratic politics for a big disconnect between those two

38:24

groups of people. Yes. And so, I think some

38:26

of it's a mindset and tools issue, which is

38:29

if you are a digital operative...

38:31

And there's a sort of fusing of

38:33

people like Rob Farady, who are... And

38:35

we're just starting to see, I think,

38:37

a generation of people who came up

38:39

through digital communications, sort of move into

38:41

the normal hierarchy of a campaign instead

38:43

of staying

38:46

there. But often, they were

38:48

looking at the social listening

38:51

data that was content-driven. It

38:54

was what are the posts that are moving? What

38:56

accounts are they going through? They were not focused

38:58

on... Polling

39:01

is fundamentally focused on voter opinions.

39:04

This was focused mostly on what

39:06

content is moving. And so, I

39:09

think some of it took the

39:11

Biden campaign bringing

39:14

together a digital communications apparatus

39:16

that had an ability

39:18

to track and measure data and

39:21

come up with some of the targeting tactics

39:24

so that you were putting

39:26

these responses in front of the people who were

39:28

most likely to be susceptible to them. But

39:32

fusing it with opinion research, polling

39:35

focus groups that was focused on the voter

39:37

side of it. The... One of the...

39:40

I think that there's this ongoing debate in

39:43

progressive politics right now about

39:47

what do you do with the conservative

39:49

media ecosystem? You have groups

39:52

like Sleeping Giants, which have been

39:54

very effective, I think, at driving

39:57

boycotts and... scaring

40:00

advertisers away from sites like Breitbart

40:04

because of their content. The Biden

40:06

campaign made a different decision. We are

40:08

going to try to buy ads on

40:10

Fox News and Breitbart because their research

40:13

suggested that at least some portion of the

40:15

people who would be exposed to this, who

40:17

were persuadable, that's where

40:20

they were getting information and that they wanted to

40:22

be adjacent to it. And so I think

40:24

that this is going to

40:27

be an ongoing problem, especially as the view

40:30

of social

40:32

media sites themselves are becoming ... It

40:36

was not morally controversial 12 years

40:39

ago for the Obama campaign to decide we're going to

40:41

put money on Facebook. People might have questioned the efficiency

40:43

of it or tactile. It wasn't like we're going to

40:45

give money to Exxon Mobil or Halliburton or something. And

40:47

now I think you have a lot of progressives who

40:49

view some of the social media companies as

40:55

analogous to tobacco

40:58

or chemical companies or whatever.

41:01

And so I think that there's going to be ... Probably

41:04

it's going to break out more into the

41:06

open going forward as to what extent do

41:08

you spend money in the places where ...

41:11

When we saw this about should

41:13

Democrats give interviews to Fox News? I was just going to

41:15

say, I've always been on the side of the

41:17

moral argument is almost beside the

41:20

point. My view is the

41:22

campaign operative or former campaign operative, which is,

41:24

is it right or

41:27

wrong to go on Fox News? Are you

41:29

helping fuel their ratings? Are you doing this?

41:32

And it's like, is it effective or not? Is

41:34

there evidence that if you

41:37

go on Fox News and you are able

41:39

to communicate in a way ... If you're

41:41

debating Sean Hannity in a three minute interview,

41:43

perhaps it's a fucking waste of time because

41:45

you're not getting your message out at all.

41:48

But are there people, even if it's a tiny

41:50

slice of people, are there people who are getting

41:52

their information from that news source that

41:55

are, like you said, that are potentially persuadable and

41:57

if so, how are you going to

41:59

persuade them? Right, like, how are you going

42:01

to reach those people? In what ways are you

42:03

going to reach those people? And if they're only

42:05

getting their information from a whole bunch of places,

42:08

that's one thing. But if they're only getting their

42:10

information from Fox, and they're still not decided that

42:12

they're going to vote for a Republican, then yeah,

42:14

maybe maybe Democrats should play in that information space.

42:16

Yeah, no, I think that this is, you know,

42:18

there are it's gotten harder to target political

42:23

advertising. I mean, it's one of these these

42:25

areas where the sort of innovation narrative turned

42:27

on its head. I think, you know, Apple

42:29

killed the cookie business. So there are a lot of things that might have

42:31

been available to campaigns 10 years ago, to

42:33

try to target individual voters online.

42:35

That's interesting. That's changed. It has

42:37

gotten more difficult now. And so

42:40

I think increasingly,

42:42

it's become important to get into the networks

42:44

where the people you want to reach are

42:46

as opposed to using the sort of infrastructure

42:49

of ad targeting to, to get

42:51

to them as they move around the web. This

43:00

episode is brought to you by PayCore, the

43:02

HR and payroll software made for leaders. It's

43:05

never been harder to recruit, hire

43:07

and engage workers. That's

43:09

why HR leaders and frontline managers

43:11

depend on PayCore for all things

43:13

people management, from onboarding and performance

43:15

reviews to compensation and benefits. Learn

43:18

more at paycore.com/leaders.

43:21

That's paycore.com/leaders.

43:27

A lot of debate out there on what

43:29

to put on your Johnsonville brat. Sauerkraut, no

43:31

sauerkraut, peppers and onions, no peppers and onions,

43:33

or maybe peppers, but no onions. But what

43:36

kind of peppers? And then mustard, we

43:38

doing brown spicy or are you going to go amateur hour?

43:40

And then ketchup's its own situation, which by

43:43

the way, is absolute heresy. But

43:45

here is the beautiful truth. Johnsonville don't

43:47

care. If people are debating relish, bun

43:49

or no bun, at least we're all

43:51

talking and sharing again. And yeah, the

43:53

ketchup's a personal choice. Keep it juicy.

44:00

Maryland and discover a

44:02

place that just feels lighter. Where every

44:04

day feels like Saturday and french fries

44:06

are a food group. Where

44:08

flip-flops are always in fashion and

44:11

seafood is always in season. Where

44:14

the boardwalk is bustling and the

44:16

beach is right outside your door. Where

44:18

you can rise with the tide and feel

44:20

like a kid again. Ocean

44:22

City Maryland, somewhere to smile

44:25

about. Book your trip at

44:27

oceocean.com I

44:34

was talking to a senior Biden person

44:37

last week and they

44:40

were saying that the the biggest

44:42

challenge, they said the entire campaign comes down

44:45

to you know there was this

44:47

this chart of like Biden's you know winning

44:49

by 50 points with people who

44:51

get their news from newspapers and and

44:54

and network news also you know and

44:56

the places where Biden's losing

44:58

the most and is behind Trump by the

45:00

most are people who either get

45:02

most of their news from social media or people

45:05

who don't tune into the news

45:07

at all don't follow the news closely at all and

45:10

that Biden person was saying like their big challenge

45:12

is trying to figure out

45:14

how to reach those people. How

45:17

would you assess how the Biden campaign is

45:19

is doing that

45:21

in 2024 and how they're handling disinformation and

45:23

the information

45:26

environment in 2024 particularly

45:29

compared to what you wrote

45:32

about in 2020? So I you know

45:34

I think that there's far more focus on

45:36

recruiting influencers you

45:39

know not just the people who do makeup

45:41

routines and but

45:43

people who have who are

45:46

not necessarily political communicators foremost

45:48

but have circles

45:50

of readers, followers

45:52

who who rely on them for

45:54

information and you know that

45:56

the fundamental tension here is between

46:01

the command and control mentality of the

46:03

modern political campaign

46:07

and the fact that the internet does not

46:09

lend itself to top-down communication.

46:14

And you go back to, oh wait, the

46:16

things that the Obama campaign chose to use

46:19

the internet to do very effectively were

46:24

get contact information for people so that you could ask them

46:26

for money in new ways and ask them to volunteer in

46:29

new ways and cut down the friction of having to go

46:31

into a field office and allow them to make phone calls

46:33

from their dining table. But in every

46:35

instance, you are either asking them to do

46:38

something very specific and targeted. And

46:40

if they were going to talk to other voters, usually

46:42

it was offline. And

46:44

it was with messaging and training provided

46:46

by the campaign. And

46:50

this is why when people come into a field

46:52

office, campaigns don't tell

46:54

them, hey, run over to whoever you love

46:56

to talk about politics with and argue

46:59

with them for our candidate. Here's

47:02

a list of people that we have identified and

47:04

here's the messaging that we think will work with

47:06

them. And don't go beyond

47:08

the bounds of what we tell you to

47:10

do. And what

47:12

Trump did, I think, more a function

47:14

of his instinct than any great strategy

47:16

was he helped cultivate a community where

47:19

he sent a signal by his own

47:21

behavior because at 3 a.m. he was

47:23

tweeting about whatever crazy

47:26

shit came to mind or whatever he saw that

47:28

he thought was cool or funny. And

47:30

he sent a signal to his

47:32

supporters that it's good.

47:34

It is useful to my campaign for you to just

47:36

post whatever you want about

47:39

me. That was never

47:41

a signal that traditional campaigns, not

47:43

just Democratic campaigns, John McCain was not sending a

47:45

signal to people, hey, you know, say whatever you

47:47

want about me or my opponent. But

47:49

I think that, you know, what

47:51

we have seen is that Democratic

47:54

campaigns, I would argue that the

47:56

Obama campaign, which we certainly until

47:58

2016 said was the most trying

52:00

to influence what the newspapers and TV networks are saying

52:02

and your paid advertising is still a thing. But

52:05

you need to play in a world

52:08

where word of mouth and rumor and

52:11

person to person communication and homemade

52:14

messaging are going to be

52:16

a way of reaching people. And

52:18

I don't see that the Biden

52:20

campaign has gone quite that far yet.

52:23

No, it's funny because it's different parts of the

52:25

campaign. It's sort of like your

52:27

background, right? Like you said, there's the

52:29

difference between the digital organizers and the

52:31

message polling people, also the comms people,

52:34

right? The

52:37

generation of communications people today are still trained

52:39

by our generation of communications people, which is

52:41

like, here's the main message for today and

52:43

here's the talking points, but it all feels

52:45

like it's very stilted, like if it leaked,

52:47

we would all be okay kind of thing.

52:50

And the other part of it is I think a

52:52

lot of those people were shaped by the Swiftboat Experience

52:54

in 2004, which was this traumatic

52:56

event for democratic communicators. And the lesson

52:59

from it that you guys internalized in

53:02

08 was

53:04

don't let anything that anybody bad

53:06

says about you go unanswered because it

53:08

could start as this little thing in

53:10

a small media market in Pennsylvania. And

53:12

if you let it move, it's going

53:14

to get everywhere. And

53:17

that's terrible advice now. But

53:19

I do think that... I think

53:21

I see so many examples of fucking

53:23

swinging at every pitch. And

53:26

I get it because, I mean,

53:29

part of this is the challenge again about a

53:31

challenge of running against Donald Trump.

53:34

And Trump gives you so many targets. And

53:37

he says so many crazy... He

53:39

floods the zone with shit, whether intentionally or not.

53:42

And so you feel like you have to

53:44

swing at every pitch. And

53:46

there's a ton of attacks on you, so you have to

53:48

defend each one. But even the ones that you know are

53:51

damaging, right? Just today, we're

53:53

talking on Wednesday. And Wall Street

53:55

Journal had some story where they

53:57

talked to a bunch of... Republicans

54:00

basically who said that Biden is

54:02

slipping, right? And they

54:05

don't have any Democrats on the record. They only

54:07

have like Kevin McCarthy on the record, which is,

54:09

it's like ridiculous, you know? And

54:11

I only noticed the story

54:13

because I went on Twitter and like all

54:16

like Biden supporters are sharing the story to

54:18

dunk on it. And I'm like, I

54:21

don't know that this would have been a big deal otherwise.

54:23

And I realized that the age thing is

54:26

that damaging narrative. I'm sure it's even more

54:28

so now than into 2020 for

54:30

a host of reasons. But

54:32

I wonder if there's

54:35

like best practices in, you

54:37

know, prioritizing certain narratives and then

54:39

figuring out how to respond in

54:41

a way that doesn't amplify the

54:43

attack itself. Yeah, and I think

54:45

in this case, you know,

54:48

I think that they're sort of torn between a

54:51

very old fashioned idea of trying to send a signal

54:54

to other reporters about what, you

54:57

know, about the quality of reporting. I

55:00

think that there's a, but the fact is that,

55:02

yeah, so the people in the White House are

55:04

on the campaign who are pushing back are probably

55:06

trying to communicate with basically nine reporters at the

55:08

Washington Post and Bloomberg who might do their own

55:10

version of this story. And try to

55:12

shame the wall. Try to shame, yes. With there's value

55:15

in doing, of course. Absolutely, but the fact is that

55:17

then you have a whole community of, you know, lefty

55:20

online personalities who

55:22

have their own audiences who take a signal from,

55:25

oh, well, you know, Andrew Bates at the White House

55:27

is shaming the Wall Street Journal. That's what

55:29

we should be doing today. And then it becomes the

55:32

dominant thing that the pro-Biden internet

55:34

is on today is shaming a reporter at

55:37

the Wall Street Journal about a

55:39

story that probably very few

55:42

voters who, you know, whatever have seen to begin

55:44

with. Or if they did, it would be like,

55:46

oh yeah, I do think Joe Biden's hold up

55:48

for that for a while. Yeah, right. But

55:51

it's funny because like I tweeted, you know,

55:53

a story where, the only thing I

55:55

did is I did not respond to what I tweeted a

55:57

story where there's like seven Republican senators who were like, yeah,

58:00

that people have sort

58:02

of adjusted by degrees, but have not

58:04

had a wholesale rethinking of everything we

58:06

thought about how to communicate was

58:08

tethered to this environment. And those 12 news

58:10

organizations, we should note, like, you know, we're

58:12

at a moment where American

58:15

media aspire to a certain type of

58:17

neutrality that is close

58:19

to unique in the world. And

58:22

we just have this peculiar confluence of factors

58:24

in the late 20th century that

58:28

the created communications playbook that is

58:31

no longer applicable. And it's

58:34

going to take, you know, a significant

58:36

shift. And I think the minds of

58:38

people who are making decisions inside campaigns

58:40

to fully break away from

58:43

it. Yeah, and again, it's like, you

58:45

know, mission accomplished with trying to

58:47

shape the

58:50

traditional media in a way that's going to be

58:52

beneficial to Joe Biden and not to Donald Trump.

58:54

Because again, all the people who are reading the

58:56

New York Times, all the people who are watching

58:58

CNN, they're voting for Joe Biden. Or at least Joe

59:00

Biden's winning them by a huge amount, way more than

59:03

he'll ever win the presidency by. But then

59:05

I think of all the platforms now, especially in 2024, like

59:07

TikTok, one

59:09

of the reasons it is so scary to

59:12

I think a lot of Democrats right now and problematic

59:14

is because you can't even track the

59:17

misinformation, the disinformation on TikTok or the

59:19

propaganda or whatever you may want to

59:21

call it. And that I don't know

59:24

how you solve for, because

59:26

you can, you know, like the Biden

59:29

campaign is on TikTok, great. And

59:31

they're talking to some influencers, but then like beyond

59:33

that, how do you even know what's on there

59:35

that's damaging? Yeah, no, I think that's, you're absolutely

59:38

right. I mean, so, you know, a TikTok content

59:40

doesn't have a link that exists in the world,

59:42

right? So just as a simple fact, you can't

59:44

send somebody in a conventional way the way you'd

59:47

send them a tweet or a Facebook post. And

59:49

that makes it for researchers impossible to see how

59:51

something moves. And in addition to

59:53

the fact that the algorithm is particularly

59:55

opaque and you're getting stuff from people that

59:57

not only do you not follow, but. but

1:00:00

it's hard to tell how and why you're getting it.

1:00:02

And so

1:00:05

it is, I think

1:00:08

rightfully people are sort of

1:00:12

vexed by it. One of the main characters in the book,

1:00:14

this woman, Jory Craig, the first thing that she does when

1:00:16

she goes to work for a campaign, she was basically the

1:00:18

first counter disinformation operative in American politics. And

1:00:20

the first thing she does when she goes to work for

1:00:22

a campaign or a party is

1:00:24

what she calls a landscape analysis.

1:00:27

And it's basically, let's map the

1:00:31

communications landscape in the area we care about.

1:00:34

It could be an issue area like abortion,

1:00:36

it could be a geographic area like Arizona.

1:00:39

And let's see what are

1:00:41

the crucial nodes for

1:00:45

information to move, how are things networked,

1:00:47

and what is a baseline level of

1:00:49

communication on a given topic or theme

1:00:51

every day? So I think one of

1:00:53

the things that people who are not

1:00:55

looking at this in a sort

1:00:57

of ongoing holistic way do is

1:00:59

something's trending. Okay, you know, Sleepy

1:01:02

Joe is trending today. Well, do

1:01:05

you know if it's trending any more than it trended yesterday?

1:01:07

Do you know if it's trending when the same people with

1:01:10

whom it trended yesterday? And

1:01:13

so some of it is just to create that

1:01:15

baseline map so that when

1:01:18

you get into the ebb and flow

1:01:21

of a campaign and stuff is moving,

1:01:23

you have some sense of context or

1:01:26

proportion. And it

1:01:28

has been over the years gotten easier and

1:01:31

more difficult on some of the platforms to

1:01:33

have the data to intelligently

1:01:35

assess that. But TikTok

1:01:37

is orders of magnitude

1:01:39

more difficult to make sense of than

1:01:41

Facebook or Twitter or Instagram or any

1:01:43

of the places. Yeah.

1:01:46

Last question, I'll let you go. Other

1:01:49

than this, any individual political

1:01:51

campaign is, I think, the

1:01:54

concern that this is just incredibly destabilizing

1:01:56

to democracies, not just here in the

1:01:58

United States, but all the world. Have

1:02:01

you seen any examples

1:02:03

or signs of hope

1:02:06

that people are

1:02:08

developing tools and strategies

1:02:11

to actually push

1:02:14

this in a better direction so that we can sort of

1:02:16

get a handle on this? Or is this just still

1:02:19

sort of the Wild West and people are just trying to, you

1:02:22

know, just do what they can and individual

1:02:24

campaigns? I think it's closer

1:02:26

to the Wild West. And, you know, the thing

1:02:28

we haven't talked much about is the role of

1:02:31

the platforms themselves and their policies in this. I

1:02:33

think if you want to find causes

1:02:36

to be optimistic, it

1:02:39

would be to hope that the sort

1:02:41

of posture of the platforms is cyclical

1:02:43

here. You know, we have since 2017,

1:02:45

roughly, you know, which I think if

1:02:50

you had to pinpoint the moment where

1:02:53

the techno-optimism we started with turned to

1:02:55

whatever we're in, whatever bleak reality we're

1:02:57

in now, it's Cambridge Analytica story broken

1:02:59

in 2017. And I

1:03:02

think that was the moment that probably the consensus view of

1:03:04

people was like Facebook's probably

1:03:06

a net neutral or good

1:03:08

force in American life to a net

1:03:10

bad force in American life turned. What

1:03:14

has the platforms

1:03:17

have gone through, I think, like

1:03:19

their third sort of cycle of

1:03:22

just their posture towards content moderation

1:03:24

on politics. There

1:03:26

was a big show in 2018 and into

1:03:28

2019 of saying we take seriously, basically

1:03:30

policing content

1:03:36

around elections in particular on our

1:03:39

platform, not just

1:03:41

in the US, but, you know, Facebook

1:03:43

created this global election war room where they

1:03:45

hired a lot of people and they went

1:03:48

to the media and said, like, we have

1:03:50

our hands around this and we are committed

1:03:52

to making sure that people don't get bad

1:03:54

information on our platform. We're putting resources behind

1:03:56

it. That

1:03:59

was the. Generally,

1:04:01

I think that attitude was shared by the

1:04:04

major online platforms through

1:04:06

the 2020 election. After

1:04:11

obviously right after January 6th, basically

1:04:13

every platform took down and

1:04:15

their biggest show of commitment to this

1:04:17

principle took down Trump's accounts.

1:04:21

And then a few things happened afterwards.

1:04:26

The Republican Party organized

1:04:28

itself against content moderation

1:04:30

as a policy position

1:04:33

to Twitter,

1:04:36

which had been,

1:04:38

I think, generally putting

1:04:40

upward pressure on other platforms

1:04:42

by taking this relatively seriously.

1:04:46

After the Musk acquisition, obviously, you

1:04:48

know, created a floor whereby there

1:04:50

was very little motivation for other

1:04:52

platforms to spend more

1:04:55

or expose themselves more to public

1:04:57

scrutiny through this. And now I think, you

1:04:59

know, certainly for the rest of this year,

1:05:02

we are in an environment where

1:05:05

the companies that run the platforms have

1:05:07

basically concluded there's more to lose by

1:05:09

antagonizing the right than

1:05:12

there is to gain either in

1:05:14

civic terms or in sort of PR

1:05:16

terms by placating, you know,

1:05:18

a broad center and left that want to see a more

1:05:21

active sort of content

1:05:24

moderation. Now

1:05:27

that changed quickly and it could change again. And I think

1:05:29

it is worth, if you go back to the utopian times,

1:05:31

some of that was that, you

1:05:33

know, people like us looked at campaigns and

1:05:35

thought, hey, generally this is getting people more

1:05:38

involved and that seems good. But part of

1:05:40

it was that companies like Facebook wanted to

1:05:42

be associated with politicians like Barack Obama. Like

1:05:44

that was a very successful PR effort. You

1:05:46

know, Google and Twitter loved

1:05:49

being associated with the

1:05:51

I mean, Google with the Arab Spring.

1:05:53

I mean that and to you, I

1:05:55

could imagine, you know, there was there

1:05:57

were just the valence of this is.

1:06:00

Social movements are being created on our

1:06:02

platforms. We are bringing people into a

1:06:04

community. They worked very hard

1:06:06

to be associated with politics during

1:06:09

that period. They also, it's a

1:06:12

big advertising, you know, $16 billion or

1:06:14

something is gonna be sent on the

1:06:16

election. That's a significant share of

1:06:18

potential advertising revenue. I

1:06:20

could imagine it would

1:06:23

not take a whole lot over the next

1:06:25

two years or four years for some of

1:06:27

these tech platforms to think, okay, there's a,

1:06:30

you know, these platforms have all sent some

1:06:32

signal, like we're not really a place for political communication

1:06:34

anymore, to start saying, hey, we want

1:06:36

some of that advertising business back. And part of

1:06:39

that is we need to send a signal to

1:06:41

political communicators that, you know, in

1:06:43

the same way that Fortune

1:06:45

500 companies don't put ads in porn

1:06:47

magazines, that we are like a place

1:06:49

where they want their logo to appear.

1:06:52

Two, I don't necessarily

1:06:54

see this person on the horizon, but

1:06:56

I think one charismatic candidacy or

1:06:59

one major geopolitical movement that seems tethered

1:07:01

to the uplifting

1:07:04

potential of social media could

1:07:06

incline some of these companies to want to be

1:07:08

associated with it, or a

1:07:10

change in control of Congress

1:07:13

and the White House that align

1:07:15

in such a way that there's

1:07:18

a serious fear of regulation as there was in the

1:07:21

2018 period that

1:07:23

makes the companies say, we're gonna try

1:07:26

to stave off the regulation by doing more ourselves. I

1:07:28

think that is a conceivable

1:07:30

shift. And I think it, we

1:07:32

have seen that they can turn

1:07:35

the switch on both

1:07:37

specific policies and putting the resources behind

1:07:39

them pretty quickly and could do it

1:07:41

again. Yeah, I think that's all right.

1:07:43

And I think to your point about

1:07:46

the potential charismatic candidate,

1:07:48

I just think whether it's a single candidate or

1:07:50

a movement, I do think that the

1:07:53

pro-democracy forces do need to find

1:07:55

a way to develop these same

1:07:58

strong communities online. that are

1:08:01

bound by sort of joy

1:08:04

in what they're doing. Like,

1:08:06

we think that there is

1:08:08

this sense of belonging you get from whether it's

1:08:10

QAnon or being at a Trump rally on

1:08:13

the right, and I think we need to

1:08:16

figure out on our side how

1:08:18

to create that, or at least how to encourage

1:08:20

that, even if we can't necessarily

1:08:22

control it. So, Sasha Eisenberg, thank

1:08:24

you so much. The book is The

1:08:27

Lie Detectives. It's a great book, everyone

1:08:29

should read it. It's a quick read,

1:08:31

but just fascinating stuff if

1:08:33

you're a campaign nerd and an internet

1:08:35

nerd like me. So, thanks for sitting down

1:08:37

with me. I really enjoyed it. Thank you,

1:08:39

John. Before we

1:08:42

go, some quick housekeeping. If today's conversation with

1:08:44

Sasha about people in politics screwing shit up

1:08:46

freaked you out, or better yet, if it

1:08:48

fired you up, then boy, do

1:08:50

we have a book for you. We're officially

1:08:53

less than one month away from the release

1:08:55

of Democracy or Else, How to Save America

1:08:57

in 10 Easy Steps. It's

1:08:59

full of lessons on politics and organizing that

1:09:01

will hopefully transform you into a savvier, saner,

1:09:03

well-armed citizen. I just want to say thank

1:09:05

you if you've already grabbed a copy. Tommy,

1:09:08

Lovett, and I really appreciate it. And if

1:09:10

you haven't pre-ordered your copy, there's no better

1:09:12

time than now. Crooked Media is donating its

1:09:14

profits from the book to support Vote Save

1:09:16

America and its partners in 2024 and

1:09:19

beyond, so every book makes a real difference.

1:09:21

Also, you buy this thing early. You're helping us

1:09:23

get on the New York Times bestseller list and

1:09:25

bump some of the worst people off it. Looking

1:09:28

at you, Mark Levin. Head to

1:09:30

crooked.com/books now to pre-order your copy.

1:09:33

Also, it's Pride Month, and the Crooked

1:09:35

store is stocked with tons of stuff

1:09:37

to help you celebrate. The Pride or

1:09:39

Else collection has everything from tees with

1:09:41

two steamboat willies kissing on them, to

1:09:43

mugs with two steamboat willies kissing on

1:09:45

them, plus a ton more stuff for

1:09:47

protesting, partying, and showing up as an

1:09:49

ally. A portion of proceeds from every

1:09:52

order go to Crooked's Pride or Else

1:09:54

fund in support of organizations working to

1:09:56

provide gender-affirming care and life-saving resources to

1:09:58

queer and transgender communities across America. America.

1:10:00

Shop over 50 pieces at

1:10:03

crooked.com/store. Offline

1:10:17

is a Crooked Media production. It's written and

1:10:20

hosted by me, Jon Favreau, along with Max

1:10:22

Fisher. It's produced by Austin Fisher. Emma

1:10:24

Illick Frank is our associate producer. Mixed

1:10:27

and edited by Jordan Cantor. Audio support

1:10:29

from Kyle Seglin and Charlotte Landis. Jordan

1:10:31

Katz and Kenny Siegel take care of

1:10:33

our music. Thanks to Ari Schwartz, Madeline

1:10:35

Herringer, and Reed Cherlin for production support.

1:10:37

And to our digital team, Elijah Cohn

1:10:39

and Dilan Villanueva, who film and share

1:10:41

our episodes as videos every week. David

1:11:00

taught himself how to make bread. Good

1:11:02

bread. He wanted to get

1:11:04

even better. So he asked Chachipitio on

1:11:07

Expedia if there's such a thing as

1:11:09

a bread vacation. Chachipiti

1:11:11

said, sure. Do you

1:11:13

want to go to Normandy, Morocco, Ireland,

1:11:15

or Tuscany? And

1:11:17

that's how David became a master

1:11:19

pizzaiolo. You were made to

1:11:22

learn new things. We were made to give

1:11:24

you trip ideas with Chachipiti, right in our

1:11:26

app. Expedia. Made to

1:11:28

travel. So

1:11:57

go ahead and give them Miracle-Gro. you

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features