Podchaser Logo
Home
Reset: Reclaiming the Internet for Civil Society | Tech Expert Ron Deibert

Reset: Reclaiming the Internet for Civil Society | Tech Expert Ron Deibert

Released Friday, 26th April 2024
Good episode? Give it some love!
Reset: Reclaiming the Internet for Civil Society | Tech Expert Ron Deibert

Reset: Reclaiming the Internet for Civil Society | Tech Expert Ron Deibert

Reset: Reclaiming the Internet for Civil Society | Tech Expert Ron Deibert

Reset: Reclaiming the Internet for Civil Society | Tech Expert Ron Deibert

Friday, 26th April 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:01

Have you ever longed to escape reality

0:03

or fantasized about stepping into someone else's

0:06

shoes, even for just a little while?

0:09

Hi, I'm Laura Mullen. And

0:11

I'm Chris Tully. We host

0:13

CBC's Play Me, the immersive

0:15

podcast that transforms theatre into

0:17

addictive audio fiction. Join

0:20

us for a new season

0:22

and disappear into a world

0:24

rich with drama, where every

0:26

show delivers hypnotizing stories and

0:28

unveils intriguing characters with secrets.

0:31

Play me wherever you get your podcasts.

0:36

This is a CBC Podcast. Welcome

0:40

to Ideas. I'm Nala Iyad.

0:45

In 2020, Ron Deibert

0:48

came to the nearly empty CBC

0:50

building in downtown Toronto to record

0:52

the Massey Lectures in studio. The

0:55

COVID-19 pandemic made our

0:57

usual multi-city tour impossible. Ron

1:01

is the founder and director of Citizen

1:03

Lab. It explores global

1:06

security, information and communications

1:08

technologies and human rights.

1:11

So he's had a front row

1:13

seat to online realities that were

1:15

spiraling out of control. We've

1:18

read about high-tech mercenary companies selling

1:20

powerful cyber warfare services to dictators

1:23

who used them to hack into

1:25

their adversaries' devices and social networks,

1:28

often with lethal consequences. First

1:32

it was Jamal Khashoggi's inner circle. Then,

1:35

allegedly, Jeff Bezos' device.

1:38

Maybe I've been hacked too, you wonder to

1:40

yourself, suddenly suspicious of that

1:43

unsolicited text or email with an

1:45

attachment. We're

1:48

all living in a digital ecosystem

1:51

that often feels impossible to escape.

1:54

The world you're connecting to with that device

1:57

increasingly feels Like

1:59

a major source, a personal risk, but

2:01

it's also become your lifeline now more

2:04

than ever. Run D

2:06

Birds Massey lectures were called

2:08

Reset Reclaiming the Internet for

2:10

Civil Society. In the

2:12

language of computers and networking, the term

2:15

recess is used widely to refer to

2:17

a measure that have a system and

2:19

returns with to an initial state. In

2:23

his lecture run. Suggests.

2:25

Let's set reset Said look like. To

2:28

get us started, I propose

2:31

a single, simple but hopefully

2:33

potent principal restraint. This

2:36

episode as part of a

2:38

special series we've put together

2:40

to Marxists sixtieth Anniversary of

2:42

Massey College one of our

2:44

partners and then I see

2:46

lectures. It features the sixth

2:48

and final installment from Around

2:50

the Birds Lectures called Retreat

2:52

Reform Restraint Along the way

2:54

you'll also here other thinkers.

2:56

Anything on his way to. This

3:01

is Luxor six retreat,

3:04

Reform restraint, Part.

3:07

One a fragile ecosystem.

3:11

There. Is an undeniable gusto in the

3:13

air. A recognition that something

3:16

of our own making is contributing

3:18

to a serious kind of social

3:20

and political sickness. Are.

3:22

Very tools and techniques threatened to wipe

3:25

us out. Says sieve of i do

3:27

not in in anti social media. We.

3:30

Display unlimited talents but no

3:32

mastery. We. Process infinite

3:34

data but display know

3:36

wisdom. Is. Remarkable

3:39

to think it was only a few

3:41

short decades ago that the Internet was

3:43

heralded as a wonderful new tool. That.

3:46

Would enlighten and liberate us.

3:48

Dissolve. The boundaries of time and

3:50

space. And. Ties together more

3:52

closely into a single global

3:55

village. But. In

3:57

the span of a single generation, it

3:59

has trends. More defied into

4:01

something far more complex and

4:03

disturbing. It's. A Sex

4:05

suddenly very ominous. With.

4:07

Started out as something so

4:09

simple as desktop pcs network

4:11

together via a common protocol.

4:14

Has. Morphed into an always

4:16

on omnipresent date a vacuum

4:19

cleaning operation. Undertaken by

4:21

gigantic corporate platforms with

4:23

unparalleled abilities to peer

4:25

inside our minds and

4:27

habits. And. Suddenly shape our

4:29

choices. Their. Operations implicate

4:32

a plan applied network of

4:34

gigantic energy sucking data farms.

4:38

The. Ecosystem has spawned a bewildering

4:40

variety of invasive species that

4:42

thrive by feeding on that

4:45

continuously expanding pools of data

4:47

that spew forth each millisecond

4:49

of every day. App

4:52

developers, data brokers, location

4:54

trackers, data fusion companies,

4:56

artificial intelligence start ups

4:58

and private intelligence firms.

5:01

Accountability. Is weak and

5:04

insecurity is endemic throughout

5:06

the entire system, creating

5:08

seemingly endless opportunities for

5:10

malevolent exploitation by spies,

5:12

kleptocrats, dark Pr firms,

5:14

and other bad actors.

5:17

It's. As if we have sleep walked

5:20

into a new machine base civilization

5:22

of our own making. And.

5:24

We're just now waking up

5:26

to its unforeseen consequences. an

5:29

existential risks. It's.

5:31

Clear that there is a

5:33

growing consensus that many things

5:35

are wrong with our social

5:37

media habits. Symptoms of this

5:39

Malays are seemingly everywhere, fairly

5:41

easy to identify, an increasingly

5:43

enumerated by scientific studies. But.

5:46

What to do about them is less obvious. There.

5:49

Is nowhere near a consensus when it

5:51

comes to a cure. This

5:54

lack of clarity around solutions

5:56

is certainly understandable. The.

5:58

Challenges thrown up by so. The media

6:00

surveillance, capitalism, and near total

6:02

state surveillance have arisen so

6:04

swiftly that we have barely

6:07

had time to understand how

6:09

they work, let alone six

6:11

them. But.

6:13

Of course it's not all just

6:15

on you. We, and fatalistic acquiescence

6:17

in the face of social medias

6:19

harms. Proposals. For

6:21

mitigation or alternatives to social

6:23

media, as currently constituted, are

6:26

frequently raised and plentiful. Summer.

6:29

Quite interesting and worth considering. Others

6:32

may be flawed in various ways or

6:34

self serving. Many. That

6:36

are worthwhile or incomplete. Many.

6:39

Of them feel like fragments of a missing a

6:41

whole of which they might be a part. Take

6:45

the family of recommendations to mitigate

6:47

social media harm that I call

6:50

retreat. These. Are

6:52

the solutions that advocate for some

6:54

variation of going off the grid

6:56

either by throwing out our devices

6:58

and applications completely and going back

7:01

to a time before social media.

7:03

Or. And slightly more reasonable form

7:06

simply taking a break from them

7:08

once in awhile. Proposals.

7:10

Such as these can be found him

7:13

please to unplug. Disconnect.

7:15

Or. Perform periodic cleanses,

7:18

Digital Detoxification as it's

7:21

widely described. The.

7:24

Concept of a digital retreat as appealing

7:26

on many levels. There. Is a

7:28

simplicity to at the makes it alluring.

7:31

It is true that one thing we need

7:33

to recover his our connection to the natural

7:35

world. Slowing. Down as a good

7:37

idea to. But. There's a

7:39

major problem so obvious that it may

7:42

be easy to overlook. Sure,

7:44

It's fine if a few

7:47

isolated communities completely detached, unplug

7:49

and retreat. And. There's

7:51

no doubt we could all use

7:53

a little digital detox once in

7:55

awhile. Meditation. Is great

7:58

to and best ton without a nap. But.

8:01

Candid Scale. Would. Have

8:03

every quite literally unplugged.

8:06

How. Would we then manager ourselves

8:08

for social relationships are problems

8:10

and our politics. Detachment.

8:14

And retreat also ignored or at

8:16

least slight. The. Many positive

8:18

uses of digital technologies

8:20

social media included. In

8:23

spite of disinformation and

8:25

overreaching surveillance. Social. Media

8:28

have proven highly useful for

8:30

many problems. Even.

8:32

Prior to the Kobe nineteen

8:34

pandemic, digital technologies have been

8:36

used extensively to monitor the

8:39

environment. To. Share information

8:41

in ways that the original

8:43

designers intended. Think of: Wikipedia.

8:46

To. Mobilize social movements and hold

8:48

bad actors to account. Furthermore,

8:51

It is frankly impossible to live

8:53

in complete detachment from social media

8:56

today. Even. The

8:58

self isolation and social distancing in

9:00

response to the covert emergency. Did.

9:02

Not dissolve digital connections among

9:05

people. In fact, it deepened

9:07

our reliance on them. Even.

9:09

If you were to try and completely

9:11

escape from social media, Unplug.

9:13

Your router. Throw. Away all

9:16

of your devices and never look

9:18

at social media again. You'd still

9:20

be subject to surveillance. Facebook.

9:23

And other social media have shadow

9:26

profiles of people. Who. Do

9:28

Not Even use their services. Cctv.

9:31

Cameras are everywhere. Go.

9:33

Ahead, live the dream and moved

9:35

to rural Montana. you'd still be

9:38

watched by drones, planes, and satellites.

9:41

No. Matter where you go, you will be

9:43

counted. Digital.

9:45

Technologies of so deeply embedded themselves

9:47

into everything that we do. It

9:50

is unrealistic to expect we can

9:52

turn the clock back entirely. Nor.

9:55

Should We. We. need an

9:57

open and secure means of communicating

9:59

globally in order to manage

10:01

our planet and our collective affairs. It

10:04

is just that the current design for it, based

10:07

around personal data surveillance, is

10:10

counterproductive to those aims. Outright

10:13

rejection of social media is

10:15

thus both undesirable and futile.

10:18

As Buckminster Fuller once said, we

10:21

do indeed live in spaceship Earth,

10:23

but we are stuck with a poorly

10:26

designed operating manual. Hi,

10:33

Tamsen Shaw here. I

10:36

like Ron's metaphor of

10:38

a poorly designed operating manual,

10:40

because if you

10:43

think about the hybrid economy of Silicon

10:45

Valley and what it was set out

10:47

to accomplish, a huge

10:49

amount of the funding came

10:51

from national security agencies, from

10:54

their venture capital wings, in

10:56

order to accomplish national security

10:58

goals. But of course, the

11:00

companies wanted to commercialize those

11:02

products in order to

11:04

make a profit. So those are

11:07

the two basic

11:09

functions of that machine,

11:11

capitalism and national security. And

11:14

I think now there's no technological

11:16

reason that we can't create social

11:18

media platforms that don't just serve

11:20

those two ends. But the

11:22

problem is where does the incentive come from?

11:25

It's not going to be from

11:27

the people who own those tech platforms

11:29

and make billions of dollars from them.

11:32

It's not going to be from politicians who

11:34

now rely on them to get elected. And

11:37

it's not going to come from people whose only

11:39

real form of social

11:41

and community organization is

11:43

actually on those platforms.

11:46

And I think this is symptomatic of

11:48

something that's true of humans in

11:50

general, which is that technological

11:52

problems are the easy part.

11:55

It's the very big problems

11:58

like climate change. or

12:00

global inequality that are transnational

12:03

and that require political solutions

12:05

that we have problems with.

12:08

Marxists used to imagine

12:10

that there were these contradictions

12:13

of capitalism and that

12:15

eventually they would be over-determined and

12:17

that would lead to us creating

12:20

something better. I

12:22

think the worry now is that maybe they

12:24

don't lead to something better,

12:27

over-throwing capitalism. Maybe they

12:29

just lead to our self-destruction and

12:31

I think that's the problem that we have to confront now.

12:34

That's what we'll need to address in

12:37

the future. It's Misha

12:39

Glenny here. I've felt

12:41

for a long time that what we face now

12:43

are what I call the four horses

12:45

of the modern apocalypse. One

12:48

of them we're going through at the

12:50

moment, pandemic and the other three are

12:52

climate change, which is the ultimate threat,

12:55

weapons of mass destruction, which

12:57

over the past 20 years have been

13:00

proliferating quite significantly

13:03

and the regulatory systems are deteriorating

13:05

for that. But the

13:07

fourth one is our

13:10

over-dependency on network

13:12

computer technologies. I think

13:15

this is a really, really serious danger

13:17

that we face. And so

13:19

for me, if we

13:21

do not address these issues and

13:24

come up with a new

13:26

system of managing network

13:29

technologies, then there

13:31

is little doubt in my mind that

13:33

they will destroy us. Part

13:36

Two. Reform. Reset. Then

13:44

there are the proposals that advocate

13:47

for some variation of reform, that

13:49

is adjustments to one or another

13:52

element of social media's business practices.

13:55

Reform proposals range along a spectrum

13:58

from minor to major adjusted. adjustments,

14:00

and lesser to greater degrees

14:02

of formal government intervention. The

14:06

intent is not to dismantle social

14:08

media, but to fine-tune them instead.

14:11

For example, corporate social

14:13

responsibility, ethical design, and

14:15

other such initiatives typically

14:18

involve the least intrusive measures

14:20

and entail only minor fixes

14:22

to superficial elements of social

14:25

media. These

14:27

initiatives typically advocate for norms

14:29

rather than laws, persuasion

14:31

rather than coercion. The

14:34

goal is to have business

14:37

executives acknowledge certain principles and

14:39

govern their business practices accordingly,

14:41

but without any kind of specific

14:44

enforcement mechanism to hold them to

14:46

their promises. Regardless

14:48

of any particular CEO's

14:51

sincerity, however, there

14:53

is a hard limit to reforms

14:55

associated with this type of corporate

14:57

self-governance. However much

14:59

the promises made by the companies

15:01

to better protect privacy or police

15:03

their networks may be genuine, the

15:06

effects will be questionable as long

15:08

as the core business imperative to collect

15:11

it all, all the

15:13

time, remains unchanged. As

15:16

long as social media are propelled

15:18

forward under the regime of surveillance

15:21

capitalism, pledges to do

15:23

better to protect privacy will

15:25

remain little more than window

15:27

dressing, a coat of paint

15:29

to make their platforms more appealing

15:31

and ultimately draw more consumers into

15:34

their fold. Among

15:36

the partial or fragmented solutions

15:38

are the technological fixes. We

15:41

just need a new app to help us

15:43

correct the errors and false information circulating on

15:46

all the other apps. Many

15:48

believe that governments need to enact

15:50

strong data protection regimes with independent

15:53

regulators who have the power and

15:55

authority to punish social media platforms

15:57

that violate the rules of the

15:59

world. those regimes. It

16:02

is common to hear calls

16:04

for more scrutiny of the

16:06

machine-based algorithms companies use to

16:08

sort their users, what's known

16:10

as algorithmic accountability. Proposals

16:13

have been made to legislate

16:15

greater transparency in the social

16:18

media advertising space, particularly around

16:20

political advertising. Some

16:22

believe we should treat social media as

16:25

publishers or regulate them in

16:27

the same way we regulate large

16:29

utilities like electricity and water. Others

16:32

believe they should be broken up

16:34

using antitrust tools instead. One

16:38

potentially helpful way to think of these

16:40

various proposals is as if

16:42

they are ingredients of a long-lost

16:44

recipe. We know the

16:46

basics from memory pass down through

16:48

the generations, but without the

16:51

original formula we hesitate and

16:53

second-guess ourselves. Do I

16:55

finish the pasta in the sauce or pour the

16:57

sauce over the cooked pasta? Do

16:59

I roast the garlic or mince it? The

17:02

term reset is most often associated

17:05

with computers and refers to the

17:07

process of shutting down processes or

17:09

systems that are hanging. It

17:12

can also refer to starting over

17:14

completely from fresh with factory settings.

17:18

A reset provides an opportunity to take

17:20

stock of the big picture. It

17:23

gives us breathing room to evaluate what's

17:25

working and what isn't and

17:28

make adjustments moving forward accordingly. Most

17:31

importantly, it provides us with space and

17:33

time to start over from first principles

17:36

and a solid foundation in

17:38

dispense with those practices that have

17:40

become a hindrance to larger aims.

17:43

There are several compelling reasons to have a

17:46

solid framework to guide us after a

17:48

reset. Having an

17:50

underlying bedrock of principles to

17:52

which we can continuously refer

17:55

helps steer our strategies and

17:57

inform our decisions, especially as

17:59

knowledge is problems arise.

18:02

For at least a decade now, technological

18:04

innovation has been undertaken

18:07

mostly in the absence of any

18:09

such foundation, other than

18:11

a simple imperative to collect

18:13

more data. A

18:16

well articulated set of principles, particularly

18:18

one that links technologies to

18:21

wider political ideals, can

18:23

help remind us that political

18:25

principles should take priority over

18:28

technology, and technology should

18:30

be designed and developed to

18:32

further our political aims, rather

18:34

than work against or be insulated from

18:36

them. It can

18:39

also help us understand the

18:41

relationship between reform proposals that

18:43

otherwise may seem disparate or

18:45

unrelated. It can help us

18:47

evaluate and prioritize, see

18:50

the larger whole of which the

18:52

various fragments are apart. Second,

18:55

having a principled foundation can anchor

18:57

our approach in a rich historical

18:59

tradition, and help us

19:02

feel connected to well-tested and

19:04

long-established wisdom and practical experiments

19:06

on analogous challenges that societies

19:08

have faced in the past.

19:10

Our social media

19:13

universe feels uniquely novel in

19:15

so many ways that it

19:17

may blind us to the

19:19

fact that societies have experienced

19:21

technological upheavals and large-scale societal

19:24

challenges before. Human

19:26

societies have had to adjust and

19:29

adapt throughout their history in the

19:31

face of new material circumstances, much

19:33

like we are experiencing today. We

19:36

can learn from what's come before and

19:39

from the collected wisdom of those who

19:41

have experienced and reflected on it. Third,

19:45

such a foundation helps combat

19:47

fatigue, pessimism, and defeatism among

19:50

critics of social media and

19:52

surveillance capitalism by showing

19:54

there are viable and robust

19:56

alternatives. If we demonstrate

19:59

the common root of numerous

20:01

disparate efforts to detach, reform,

20:03

and regulate social media, we

20:06

show that everyone's efforts are

20:08

weaving something larger than their

20:10

own separate struggles. This

20:13

suggests that the whole is larger

20:15

than the sum of its parts,

20:17

and that there are alliances to

20:19

be forged among like-minded advocates, policymakers,

20:21

and researchers, particularly among

20:24

civil society in different

20:26

jurisdictions worldwide. It

20:29

can help chart a path towards an

20:31

alternative agenda on which groups can work

20:33

together more confidently. This

20:41

is Astra, and there's a lot about this

20:43

idea that I agree with. There

20:45

are a lot of powerful forces I would like to

20:47

be restrained, but I suppose

20:50

there's something about it that

20:52

frames our power as inherently negative,

20:55

that all we can do the best we can hope for

20:58

is to have these powerful

21:00

actors, be they corporations or governments,

21:02

do less harm. So

21:04

it's a kind of negative

21:06

framework, right? Be less invasive,

21:09

be less destructive, be less

21:12

powerful. And I suppose

21:14

I also would want to couple that with

21:16

a framework that thinks about what power

21:18

we have and what other things we

21:21

could do, including with technology.

21:24

So I guess in my

21:26

affirmative vision, I want to think about

21:29

how to redesign our technology and maybe

21:31

into three buckets. So one is

21:35

what technology do we want to democratize? What

21:38

technology do we want to make public

21:40

or socialize? And then what do we

21:42

want to abolish? So when

21:45

I'm thinking about democratizing, well,

21:47

what businesses could be

21:49

run as co-ops, right? Instead of Uber

21:52

as this huge multinational corporation,

21:54

what if there were local

21:56

municipal co-ops of brideshare drivers,

21:59

you know who serve their communities. You

22:01

know, then what platforms do we want

22:03

to socialize or make public? So maybe

22:06

Facebook should be a public utility. This is

22:08

something that Ron Deibert, you know, mentions

22:10

in passing in the book. Yeah,

22:13

and then what things should be abolished? I mean,

22:15

for me, there are certain types of invasive data

22:17

collection that just shouldn't exist. They're

22:20

organizing campaigns around the United States.

22:22

And in some cities, they've been

22:24

successful to just ban facial recognition

22:26

as a tool that the police can use. I think

22:28

that's the right thing. Instead of trying to regulate, there

22:31

are some things that just need to

22:33

be left alone, some data that does not need

22:35

to be collected. So there's an analogy

22:37

there to the environmental justice movement. Some fossil

22:39

fuels need to stay in the ground. Some

22:42

private data just needs to stay off

22:45

the cloud. We should

22:47

go beyond restraint to just the

22:49

framework of abolition of saying, well,

22:51

actually, that's just because it's technically

22:53

possible doesn't mean it's morally

22:55

desirable. On

23:02

ideas, you're listening to an

23:04

encore presentation of retreat, reform,

23:06

restraint, the final installment of

23:08

Ron Deibert's 2020 Massey

23:11

Lectures. You can

23:13

find ideas wherever you get your

23:15

podcasts. And on CBC Radio 1

23:17

in Canada, across North America, on

23:20

US Public Radio and on Sirius

23:22

XM, in Australia,

23:24

on ABC Radio National

23:26

and around the world

23:29

at cbc.ca/ideas. I'm

23:31

Nala Iyad. The

23:33

past is shrouded in mystery. To

23:36

understand it, you have to get

23:38

up calls. Something

23:40

happened to our collective psyche after

23:43

the atom bomb. An

23:47

NPR throughline, we reopen stories from

23:49

the past to find clues to

23:51

the present. Find through

23:53

line wherever you get your podcasts. Part

24:00

of the series marking the sixtieth

24:02

anniversary of Nancy College. One

24:04

of our partners and them and see lectures.

24:09

It features the sixth and

24:11

final Matter of Run The

24:13

Birth Series Reset Reclaiming The

24:15

Internet, First Civil Society. In

24:19

this lecture, he explores the kinds

24:21

of restraints we need to place

24:23

on government and corporations and on

24:25

our own endless appetite for data.

24:32

Part. Three, Restraint. To.

24:35

Get us started! I propose a

24:38

single, simple, but hopefully potent principle.

24:41

Restraint. Restraint.

24:43

Is primarily defined as quote a

24:45

measure or condition the keep someone

24:47

or something under control her within

24:50

limits. Secondarily,

24:52

It also means quote self

24:54

control as in an emotional

24:57

dispassionate are moderate behavior. Both.

25:00

Senses of the term point

25:02

to general qualities that will

25:04

be essential to preserving rights

25:06

and freedoms in our supercharge

25:08

hyper network world of data.

25:11

We. Need to restrain what governments

25:13

and corporations do with all of

25:16

the extraordinarily powerful tools of surveillance

25:18

that are now in their hands.

25:21

We. To restrain with they do with all

25:23

of the data about us and our behaviors.

25:26

Restraints. Will be essential to

25:28

ensure the security of the broader

25:30

information and communication space in which

25:32

we live. Particularly.

25:35

Restraints on bad actors exploiting

25:37

costs for despotic, corrupt, or

25:39

criminal ends. Of government's

25:42

exploiting up for their narrow

25:44

national security aims. Will.

25:47

Need personal restraints to. Restraints.

25:50

On are endless appetite for data.

25:52

Restraints. On our emotions and anger

25:54

as we engage on line in the

25:56

absence of the physical cues that normally

25:59

help Can. them. We

26:01

will need restraints on each other, mutual

26:04

restraints that apply to individuals,

26:07

organizations, and even sovereign states.

26:10

If there is one single mantra,

26:13

one simple concept, that should become

26:15

our slogan and guide us as

26:17

we chart our path forward, I

26:20

believe it should be restraint. While

26:24

most everyone is familiar with the

26:26

concept of restraint, what may be

26:28

less familiar to many is that

26:30

this seemingly simple term is derived

26:32

from and is essential to a

26:35

long tradition of theorizing about political

26:37

liberty and security going back centuries.

26:41

It is most intimately connected to

26:43

that broad family of political thought

26:45

that for most of us is

26:47

so entrenched in our habits and

26:49

dispositions, it is more

26:51

like an instinct than a self-conscious

26:53

philosophy. I'm talking of

26:56

course about liberalism. Broadly

26:59

defined, liberalism is a tradition

27:01

that supports individual rights, civil

27:04

liberties, and political reform that

27:06

pushes societies in the direction

27:08

of individual freedom, democracy, and

27:11

social equality. Political

27:13

theorists will be quick to point out

27:15

that liberalism is not a single theory,

27:18

but a large family of ideas

27:21

and prescriptions for how to manage

27:23

societies that goes back hundreds of

27:25

years. Most of

27:27

us can rhyme off the key

27:29

features of liberalism so ingrained are

27:32

they into our collective approach to

27:34

politics, free and fair

27:36

elections, constitutions, limited

27:39

government, the rule of law, separation

27:42

of powers, pluralism, social

27:44

justice, and protection for human

27:46

rights. There

27:49

are many tangled threads that weave

27:51

their way through liberalism. There

27:53

are also competing schools of thought

27:56

within its large and diverse tent.

27:59

Those who consider the Themselves adherents range

28:01

from Libertarians and free market

28:03

fundamentalists on one end of

28:05

the spectrum. To. Democratic Socialists

28:07

on the other. In.

28:09

Spite of competing schools, however,

28:12

Liberalism. Of all stripes

28:14

shares a fundamental principle. At

28:17

it's core is the belief that

28:19

in order to preserve and maximize

28:21

freedom, While. Countering insecurity and

28:24

fear. We. Must build

28:26

legally binding restraints on those

28:28

we interest to discharge authority.

28:31

Restraints. On the exercise of

28:34

political power. To prevent

28:36

abuse. Within. The

28:38

Liberal Tent The specific school

28:40

of thought that is most

28:42

associated with this principle of

28:44

restraint is known as Republicanism.

28:47

In fact, the idea of applying

28:49

restraint as a design principle for

28:51

political systems is one of the

28:53

most venerable in political theorizing, with

28:56

roots reaching all the way back

28:58

to Ancient Greece. Although

29:01

Republicanism has something to say about

29:03

many principals and it's hard it

29:06

is about preventing the centralization and

29:08

thus abuse of power. For.

29:11

Republicans unchecked concentrations of power

29:13

threatened liberty and security because

29:15

they are apt to be

29:18

abused. And. So Checks and

29:20

Balances A widely known phrase that

29:22

comes from Republican thought. Must.

29:25

Be institutionalized to distribute power

29:27

pluralistic lee and keep it

29:29

that way. The.

29:32

Republican opposition to concentration of

29:34

power rests on assumptions of

29:37

human frailty. Humans

29:39

tend to be both self interested

29:41

in prone to lapses in judgment.

29:44

When. Opportunities present themselves. They are

29:46

tempted by power which can in

29:48

turn bring about corruption and other

29:51

abuses. As. Montesquieu famously

29:53

observed. Every man invested

29:56

with power is up to abuse

29:58

It. Republic. American

30:00

theorists were acutely aware of the

30:02

accumulation of unchecked and oppressive power

30:04

in the hands of government as

30:07

a threat to individual liberty and

30:09

security, and so they

30:11

devised an elaborate system of power-restraint

30:13

devices to thwart it. These

30:16

devices are a form of

30:18

friction introduced into political processes

30:20

to make the exercise of

30:23

authority less efficient. Strange

30:26

as it may sound, at a

30:28

time of near-total surveillance potential in

30:30

the hands of government agencies, we

30:32

need to challenge ourselves to think

30:35

of ways to artificially reduce the

30:37

efficiency of our government's security agencies.

30:41

The Republican attention to material factors

30:43

tells us why. Oceans,

30:46

mountains, and other rugged terrain,

30:48

as well as inhospitable climates

30:51

and even endemic diseases, created

30:54

obstacles for invading armies,

30:56

slowing them down, impeding

30:59

conquest, and protecting long-term

31:01

control of populations, a

31:03

kind of accidental friction by

31:06

circumstance. All

31:08

things being equal, the further away

31:10

people are from the center of

31:12

control, or the more

31:14

natural barriers shelter them, the less

31:16

efficient the exercise of that control

31:18

tends to be. Once

31:21

in times prior to social media and the

31:23

internet, activists could flee their

31:26

home countries out of fear of

31:28

persecution and feel safe, thousands of

31:30

miles away from the reach of

31:32

the power elites they left behind.

31:35

However, with social media and other

31:37

digital technologies, these natural

31:40

barriers have been almost

31:42

entirely dissolved. We

31:44

now live in something approximating

31:47

a friction-free environment in which

31:49

outside forces, being they companies

31:52

or governments, can pry

31:54

into our most intimate details, even

31:56

when we're behind closed doors. new

32:00

technologies, all of us

32:02

can be tracked to a degree

32:05

and in a fashion that is

32:07

both unprecedented in human history and

32:09

nearly comprehensive in its potential totality.

32:12

When our fridges, baby monitors,

32:15

video conferencing facilities, smart TVs,

32:17

and even brains are all

32:19

networked to the outside world,

32:22

natural restraints that we once took for

32:24

granted no longer serve the

32:26

same function. We

32:30

now face an entirely new challenge

32:32

from the material context, thanks

32:34

to the changing nature of technology. Implanted

32:38

technologies have the potential to

32:40

pinpoint details even down to

32:42

a biological level with a

32:44

precision that borders on precognition.

32:47

This great leap forward in

32:49

remote control raises the prospect

32:52

of severe abuse of power

32:54

and near totalitarian control. The

32:57

application of restraint measures to the

32:59

design and functioning of both private

33:02

and public sectors will thus be

33:04

critical to preserving liberty and security.

33:12

This is Daniel Doudney. One

33:14

way to think about why restraint in

33:16

this way is so important

33:19

is to just recognize the

33:21

extent to which civilization is

33:24

a series of restraints that

33:26

have over time emerged to

33:29

solve various problems that have

33:31

arisen from human

33:34

exploitations of technology. We

33:37

can think of it this

33:39

way that over the longer

33:41

term, there's a cornucopia of

33:43

increasingly potent double-edged technological swords.

33:47

For example, advances which are

33:49

coming at ever-increasing rates.

33:52

When one looks at the longer horizon

33:55

of human history, there's been

33:57

this last century or so which has just

33:59

been explosive. with

34:01

regard to technological capabilities.

34:04

And in all cases, we're

34:06

basically getting human enablement, a

34:09

science-based technology is enabling humans

34:11

to do more things. Many

34:15

of the things that it enables are

34:17

things we want, but many

34:19

of the things that it enables are things

34:21

we don't want. And

34:23

our capacity to continue

34:25

to advance, to

34:28

continue to make technology serve us

34:31

rather than technology is a

34:33

source of disasters

34:35

and oppressions, depends

34:38

upon how we regulate

34:40

and restrain the technologies.

34:44

In every aspect of our life, we've

34:46

got restraints on technology, and we take

34:48

them for granted. They're like baked in.

34:51

Think about the automobile. The first

34:53

automobiles did not have mufflers. And

34:56

so urban areas were quickly

34:59

overwhelmed by enormous amounts

35:01

of noise. And so

35:03

citizen groups got together and demanded

35:06

regulation. And of course,

35:08

the nascent automobile industry said,

35:11

oh, this is an impingement on our freedom.

35:13

But over time, the

35:15

automobile manufacturers started putting

35:19

mufflers on automobiles

35:21

in the factory. And

35:23

the problem just went away. And

35:25

no one thinks about it anymore.

35:27

We don't have to organize ourselves

35:29

to control urban automobile noise. So

35:33

when we say technological progress,

35:35

what we mean is not

35:37

just that technology can

35:39

do stuff we want, it

35:42

also means that we have the

35:44

ability to make sure the

35:46

technology doesn't do stuff we don't want.

35:50

Part 4, mechanisms for

35:52

change. The

35:58

first place to start is by reading. reviewing the

36:00

type and effectiveness of existing

36:03

restraint mechanisms around governments. While

36:06

companies can abuse power too and

36:08

collect a lot of sensitive, fine-grained,

36:10

and highly revealing data about us,

36:13

which they can in turn share

36:15

with governments, only a

36:17

government can take away a person's liberty

36:19

by force. The

36:21

security arms of the state, the

36:24

police, armed forces, and

36:26

other security agencies have a

36:28

monopoly on violence, one of

36:30

the definitions of sovereign statehood.

36:33

They have lethal means at their

36:35

disposal, can arrest people and lock

36:37

them up, and in some

36:40

jurisdictions can even end their lives

36:42

through capital punishment. In

36:44

response to emergencies, governments can

36:47

also take all sorts of

36:49

exceptional measures that infringe on

36:52

liberties, including declaring martial law,

36:54

or simply suspending constitutionally protected

36:57

rights that we take for granted as

36:59

we discovered with the COVID pandemic. The

37:02

first task of our reset should

37:05

be to evaluate the effectiveness of

37:07

the restraint mechanisms we have inherited.

37:10

Do we need to supplement them

37:13

with new resources, capabilities, and

37:15

authorities? One

37:17

simple rule of thumb that may help

37:19

guide us is as follows. Restraint

37:22

should increase proportionately to the

37:25

intrusiveness of the practices in

37:28

question. The

37:30

more invasive a technology is, the

37:32

more likely it lends itself to abusive

37:35

power, and so the stronger and

37:37

more elaborate the restraint should be. Consider

37:41

location tracking via cellular and

37:43

telecommunications data. Most

37:46

everyone carries around a network device

37:48

with them all the time that

37:50

pings cell towers and local telco

37:53

networks on a continuous basis, and

37:55

is standard outfitted with GPS and

37:57

Bluetooth beacons. Most

38:00

all of us have dozens of apps

38:02

that routinely grab location history and data

38:04

too, and use them primarily

38:07

for advertising purposes. Prior

38:10

to COVID-19, these data could be

38:12

accessed by law enforcement, military and

38:15

intelligence agencies, but under

38:17

widely different conditions. In

38:20

some countries, certain of those agencies might

38:22

require a warrant or a production order,

38:25

while in others they might simply walk

38:27

into the headquarters of telecommunications companies and

38:29

demand them. In

38:32

a world where we reflexively look

38:34

to big tech for solutions, it's

38:37

not surprising that Google and Apple

38:39

have teamed up to develop a

38:41

protocol for anonymized contact tracing through

38:44

smartphone apps. Like

38:46

others who have already weighed in on the

38:48

issue, I believe that, however

38:50

much these prove to be useful

38:52

in public health emergencies, the

38:55

safeguards around them must be

38:57

exceptionally strong too. Some

39:00

basic restraint mechanisms should include

39:02

strict limits on data retention,

39:05

clear limitations on use, and

39:07

restrictions on access to ensure

39:09

that the data are not

39:11

illegitimately redeployed for other reasons,

39:14

like catching chicken wing thieves

39:16

and jaywalkers or monitoring

39:18

human rights defenders. Similarly,

39:21

strong restraints should be applied

39:23

to the use of commercial

39:25

spyware and hacking tools by

39:27

government security agencies, which

39:30

are among the most intrusive and,

39:32

as we at Citizen Lab have

39:34

demonstrated in our research, highly prone

39:36

to abuse. States

39:39

purchasing spyware are at liberty to

39:41

abuse it with limited or no

39:43

transparency or regulation.

39:46

Companies that manufacture and sell it have

39:49

unbridled freedom to rake in

39:51

revenues by the tens of

39:53

millions, largely without fear

39:55

of criminal liability or concern for

39:57

how their technology impacts human rights.

40:00

rights, the net result,

40:02

harassment, blackmail, and

40:05

even murder of countless innocent

40:07

civilians worldwide. Abuses

40:10

and built-in discrimination around the use of

40:12

some of these technologies today, in

40:15

policing, immigration, and criminal justice

40:17

practices, are already well

40:20

documented. The prospects

40:22

for even greater harms down the

40:24

road are impossibly large and daunting

40:26

to contemplate. Having

40:29

restraints to what governments can do is only

40:32

part of the solution. We

40:34

live in an age in

40:36

which gigantic corporations, and especially

40:38

technology giants, dominate the

40:41

social and political landscape. The

40:44

powers of unbridled surveillance

40:46

capitalism are truly awesome,

40:49

and when combined with state

40:51

authority, are potentially totalitarian. Keeping

40:54

our desires to persuade us to

40:56

consume this or that product is

40:59

disturbing enough, but the

41:01

prospect of corporations and states

41:03

colluding in broader population control

41:05

is downright dystopian.

41:08

Just ask a Tibetan or Uyghur. In

41:12

addition to the risks of abuse

41:14

of power related to fine-grained remote

41:16

control technologies, there is

41:18

another reason to impose restraints on

41:21

surveillance capitalism. The

41:23

engine at the heart of the business model,

41:26

which prejudices sensational, extreme,

41:28

and emotional content, amplifies

41:31

our baser instincts, creates

41:34

irresistible opportunities for

41:36

malfeasance, and helps pollute

41:38

the public sphere. It

41:41

also continuously accelerates our consumption

41:43

of data. More

41:45

is better, faster too. But

41:48

endlessly accelerating consumption of data on

41:50

the part of both consumers and

41:53

firms mining our behavior, taxes

41:55

the planet's finite resources, draws

41:58

volumes of fossil fuel

42:00

power energy and contributes

42:02

to one of humanity's most

42:04

pressing existential risks. It

42:06

certainly doesn't solve it. Introducing

42:09

friction and other restraints on

42:11

surveillance capitalism can help improve

42:14

the quality of our public

42:16

discourse while tempering the insatiable

42:18

hunger for more data, faster

42:20

networks and disposable gadgets. One

42:23

thing is for sure, business as

42:25

usual can no longer be tolerated.

42:28

Problems that have major effects on

42:31

the public's sharing and consumption of

42:33

information should not happen in the

42:35

shadows or behind a proprietary algorithm.

42:38

To be sure, there are balances

42:40

to be struck around free expression,

42:42

content moderation and abuse prevention. There

42:46

are very real risks that

42:48

mandatory or poorly constructed measures

42:51

could be perverted as an

42:53

instrument of authoritarian control abused

42:56

by despots and autocrats to enforce

42:58

their rule. The

43:01

key will be to make sure that

43:03

social media platforms manage content in ways

43:05

that are transparent, limited,

43:08

proportional and in

43:10

compliance with internationally recognized human

43:12

rights. These

43:14

standards may not be possible in all

43:16

jurisdictions right now, but they

43:18

should be an imperative for

43:20

those that consider themselves liberal

43:22

democracies. I

43:25

think this is a great place to

43:28

start examining how we might push back

43:30

on these systems. I

43:33

am also someone who believes

43:35

we need to examine the

43:37

roots of liberalism with critique,

43:41

understand that small are republicanism, at

43:44

least in the US context where I'm from was

43:47

concomitant with settler colonialism and

43:49

slavery and some of these

43:51

racial projects that are still with us today

43:53

and be very

43:55

discerning about which of

43:58

those principles we carry over and

44:00

which of those we reinvent. I

44:03

think there is also an omnipresent

44:06

question here about who

44:09

the we is, who gets

44:11

to determine what a limitation is, what

44:13

transparency is, how to interpret a

44:15

human rights principle. And

44:18

this is something that I think points

44:20

to the requirements for robust

44:23

social movements and

44:25

small d, democracy, that

44:28

can continue to foment and

44:30

push back against the urge

44:32

for again, that small council

44:34

of elders with a view from

44:37

nowhere, making these determinations on behalf

44:39

of everyone else. It's

44:41

John Norton here. I think that Ron is right that we

44:43

need to have a comprehensive approach

44:45

to this. And one of the

44:47

great things about the lectures, I

44:50

think is that they

44:52

constitute one of the first attempts I've

44:54

seen to try and convey the comprehensiveness

44:57

of the problem that we face as

44:59

a whole. But if we go back

45:01

to social media, what they

45:03

have done is they

45:05

have transformed humanity's media ecosystem.

45:09

Now the word media is interesting.

45:11

It's the plural of medium. And

45:14

medium has actually two meanings. One

45:16

is the conventional meaning. That's to say a

45:19

medium is a communication channel. But

45:21

to a biologist, a medium is

45:23

something quite different. It's

45:25

a mixture of substances in

45:28

a Petri dish in which organisms

45:30

grow. And if

45:32

you change the medium as a biologist, then

45:35

different kinds of organisms grow. And

45:39

our information ecosystem, you could think of

45:41

it as a medium in which human

45:43

culture grows. We've changed

45:45

that medium. We've changed the medium

45:47

in our global Petri dish. And

45:50

therefore we shouldn't be surprised that some strange organisms

45:52

are now thriving in it. Organisms

45:54

that in the end might be really toxic,

45:57

for example, for democracy. Iran

46:00

is right. We need to think

46:02

on that kind of global scale. Part

46:09

5. The Road Forward

46:13

Where global governance in general goes,

46:15

so too does governance of the

46:17

internet and social media. Discussions

46:21

to develop norms of appropriate

46:23

state behavior in cyberspace, while

46:26

laudable on one level, seem

46:28

entirely theoretical at the current

46:30

time against the practical reality

46:33

of massive investments by states

46:35

in offensive hacking, super power

46:38

policing, mass surveillance and influence

46:40

operations. We should

46:42

not expect international institutions to

46:45

be anything other than reflections

46:47

of self-interested and power-seeking sovereign

46:49

states, as long as

46:51

restraints on the abuse of power

46:53

are not deeply entrenched in domestic

46:55

spheres. Only once

46:58

restraints, divisions and separations are

47:00

established in individual republics can

47:03

they then be extended internationally,

47:05

first to other like-minded states

47:07

and then gradually to others.

47:11

Liberal democratic systems of government

47:13

can ensure that social media

47:15

platforms and other technology giants

47:18

are subjected to common standards of

47:20

governance so that they cannot play

47:22

jurisdictions against each other. Only

47:26

with such a united front can they

47:28

develop a truly robust response to

47:31

the state-centric model of social

47:33

media and internet governance being

47:35

propagated by authoritarian countries like

47:37

Russia and China. The

47:41

painful truths outlined in these talks

47:43

paint a very bleak picture. They

47:47

also present a troubling forecast for the

47:49

future of the human condition. It

47:53

seems undeniable now that the

47:55

disturbing worldwide descent into

47:58

neo-fascism, tribal tactics, unbridled

48:01

kleptocracy, along with

48:03

the accompanying spread of ignorance and

48:05

prejudice we have witnessed in recent

48:07

years, is at

48:10

least in part because the social media

48:12

environment, presently constituted under

48:14

the regime of surveillance capitalism,

48:17

created conditions that allowed such

48:19

practices to thrive and flourish.

48:23

Personal data surveillance and authoritarian

48:25

state controls present a perfect

48:27

fit. Seemingly

48:30

endless lucrative business opportunities

48:33

that undermine public accountability

48:35

and facilitate despotic rule.

48:38

These negative externalities may be amplified

48:40

by the surge in demand for

48:42

social media during the COVID pandemic,

48:45

the enhanced power of the platforms that went

48:47

along with it, and the

48:49

unprecedented emergency measures that tapped

48:52

into those platforms surveillance potential.

48:55

On top, our insatiable lust

48:57

for data and disposable devices

48:59

is silently taxing resources, sucking

49:01

up vast amounts of energy

49:03

and thus contributing to, rather

49:06

than helping to mitigate, the

49:08

climate crisis. While

49:10

the COVID pandemic has given some

49:12

reprieve to carbon dioxide emissions, thanks

49:15

to a short reduction in

49:17

airline and other fossil fuel

49:19

powered transportation, that reprieve

49:21

will eventually pass. However,

49:24

the sudden embrace of digital

49:26

network technologies will not and

49:29

may in fact deepen. Combined,

49:31

both real and virtual

49:34

consumption could increasingly strain

49:36

natural resources, draw

49:38

from dirty energy sources, drive up

49:40

emissions and contribute to waste. As

49:44

climate scientists warn, continuing down

49:46

that path will lead to

49:48

collective ruin. Our

49:50

precious apps will mean little when

49:52

humans are either wiped out altogether

49:54

or consigned to a Hobbesian state

49:56

of nature, dispersed in small tribes, strays,

49:59

and forests. We're going against each

50:01

other for the scarce resources needed

50:03

for survival. It. Is

50:05

not unrealistic to imagine a time.

50:08

When. The Internet and all

50:10

of it's associated infrastructure. Will.

50:12

Be reduced to rusting artifacts.

50:15

Submersed and rising oceans are

50:17

covered over by tangled weeds.

50:20

If consumption practices continue

50:22

apace, That. Will be

50:24

one way to mitigate social medias

50:27

negative consequences but obviously not worth

50:29

the price. Much.

50:31

Of what I've been talking about in

50:34

this series is sadly not surprising. These.

50:37

Are Truths because they are

50:39

widely recognized by a growing

50:41

community of experts. But.

50:44

The time is now come to move

50:46

beyond diagnosis and start the hard work

50:48

on solutions. We. Must squarely

50:51

and comprehensively address the intertwined

50:53

pathologies of social media and

50:55

surveillance capitalism. Starting.

50:57

With that device you hold in your hand. Fortunately,

51:01

We have a recipe the set of

51:03

principles that can help guide us on

51:05

this task. We. Do you not

51:07

even need to invent something new? Humans.

51:10

Have faced enormous political challenges

51:12

thrown up by new material

51:15

situations before. And there

51:17

is a long tradition of practical

51:19

theorizing that can be adapted to

51:21

our own unique circumstances. It.

51:24

Is time to reset. To. Start

51:26

over from first principles and

51:28

to work towards the construction

51:30

and stewardship. Of a

51:32

communications ecosystem that serves rather

51:35

than diminishes human well being.

51:38

We. Need to realistically of that,

51:40

there are major hurdles to

51:42

overcome and deeply entrenched and

51:44

powerful interests that will work

51:46

and opposition. And not always by

51:48

the rules. A comprehensive

51:51

strategy of long term reform

51:53

is therefore required. Extending.

51:55

From the personal to the political. From.

51:58

The local to the global. We.

52:01

Must begin now with practical

52:03

and manageable small steps simultaneously

52:05

undertaken by many of us

52:07

spread across the planet. The.

52:10

Covert Emergency reminds us of

52:12

our shared fate. We.

52:15

Have a once in a lifetime opportunity.

52:17

To. Reset. We

52:19

can reclaim the internet for

52:21

civil society. The principle

52:24

of restraint should be our God.

52:45

Given special encore selection from

52:47

Run Be Birds Twenty Twenty

52:50

Massey Lectures called Reset Reclaiming

52:52

the Internet for Civil Society.

52:59

Along the way you also heard from

53:01

Astra Taylor. Ten since

53:03

our Meredith Whitaker, Misha, Glenny,

53:05

Daniel Dewdney, and John Not

53:08

in. That

53:11

series has produced for Ideas

53:13

by Philip Culture. This episode

53:15

was produced by Pauline Holdsworth

53:17

with production assistants from Any

53:19

Bender. It's

53:23

part of a series of conversations

53:25

and archive lectures we put together

53:27

to mark the sixtieth anniversary of

53:29

Massey College. One of our partners

53:32

in the Massey electors. Thanks

53:34

to Massey College and former

53:36

principal Natalie, the opposing. Ideas

53:42

as a podcast and a broadcast,

53:44

so she liked the episode you

53:46

just heard. Check out our vast

53:49

archive at Cbc.see A/ideas where you

53:51

can find more than three hundred.

53:53

or past so Technical

54:00

Production, Danielle Duval. Our

54:03

web producer is Lisa Ayoosso. Acting

54:06

Senior Producer, Lisa Godfrey. Greg

54:08

Kelly is the Executive Producer of

54:11

Ideas. And I'm Nala Aied.

55:00

Go to cbc.ca/podcasts.

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features