Podchaser Logo
Home
AI Safety with Shazeda Ahmed

AI Safety with Shazeda Ahmed

Released Tuesday, 9th April 2024
Good episode? Give it some love!
AI Safety with Shazeda Ahmed

AI Safety with Shazeda Ahmed

AI Safety with Shazeda Ahmed

AI Safety with Shazeda Ahmed

Tuesday, 9th April 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:25

David, the U. S. State Department recently

0:26

commissioned a report about how A.

0:32

I. employees feel about the safety of their

0:32

work and the incentives driving it.

0:38

And let's just say the report

0:38

was not particularly positive.

0:41

It turns out that a lot of people

0:41

working in the field of AI have a

0:47

lot of concerns about its safety.

0:50

Yeah, and the fact that it's coming

0:50

from people who work in this industry, I

0:54

think is, particularly telling indicator

0:54

of the fact that the general population,

0:59

ourselves included, don't really know

0:59

the full extent of the risks that are

1:05

associated with these new technologies. It's similar to, all the people who work

1:07

for places like Facebook that don't allow

1:12

their children on social media because

1:12

they know just how addictive it is.

1:17

So the fact that the people with

1:17

knowledge of the inside are the

1:20

ones that are, raising the red flag

1:20

should be a sign that we should pay

1:24

attention to this a little bit more.

1:26

Yeah, the authors of the

1:26

report that I mentioned spoke with

1:28

over 200 experts for it, including

1:28

employees at OpenAI, Google,

1:34

DeepMind, Meta, and Anthropic.

1:36

And these are AI labs that are

1:36

working towards what's known as

1:39

artificial general intelligence. Now the term AI is a very odd one,

1:41

like artificial intelligence, but

1:47

the basic idea is that techies are

1:47

working to create a simulation of the

1:54

intelligence that humans recognize

1:54

as such, which is like mostly

1:57

cognition, but also imagination, right?

1:59

So the development of Dall-E was not just

1:59

to stiff artists and have AI become like

2:07

the new creator of images, but actually,

2:07

I heard the head of frontiers research

2:12

at OpenAI, Mark Chen speak about in 2022.

2:17

He said that the idea was that In

2:17

order to have an artificial general

2:20

intelligence, you also have to

2:20

have image making capabilities.

2:23

So basically, even if artificial

2:23

intelligence is a misnomer, the

2:26

idea behind it, or the idea behind

2:26

artificial general intelligence, is

2:30

that we're trying to simulate the

2:30

full range of what humans recognize

2:36

to be intelligence, and in simulating

2:36

it, find ways to make it even better.

2:43

And so the reason I hesitate in talking

2:43

about this a little bit is because what

2:46

humans consider to be intelligence,

2:46

I think, is a pretty open question.

2:50

The fact that it's only recently

2:50

that those in the AI community have

2:53

considered image generation to be

2:53

a big part of it is pretty telling.

2:56

A lot of our intelligence notions have

2:56

to do with cognition specifically,

3:01

but also this, idea behind, what's

3:01

the point of it, I think, is like

3:06

something I have no good answers for,

3:06

but I have in the back of my mind.

3:11

Yeah, and I think one of the questions that we should be asking about a lot of this work

3:12

in artificial intelligence is what

3:17

definition of intelligence are they

3:17

working with to begin with that

3:20

they're trying to replicate, right? Because often a lot of the people in this

3:21

area are computer scientists, engineers.

3:28

People who define intelligence largely

3:28

just as information processing.

3:32

And when you adopt a broader

3:32

interpretation of intelligence that

3:35

includes, let's say, social intelligence,

3:35

emotional intelligence, moral thinking,

3:41

it starts getting more difficult to

3:41

know whether that's something that

3:44

you can replicate in a machine.

3:47

Also, one thing that I want to

3:47

mention in this context is that

3:50

it's not just that they're trying

3:50

to replicate human cognitive

3:52

capacities, as you mentioned, Ellie. Is that especially when you're thinking

3:54

about people who are closer to the

3:57

transhumanist community, they believe

3:57

that they are going to replicate

4:02

human intelligence in a machine. And once that intelligence is

4:05

created, we will merge with it

4:10

through computer-human interfaces such

4:10

that a new super intelligence will

4:15

emerge that is neither exclusively

4:15

human nor exclusively machinic.

4:20

And this idea of a possible

4:20

future where humans and machines fuse

4:25

is both for some a doomsday scenario

4:25

and for others a utopian scenario.

4:32

The idea is that machines won't be working

4:32

for us or, us working for machines.

4:39

But it's not even that we'll be

4:39

working together, it's rather

4:42

that like we will be one, right?

4:47

Which I think is attractive in certain

4:47

respects because I think a lot of

4:50

people's concerns around AI safety

4:50

have to do with this idea that in the

4:54

future we're going to have AI overlords.

4:57

But I think, for others risks losing

4:57

what makes humans or like what makes the

5:06

organic and especially once you bring

5:06

in the realities of the late capitalist

5:11

world that we live under even if you

5:11

think that in principle humans becoming

5:16

machinic and machines becoming human

5:16

like such that we ultimately merge

5:22

could be ideal or could be a good thing

5:22

I think many of us are concerned that

5:27

we'll end up getting co opted by a

5:27

profoundly exploitative economic order.

5:32

Henry

5:33

No, and I love how the same

5:33

scenario is literally utopia for one

5:37

group and dystopia for another group.

5:39

And it's interesting to think that

5:39

distinction between utopia and dystopia

5:43

might map onto the distinction between

5:43

employee and boss in contemporary

5:49

technology, because you mentioned,

5:49

that a lot of the employees who work

5:52

in tech today have a lot of worries

5:52

about the safety and the future of AI.

5:57

But that's very different than the

5:57

attitude that we find in what I sometimes

6:01

call the golden boys of contemporary tech,

6:01

like the Ray Kurzweil of the world, the

6:07

Elon Musk, so on and so forth, right?

6:09

For them, the bosses, it

6:09

really represents a utopia.

6:14

And I think there is no better

6:14

example of this than precisely Ray

6:17

Kurzweil's book, The Singularity is

6:17

Near, which came out a couple of years

6:21

ago, and it's a behemoth of a book. It's about 650 pages of him talking

6:23

about this utopian moment that is going

6:30

to happen in the near future that he

6:30

calls the singularity, which is the

6:35

moment when artificial intelligence will

6:35

not only surpass human intelligence,

6:39

but will also merge with it.

6:42

And he argues that at that point, we

6:42

will become a kind of super organism

6:46

that is nothing but disembodied

6:46

information and knowledge in like

6:50

a purified, crystallized form.

6:55

Kurzweil's position has been

6:55

attractive for a lot of people, but also

7:00

easy to parody because he does read as so

7:00

utopian about the potential of technology.

7:07

There are, a lot of people who have been

7:07

voicing significant concerns recently

7:13

about artificial general intelligence

7:13

specifically, and I think that is,

7:18

where we want to begin thinking about

7:18

the topic for this episode, which is

7:23

ultimately going to be less about like

7:23

actually what recommendations we have for

7:26

AI safety, not like you and me, we, but

7:26

even like the actual community, right?

7:31

Our guest today is an expert on

7:31

the character of the AI safety

7:37

communities rather than on if we do

7:37

X, Y, Z, we will be safe from AI.

7:42

But I think a few of the concerns that

7:42

are worth pointing out come from the

7:46

philosopher Nick Bostrom, who wrote

7:46

a really influential 2014 book called

7:50

Superintelligence that ended up having

7:50

a big influence on Elon Musk and a

7:55

bunch of other people in the tech world. And in this book, Bostrom suggests

7:57

that the rise of artificial general

8:02

intelligence can cause some significant

8:02

issues, including displacement of the

8:07

workforce by artificial intelligence,

8:07

political and military manipulation,

8:13

and even possibly human extinction.

8:16

And I think this threat of human

8:16

extinction that Bostrom articulates in

8:20

the book has been one of the main drivers

8:20

of recent concerns about AI safety, in

8:26

addition to the fact that our AI has

8:26

just improved significantly since Bostrom

8:30

published this book ten years ago.

8:32

Yeah, and Bostrom has this

8:32

really fun kind of thought experiment

8:35

to highlight some of these high end,

8:35

even if low probability, doom scenarios.

8:40

A fun thought experiment

8:40

about human extinction.

8:43

about human extinction.

8:44

So philosopher of him.

8:46

Yeah, just like turning everything

8:46

into an interesting theoretical point.

8:50

No, but it's called the

8:50

paperclip maximizer.

8:54

So he says, imagine that we create

8:54

an intelligent machine that has one

8:58

specific goal, and that goal can be

8:58

very benign, like making paperclips.

9:03

The problem is that we humans have

9:03

no way of guaranteeing that we

9:07

can predict exactly how that ends. algorithm or that machine will

9:09

go about achieving that end.

9:13

So it could be that the machine

9:13

decides, using its general intelligence

9:19

that it's been equipped with by its

9:19

human creators, that it needs to

9:23

use any possible resource in order

9:23

to create more and more paperclips.

9:28

And then it starts literally

9:28

killing human life in order to

9:32

produce more and more paperclips. And so here, it's not as if human

9:34

extinction would come about from an

9:40

algorithm that turns evil and wants

9:40

to dominate us, it would actually

9:44

just be something much more mechanical

9:44

and mundane that we can't control.

9:50

And so he's trying to make some

9:50

of the risks of AI more palpable

9:54

to a general audience by making

9:54

them seem more realistic.

9:58

By talking about

9:58

death by paperclips.

10:00

Yeah, it's not that they become diabolical. It's just that they are machines

10:02

and they are indifferent to us.

10:05

That's the point.

10:06

I feel like this thought

10:06

experiment is what you would get if

10:08

you asked Kafka to write office space. Let's talk with our expert guest

10:13

today who has far more intelligent

10:17

things to say about AI safety than

10:17

either of us could do justice to.

10:24

Today, we are

10:24

talking about AI safety.

10:27

How have philosophies stemming

10:27

from utilitarianism become so dominant in

10:31

discussions of AI safety in today's world?

10:34

What are the most salient

10:34

risks associated with AI today?

10:38

And how has the rise of

10:38

funding and interest in AI safety

10:42

come to shape today's understanding

10:42

of what the risks posed are?

10:48

Shazeda Ahmed is Chancellor's

10:48

Postdoctoral Fellow at UCLA's

10:52

Institute of American Cultures. A specialist in AI safety, AI

10:54

ethics, and technology in China, Dr.

10:59

Ahmed has held positions at

10:59

Princeton University, Citizen

11:02

Lab, and the AI Now Institute. In addition to her peer reviewed work, Dr.

11:07

Ahmed's research has been featured

11:07

in outlets including Wired,

11:10

Financial Times, The Verge, and CNBC.

11:14

Dr. Ahmed, welcome to Overthink.

11:16

We are so excited to have you.

11:18

Oh?

11:19

And there is so much

11:19

demand for us to have an expert

11:22

on AI and ethics and safety.

11:25

So there are a lot of people out

11:25

there who are our fans who are just

11:28

dying to hear this conversation. So thank you for your making the time.

11:31

Thank you so much for the invitation.

11:33

Let's jump in by thinking about

11:33

the recent explosion of interest in

11:37

AI, which has happened especially

11:37

in the wake of the recent successes

11:41

of LLMs, large language models.

11:43

I'm here thinking about especially

11:43

ChatGPT, but also other models.

11:48

And it makes sense that a lot of

11:48

people are posing questions about

11:52

the safety of contemporary AI.

11:56

But I think that when the average person

11:56

thinks about AI and safety together, they

12:01

often conjure up images of a dystopic

12:01

techno future, drawn from maybe their

12:07

favorite sci fi movie, where an algorithm

12:07

or a machine sort of malfunctions, runs

12:13

amok, and then, subjects us to its will.

12:18

And I think that most contemporary

12:18

discussions of people who work in this

12:23

field don't really take that form, right?

12:25

It's not about those scenarios. So can you tell us just by way of

12:27

beginning what people working on AI

12:32

and safety today are actually concerned

12:32

about when it comes to modern AI?

12:38

There are multiple ways of

12:38

approaching questions around AI and

12:41

safety outside of what the community

12:41

that calls itself AI safety works on.

12:45

So for many years before such a

12:45

community existed and used that

12:49

name for itself, you had people

12:49

working on issues like bias, right?

12:52

If you use a hiring algorithm in

12:52

your software that's screening job

12:57

candidates and it's trained on CVS that

12:57

are mostly white men who went to Ivy

13:02

League schools, what will that mean for

13:02

candidates who don't fit that description?

13:05

Bias, discrimination, false negatives

13:05

and positives and associating

13:09

certain characteristics with people.

13:12

The kind of classic real world examples

13:12

are face recognition systems that

13:16

are used for policing that recognize

13:16

the wrong person and can lead to

13:20

the wrongful arrest of that person. So those are the issues that had

13:22

been percolating and people have

13:25

been talking about for many years. But when it comes to these kinds

13:27

of more dystopic examples that you

13:31

brought up, that is very specifically

13:31

related to a community of people who

13:35

have come together around a series

13:35

of other interlinked movements.

13:38

So effective altruism, long

13:38

termism, existential risk, this

13:42

kind of group of people had been

13:42

thinking about how to promote human

13:46

flourishing centuries from today.

13:49

And they try to think on that

13:49

kind of long term horizon.

13:52

And thinking about that, they also had

13:52

to contemplate the possibilities of risks

13:56

to that flourishing not being possible.

13:59

And they've created a lot of hypothetical

13:59

scenarios about the possibility

14:02

that we could end up with artificial

14:02

intelligence that is smarter than humans.

14:06

And again, they would call that

14:06

artificial general intelligence or AGI.

14:10

And there's competing definitions

14:10

about what that entails.

14:14

when it comes from a company like

14:14

OpenAI, Sam Altman will repeatedly

14:17

say AGI is artificial intelligence

14:17

systems that outperform humans at

14:22

every economically valuable task.

14:24

A lot of the recent hype, a lot

14:24

of the recent headlines people are

14:28

seeing and parroting back about the

14:28

possibility that AI could kill us

14:31

all, it comes out of that community

14:31

and a lot of these speculative fears.

14:36

In the moment ChatGPT blew up and people

14:36

were really astounded by its performance

14:40

and the text outputs being in some cases

14:40

indistinguishable from something a human

14:45

being might write, you had a lot of

14:45

people who were on the fringes of those

14:49

communities and were skeptical suddenly

14:49

becoming very concerned about AI's risk.

14:54

But that is the main thing that happened, right? It's the ChatGPT was widely

14:56

released to the public and people

14:59

got to see its performance. In my opinion, that is a

15:01

marketing stunt, right?

15:03

What better way to get people to want to

15:03

use your product than to let them use it.

15:07

And also to get tons of free

15:07

feedback on the ways they're going

15:10

to try to break it and misuse it. So a lot of what my research, when I

15:12

was at Princeton working with four of

15:17

my colleagues, looking at how these

15:17

ideas came, to spread as quickly as they

15:22

did, what are their epistemic claims?

15:24

what does it mean to produce

15:24

knowledge on something so speculative?

15:28

And some of the issues that we saw

15:28

coming out of the communities of people

15:32

interested in this, they talk about

15:32

a thing called the alignment problem.

15:35

So they say we can live in a

15:35

utopic future with AGI if we

15:40

can align AGI with human values.

15:43

And that is a whole process. It's eliciting what those human values

15:44

are, figuring out how to encode them

15:48

into AI systems, testing them to

15:48

ensure that they would be aligned

15:52

under a variety of circumstances. And so the fear of the kind of scenarios

15:54

of AI posing a threat to the future of

16:00

human existence or an existential or X

16:00

risk comes from misalignment and there

16:05

are tons of people working on alignment. We did a whole study with people asking

16:07

what it even means to work on alignment.

16:11

if you really think this is the way, what

16:11

are you doing in your work every day and

16:14

how do you know you're getting there? And this is, so I hope this

16:16

unpacks a little bit of like,

16:19

why is it that policymakers are

16:19

suddenly thinking about this?

16:22

Why is it that this is in the news so

16:22

often, our research showed that this

16:27

community around these ideas, it's not.

16:30

It wasn't that big when we started

16:30

doing this research in 2022, but it's

16:33

very well funded through a series

16:33

of philanthropies and also tech

16:37

billionaires who are individual donors. There's a lot of voluntary

16:39

labor that goes into this.

16:42

There are people producing research

16:42

that they'll put up in web forums.

16:46

It's not peer reviewed, but there's

16:46

such an urgency around these ideas

16:50

within the community that people are

16:50

riffing off of each other and building

16:53

off of each other's work very quickly.

16:55

Yeah, and it's so interesting to

16:55

hear you talk about it because it's so

16:59

obvious that there's an intersection here

16:59

between technology and human values at

17:05

which philosophy is at the center and

17:05

it's funny, because on the one hand,

17:13

I feel like a lot of people in today's

17:13

world think that philosophy is this

17:16

sort of ivory tower pursuit or this like

17:16

useless, symptom of the decline of the

17:22

humanities, philosophy maybe used to

17:22

have its place, but it doesn't anymore.

17:26

And it's clear when you look into AI

17:26

safety and AI ethics that a lot of the

17:29

people who are developing some of the big

17:29

picture ideas around this are philosophers

17:34

and also some of the ideas that non

17:34

philosophers are building on, who are,

17:39

people who are important in this community

17:39

are coming from philosophy, especially

17:43

from utilitarianism in the 19th century.

17:46

And so I'm curious what you think

17:46

about this, because even though as a

17:50

philosopher, it's exciting to me to

17:50

think about there being an important

17:55

role that philosophy is playing in

17:55

the cutting edge of human technology.

17:59

It's also, I think, clear from

17:59

your research that this, is

18:03

being done in pretty problematic

18:03

ways in certain respects.

18:07

And so I think this is also a

18:07

question about who is in the

18:11

AI safety community, right?

18:13

They're coming from all these

18:13

different walks of life,

18:16

thinking about the intersection

18:16

between AGI ethics and policy.

18:21

Who are the stakeholders here and how

18:21

do you see philosophy playing into this?

18:26

Sure. I think I would break the stakeholders

18:27

down along disciplinary lines,

18:30

institutional lines, and then who

18:30

gets recruited into the community.

18:33

So along disciplinary lines, when we were

18:33

looking at who does AI safety work, right?

18:38

We started by looking at Sam

18:38

Bankman Fried's FTX future fund.

18:42

So Sam Bankman Fried is the now

18:42

disgraced former crypto billionaire.

18:45

He was a big effective altruist who

18:45

really seemed to believe in kind of

18:50

the values of the community and to back

18:50

up and explain what that is, right?

18:54

Effective altruism was a kind of

18:54

movement founded by graduate students

18:58

in philosophy from Oxford and Cambridge.

19:01

They were really interested in

19:01

applying utilitarianism and thinking

19:03

about maximizing human welfare,

19:03

not only in the current day, but

19:07

really in that far term future. And they came up with ideas

19:09

like earning to give, right?

19:11

So getting the highest paid job you can

19:11

and donating 10 percent and gradually

19:15

more of your income to charities that

19:15

they thought were related to cause areas.

19:20

So produce that vision of a future utopia.

19:23

And as I was mentioning, that gets tied

19:23

into them thinking about that long term

19:27

future thinking about threats to it

19:27

around the same time you had another

19:32

Oxford philosopher Nick Bostrom writing

19:32

books like super intelligence and kind

19:36

of Thinking through thought experiments

19:36

about that future that would involve,

19:42

either having quote unquote, super

19:42

intelligent, artificial intelligence

19:45

or artificial general intelligence. He has eugenicist thought experiments

19:46

in there about, what it would look

19:51

like if we selected out certain

19:51

fetuses in human reproduction and

19:56

try to reproduce for specific traits.

19:58

I bring this up to save the disciplinary

19:58

background of people who become

20:01

attracted to those things we noticed

20:01

were people who had studied computer

20:05

science and engineering, statistics,

20:05

math, some philosophy and physics, right?

20:10

There were like a few disciplines

20:10

that quite a few people came from.

20:14

They would start reading some

20:14

of the blogs in this space that

20:16

engage with some of these issues. They'd maybe read super intelligence

20:18

or read MacAskill's well, What We Owe

20:22

The Future is his newer book, but he

20:22

had older books on effective altruism.

20:25

They read Toby Ord's The Precipice

20:25

and this became the set of like books.

20:29

The canon for this kind

20:29

of growing community.

20:32

And then in terms of institutional

20:32

backgrounds, you have a lot of

20:35

people working in tech companies. I did graduate school at Berkeley in

20:37

the Bay Area, and I had been brushing

20:40

shoulders with these people for years,

20:40

but they had always been on the edges of

20:44

some of the conversations in academia.

20:47

You have certainly academic computer

20:47

scientists, but it's not mainstream.

20:51

Even now, there are, there's a whole

20:51

infrastructure in this, space that's been

20:56

making me think a lot about how one of

20:56

the things I really like about working

21:00

on tech from an interdisciplinary and

21:00

like justice oriented perspective is that

21:05

there was this paper a few years ago that

21:05

talked about social roles for computing.

21:09

And one of them was computing as

21:09

synecdoche or the idea that looking

21:13

at computational systems that have

21:13

social effects can make you reflect

21:17

on what is structurally, unacceptable

21:17

about those institutions, right?

21:22

So if you think about like

21:22

Virginia Eubanks book, Automating

21:25

Inequality, and she comes up with

21:25

this idea of the digital poorhouse.

21:28

She's talking about how algorithmic

21:28

systems in welfare distribution, in

21:34

determining, whether or not to allocate

21:34

housing to an unhoused person, that

21:38

these create a kind of digital poorhouse. And the point of a book like that

21:39

is to show that there are all these

21:42

social structural forces that get

21:42

baked into algorithms that create

21:45

a digital poorhouse effect, right? Not just that the technology

21:47

itself is doing it.

21:50

So computing as synecdoche has

21:50

been really interesting to apply

21:53

to looking at the institutions

21:53

popping up around AI safety because

21:56

they have their own philanthropies. They are creating their own non

21:58

profits, their own think tanks, their

22:01

own polling organizations that create

22:01

surveys and thus statistics around

22:06

what percent of the population believes

22:06

that AI is going to kill us all.

22:09

And then, pumping those things out to

22:09

the media and to policymakers in the same

22:14

way that many of these institutions that

22:14

have already existed but did not have

22:17

this worldview have been doing forever. And so a lot of the times people will

22:19

ask me, do you feel like this is a

22:23

distortion of those institutions? And I'm like, no, it's like

22:24

computing a synecdoche.

22:26

They're using all of those institutions.

22:29

The way they have been designed to be

22:29

used, there's a deep irony in AI history.

22:35

there's a recent book called How Data

22:35

Happened by Matt Jones and Chris Wiggins

22:38

where they have only one chapter on AI,

22:38

which I appreciate because AI is not

22:42

really the whole history of technology. And they talk about how AI was actually

22:44

basically like a marketing term that

22:48

the mathematician John McCarthy credits

22:48

himself as coming up with when he wanted

22:52

funding from the Rockefeller Foundation.

22:55

So they say that from its inception,

22:55

AI was a hype term to create a field

23:00

and distinguish it from other fields,

23:00

but it was really very much steeped

23:04

in, other things that were happening,

23:04

like symbolic approaches to creating

23:08

what we're now calling artificial

23:08

intelligence, and I find that history

23:12

so relevant when you look at how

23:12

that kind of Effective Altruism, AI

23:16

safety intersection, it wouldn't have

23:16

catapulted into public attention if

23:20

there weren't hundreds of millions

23:20

of dollars in these philanthropies

23:23

that this space has created. They have spun up National Science

23:24

Foundation grant money, like 20 million

23:28

grant pools to fund this work and to pump

23:28

it through more traditional institutions

23:34

that are slower and not really equipped

23:34

for what happens when something that many

23:39

researchers contest as pseudo scientific

23:39

and fear mongering kind of work takes up

23:43

residence and in terms of students asking

23:43

for courses on things like AI alignment

23:48

and forming campus groups, tech companies

23:48

potentially building teams around this,

23:52

tech companies reorienting themselves so

23:52

that suddenly everyone's working on large

23:56

language models when they could have been

23:56

working on all sorts of other things.

24:00

As I've been doing this work, I've been

24:00

paying attention to what does it say about

24:03

how fragile all of these institutions were

24:03

that they so quickly fell for this hype.

24:07

Yeah. And the infusion of cash into

24:08

this discussion is so real.

24:11

just in the last several years, I've

24:11

noticed as a philosopher and as somebody

24:15

who works in a humanities space.

24:17

And as a San Francisco resident...

24:19

yeah. Yeah. In San Francisco, I that

24:19

all over the place, right?

24:23

this venture capital interest in AI, and

24:23

also the kind of braggadocious attitude

24:29

that you get from people who are not

24:29

experts in philosophy to just appropriate

24:33

philosophical ideas and throw them

24:33

because it gives them a certain kind

24:36

of cultural and technological capital

24:36

that they are very happy to cash in on

24:41

the moment they get the opportunity. But the point here is that, yeah,

24:43

there's a lot of money rushing in into

24:46

this intersection of ethics and AI.

24:49

And a lot of this seems to involve,

24:49

in particular, the effective

24:54

altruism group, which is one

24:54

of the kind of sub communities.

24:59

that belong to this larger umbrella

24:59

of the AI safety community.

25:04

And so I want to talk about them for

25:04

a little bit because the effective

25:08

altruism group, of course, a lot of

25:08

their thinking is based on William

25:13

McCaskill's writings and publications.

25:16

And he himself was deeply influenced

25:16

by Peter Singer's utilitarianism.utop

25:20

And I want to get your

25:20

take on this community.

25:23

So tell us just in a few words

25:23

what effective altruism is and

25:27

then how you interpret the work

25:27

that it's doing in this space.

25:32

I would say there are two

25:32

things of late that have influenced

25:35

how I think about effective altruism. One is an article Amia Srinivasan wrote

25:37

in 2015 called Stop the Robot Apocalypse.

25:42

And, she was questioning, what

25:42

is the value of utilitarianism to

25:46

practical things like policymaking? What is effective altruism?

25:50

As of that date, what was it really doing?

25:52

And I think the conclusion she settled

25:52

on was that it made people feel

25:55

like they were contributing to some

25:55

kind of radical change when really

25:58

they were maintaining a status quo. And I would argue, yeah, that

26:01

was written eight years ago.

26:05

And what has really changed other than

26:05

in the last year or two when so much

26:08

money has been poured into this, it's

26:08

really changed conversations and some

26:12

of the things people are working on. But when you work on something like

26:13

AI safety and it's super loosely

26:18

defined from that community. It's very hard to point to specific

26:20

wins and, benchmarks or measurable

26:25

signs that you've made systems safer. And that is some of the critique that

26:27

I see coming from, really critical

26:31

colleagues in computer science is saying,

26:31

these are such narrow definitions.

26:34

We're not rather, these are such kind of.

26:36

vague terms and narrow definitions that

26:36

something like systems engineering,

26:43

the kinds of things you would do to

26:43

make sure that like a plane is safe.

26:45

You can do that for certain types of

26:45

AI systems, if you look at them in

26:49

their context, but a lot of this is

26:49

so decontextualized and so abstract.

26:53

So that's one thing. The other, when I think about effective

26:54

altruism, there's this really wonderful

26:58

graduate student in, Europe, Mollie

26:58

Gleiberman, who's been writing about the

27:02

difference between public EA and core EA.

27:06

And she's arguing that a lot of

27:06

the areas more related to maybe

27:10

Peter Singer's work, like issue

27:10

areas like animal welfare, right?

27:13

Effective altruists care

27:13

about pandemic prevention.

27:16

They've talked about preventing nuclear

27:16

destruction, that some of these ones

27:20

that seem reasonable to an outside

27:20

observer, more palatable, that is public

27:25

facing EA, but core EA, which is these

27:25

ideas that are more contestable, right?

27:31

The possibility that AI could kill us all. Some of these kind of more transhumanist

27:33

problematic beliefs that is at this

27:38

core and is kind of Trojan horsed

27:38

in by this other set of ideas.

27:42

I really recommend her work. I'm still digging into it.

27:44

And she makes comparisons to

27:44

the structure of this community

27:47

and the federalist society. that there is something so inherently

27:49

conservative about this while

27:53

presenting itself as radical. And I would argue that is of a piece

27:54

with how ideology becomes material

27:58

in Silicon Valley in general. this one has a twist, right?

28:01

Because there's this sort of Bay

28:01

Area meets Oxbridge tie, like those

28:06

are the two kind of core places

28:06

that are married in this movement.

28:10

But, a book I really love actually,

28:10

and I'd read in 20 20 or 2021.

28:17

And I thought, gosh, I wish

28:17

somebody would write this about, E.

28:19

A. And A. I. But there's probably not enough happening.

28:21

And then lo and behold, it's a few years

28:21

later, and I'm doing it is Adrian Daub.

28:26

What Tech Calls Thinking is such

28:26

a great little book looking at

28:30

just a few major figures, right?

28:32

Mark Zuckerberg of Meta, Peter Thiel of,

28:32

you name it, PayPal, Palantir, various

28:38

investments in other companies, and

28:38

trying to understand like, yeah, what

28:42

are the ideological underpinnings of

28:42

the things they have produced, right?

28:45

With Zuckerberg, he went to college

28:45

for a little bit at Stanford.

28:48

What, from the courses that he

28:48

was very likely to have taken

28:51

there because they were required,

28:51

are features we see in Facebook?

28:56

With Thiel, who's a big Ayn Rand

28:56

head and makes it known even now,

29:01

Oh, I didn't know that!

29:03

Yeah. Yeah. Ellie, I'll get you this book.

29:07

Thanks.

29:08

yeah, I think that idea of it's

29:08

really just maintains the status quo

29:13

and there's a public facing in a core

29:13

like that has been fueling a lot of how

29:17

I think about this at the same time. One of my questions about this is to

29:19

what extent when people enroll in these

29:24

communities, do they realize that? That there is that distinction.

29:27

To what extent do they

29:27

believe in the causes?

29:29

Because most of the people I've

29:29

interviewed are very earnest, and not

29:32

all of them are, identify as effective

29:32

altruists or believe in long termism.

29:36

There's different reasons and motivations

29:36

as this community spreads that people

29:40

who are not part of the kind of earlier

29:40

waves would come to it for, But something

29:45

that's really interesting that I noticed

29:45

is even as I learned about effective

29:48

altruism in 2019 from people I knew in

29:48

the Bay Area who were in China when I was

29:52

doing field work there, and they would

29:52

even then say while wearing EA swag, and

29:58

this actually comes up in like a recent

29:58

New Yorker article, people will wear

30:01

EA swag while saying, Oh, but I don't

30:01

really identify as an effective altruist.

30:05

I just like the community. And so this has stuck in the back

30:06

of my mind while I do the research

30:10

where it seems like some of the

30:10

biggest things people are looking

30:12

for are communities where they can

30:12

talk about these things without being

30:15

stigmatized and being taken seriously.

30:17

At the very beginning of doing this work,

30:17

we looked at something like 80, 000 hours,

30:22

which is a big EA career hub website.

30:26

It has. It's really thorough job boards that

30:26

you can scroll through based on what

30:30

EA cause area you're working on. So they have tons of jobs in

30:32

AI that they say these can

30:34

contribute to preventing AI X risk.

30:37

Go work for, OpenAI and thropic, but

30:37

even Microsoft and the UK government,

30:42

like suddenly the range of where you

30:42

can do that has expanded and they

30:45

have a framework of importantness,

30:45

tractability and neglectedness.

30:49

And so they were saying that

30:49

at the beginning, AI safety

30:52

is a marginalized issue. Everybody's working on ethics,

30:54

but nobody's working on

30:57

safety from existential risk. It's so important and it's

30:59

tractable, but it's so neglected.

31:03

And so I now point out that there's been

31:03

a flip where it's not neglected anymore.

31:06

And there's so much money going into this. What is this community actually

31:08

going to do with its power?

31:12

And I want to dive into the

31:12

ideas behind this a little bit more as

31:16

well, because as terrifying as it is to

31:16

hear that Peter Thiel is all into Ayn

31:22

Rand, who I think most people within

31:22

philosophical communities, if we consider

31:27

her a philosopher at all, consider her a

31:27

pretty straightforwardly bad philosopher.

31:31

It's just basically capitalist

31:31

ideology cloaked in,

31:35

philosophical sounding language. So if there's that side that's

31:37

pseudo philosophy or just like pretty

31:41

obviously bad philosophy, there's

31:41

also a lot of effective altruism.

31:44

This is probably the dominant movement,

31:44

as we alluded to before, that is grounded

31:48

in a much more, let's say, academically

31:48

respectable tradition, which is the

31:54

philosophical tradition of utilitarianism.

31:57

And one of the examples that comes to mind

31:57

for me here is that effective altruism

32:03

argued that it's better to provide

32:03

deworming medication in Kenya than it is

32:08

to provide educational funding because,

32:08

you can only help better educational

32:14

outcomes in Kenya if people have access

32:14

to deworming medication because, dying

32:20

due to parasites is a huge problem there.

32:23

So you have that, which is like a

32:23

utilitarian calculus that seems benign

32:27

to, like pretty straightforwardly helpful,

32:27

I think, from various vantage points.

32:32

But then more recently with this

32:32

rise of interest in AI, you get

32:35

I think far more controversial

32:35

utilitarian calcul calculacies?

32:40

Calculi? Anybody? Either of you? Or my more friends?

32:44

Calculations.

32:45

Okay, thank you. Thank you. Calculations. which is better than providing funding for

32:47

deworming in Kenya, let alone education.

32:53

People should be providing funding

32:53

for AI because AI is really what

32:58

poses an existential risk to humanity

32:58

altogether, not just to people in Kenya.

33:02

And so what you get, you mentioned

33:02

long termism before, what you get

33:05

is this idea that actually so many

33:05

human resources and so many people.

33:10

so much funding should be going into AI

33:10

at the expense of real world struggles

33:15

that people are dealing with now. And so I'm really curious

33:17

what you think about that.

33:19

Are there benefits to this or is it

33:19

just like a toxic ideology in your view?

33:24

Are like, should we really be

33:24

worried about the existential risks?

33:27

And even if it's in a different way from

33:27

the way that they're thinking of it, or

33:30

are they like getting something right? Yeah.

33:32

What do you think?

33:34

I had a really thoughtful friend

33:34

who has moved between the worlds of

33:38

people who are into this and skeptical

33:38

of it say, I do think things can get

33:42

really weird really soon, and that I

33:42

don't want to discard everything this

33:47

community is working on and I agree

33:47

because what this community has done is

33:53

they're not really challenging capitalism

33:53

or the marketing that leads people

33:57

to adopt AI systems at scale, right? They're accepting that's going to keep

33:59

happening because companies are very

34:01

good at that and people are bought in. And then they're imagining a future

34:04

where that just keeps happening.

34:06

But they've taken a bunch

34:06

of solutions off the table.

34:09

what are things we don't

34:09

have to use AI systems for?

34:12

What are you know, just in regulating

34:12

these technologies, which again, the

34:16

community is very divided on, and

34:16

they're never thinking about regulations.

34:20

People have been asking for years

34:20

around hate speech or copyright or,

34:25

these kinds of like bread and butter

34:25

issues they would see as too focused

34:28

on the present day and not deserving of

34:28

the urgency and the cash infusion and

34:32

the lobbying efforts as these kind of

34:32

more long term speculative ones are.

34:37

But I think even early in this project,

34:37

my colleagues and I were noticing

34:41

that there's short term technical

34:41

fixes coming out of some of the

34:46

work that this community is doing. But I think that they've

34:49

had to create a kind of self

34:51

aggrandizing hero narrative, right? It's, we've been able to watermark

34:53

GPT outputs so that it, somebody

34:58

can see that something was written

34:58

with ChatGPT and not by a human.

35:02

That's valuable.

35:04

Okay, this is good to know

35:04

for me since I'm on sabbatical,

35:06

not grading papers right now, but I

35:06

will be back in the classroom soon.

35:10

Yeah, I, don't use GPT

35:10

for a variety of reasons, and I am

35:15

curious about if they figured out the

35:15

watermarking or not, but I, something

35:20

like that is a bounded technical

35:20

fix that makes a lot of sense.

35:23

Will it or will it not prevent some

35:23

of the worst case scenarios that

35:27

this community is coming up with? I'm not sure, but I wonder if doing

35:28

that kind of really bounded ordinary

35:33

work and it's for a product by a company

35:33

needs to come with this narrative.

35:38

there's really a lot of questioning of

35:38

how will history remember us and the

35:41

idea that history will remember any of

35:41

the specific people working on this.

35:45

So I just, I've recently really

35:45

thought about how much of this

35:48

is both overcompensated labor

35:48

for what it actually produces and

35:54

how much it helps other people. And not very compensated and voluntary

35:56

sometimes labor that is part of this

36:01

narrative of like you're, going to

36:01

be the reason that we get utopia.

36:06

And I don't know for how long

36:06

people are going to be willing

36:09

to suspend that kind of belief.

36:13

So that's one piece of it that

36:13

I've been thinking about, in

36:16

terms of is this valuable or not. And I asked people who have access to it.

36:20

Some of the bigger names in this field

36:20

and to interview them publicly, ask them

36:24

what we stand to lose if they're wrong. They never have to talk about that.

36:28

They never have to talk about

36:28

maybe this was a huge waste of

36:30

money, resources, time, careers.

36:33

some journalism that was partially

36:33

inspired by our work coming out

36:37

of the Washington Post was looking

36:37

at how the rise of AI safety on

36:40

campus is basically how our college

36:40

students creating groups around this.

36:44

Kind of tapping themselves into this

36:44

network I described of philanthropies,

36:48

fellowships, boot camps, ways of being

36:48

in the network of people working on this,

36:52

and then getting to work in the companies

36:52

and telling yourself you're working

36:55

on these issues and, maybe making 400,

36:55

000 working at Anthropic on alignment.

37:00

And,

37:01

All right, guys, this

37:01

is my last day at Overthink.

37:04

I'm going to do that

37:04

next, I'm just kidding.

37:08

Now you understand how to do the

37:08

culture thing, if you read our research.

37:13

but if, there are some people whose

37:13

timelines of either we get it wrong and

37:18

everything's a disaster, we get it right.

37:20

And we live in utopia for people

37:20

who have shorter timelines,

37:22

like three to five years. I'm really curious about what they're

37:24

going to be saying in three to five years.

37:27

Will there be some new excuse for

37:27

why this didn't all shake out?

37:30

I also find it something that

37:30

this all makes me reflect on is

37:33

this community of people is not. They're just not as interdisciplinary

37:35

as I think they've been made out to be.

37:38

They're very limited ways of knowing and

37:38

selective cherry picking of things from

37:43

philosophy, from geopolitics, right?

37:45

Like for me in particular, most of my

37:45

work was based in China, and a lot of

37:49

my work is centered around what I'm

37:49

calling technological jingoism, or the

37:52

over securitization of issues related to

37:52

technology and society and the economy.

37:58

In the moment we're in right now with the

37:58

TikTok ban and having seen this coming

38:02

down the line for a very long time,

38:02

a lot of my research looks at how the

38:06

US government in particular is just so

38:06

afraid of the United States losing its

38:10

economic and geopolitical dominance over

38:10

the world that casting China or any other

38:16

adversary out to be this existential

38:16

threat to us and that supremacy becomes

38:22

an excuse for all kinds of things like

38:22

banning apps and making it out to seem

38:27

like all technology coming out of China is

38:27

surveilling you and part of a grand scheme

38:31

plan that the Chinese government has. Many of, these are also speculative

38:34

and not provable, and you're

38:37

supposed to just believe that it's

38:37

in the best interest of everyone.

38:41

I think with the TikTok ban, creators

38:41

are really clapping back and watching

38:45

like congressional hearings and coming

38:45

up with fairly like sharp critiques.

38:50

But it takes me back to AI safety

38:50

when I think, they're fully

38:54

bought into the jingoism as well. they will say things like, if we get

38:56

AGI and it leads to bad outcomes, at

39:00

least it would be better if it came out

39:00

of the United States than China, which

39:03

would just turn all of us authoritarian. And they've completely ignored

39:05

the authoritarian tendencies

39:08

within democracies ours.

39:10

I think this speaks to one of

39:10

the central dreams of the people in

39:15

the AI community, which is the dream

39:15

of taking their engineering technical

39:20

skills onto a social playing field

39:20

where they're no longer playing with

39:24

the engineering of systems and machines,

39:24

but they're actually tinkering with

39:28

human bonds and with time itself, right?

39:31

They're engineering. Yeah. Yeah. the future. So there is clearly a social mission here

39:33

that is essential for them to maintain

39:38

the goodness of their motivations and

39:38

their intentions in entering this terrain.

39:43

And so I want to ask you a question here

39:43

that picks up on this discussion that you

39:47

just started about the relationship to the

39:47

state and to government, because the AI

39:52

safety community, I would say as a whole,

39:52

and correct me here if I'm mistaken,

39:57

sees itself primarily as a social

39:57

watchdog, as this canary in the mine

40:02

that's keeping society from destroying

40:02

itself by blowing the whistle on this

40:06

doom that is coming down the hatch.

40:08

And by adopting this role, their hope

40:08

is to influence law and policy, right?

40:14

They want to determine the content of law.

40:17

And I want to know what some of

40:17

the more concrete legal proposals

40:23

and recommendations that grow

40:23

out of this community are.

40:27

What are they proposing for managing risk?

40:31

What role do they think that governments

40:31

should play in the regulation of private

40:36

industry, for example, in order to

40:36

ensure an AI safe future, or is it

40:42

that the people in this community are

40:42

just advocating self regulation on

40:46

the part of industry left and right?

40:48

So it's a bit of a mixed bag. You have people who are very afraid

40:49

that there will be malicious actors

40:53

who will use AI to build bioweapons.

40:56

And as my mentor here at UCLA, Safiya

40:56

Noble, and computer scientist Arvind

41:02

Narayanan and Sayash Kapoor who

41:02

write all about AI snake oil have all

41:05

commented the instructions for how

41:05

to build a bioweapon have been on the

41:08

internet since there was an internet. But what you'd really need is

41:10

like the resources and, you

41:14

need a lab, you'd need reagents. You'd actually have to build it.

41:16

You can't just tell a ChatGPT to

41:16

do it and it'll do it for you.

41:19

And so what is AI safety advocates

41:19

of basically preventing the creation

41:27

of bioweapons using AI, talking to

41:27

policymakers and saying things like,

41:30

we're afraid that there will be individual

41:30

people who will buy up enormous amounts

41:34

of compute resources that could be

41:34

used to build something like that.

41:37

There was a recent AI executive order

41:37

talking about using the defense production

41:42

act as like a way of regulating this and

41:42

making sure it doesn't happen, treating it

41:45

like you would the creation or acquisition

41:45

of other kinds of weapons that has taken

41:51

a lot of energy away from other kind of

41:51

more practical prosaic ways of regulating

41:56

AI in everyday life, because, as we

41:56

know, policymakers, time is limited.

41:59

The amount of work it takes to

41:59

get something passed into whether

42:02

an executive order or just any

42:02

other policy document is finite.

42:07

You have a subset of people

42:07

in the AI safety world.

42:11

Sometimes externally, people

42:11

would call them doomers, right?

42:14

The ones who are certain that this

42:14

will only lead to horrible outcomes

42:18

and are saying, let's just stop it

42:18

all until we have proper safeguards.

42:21

That, the thing I like to point

42:21

out about that community is it's

42:25

not particularly well funded. And as a critic like Timnit Gebru

42:26

of the Distributed AI Research

42:30

Institute, would say, They're also

42:30

fomenting hype in saying these systems

42:36

are so powerful, we must stop them.

42:38

Then you're contributing to what

42:38

computer scientist Deb Raji would

42:42

call the fallacy of functionality. And Deb has this amazing paper

42:43

with a bunch of colleagues around

42:46

typologizing the different ways and

42:46

points at which there's an assumed

42:51

functionality that is not always there.

42:53

It can come in, the design and the

42:53

marketing and the deployment, interacting

42:57

with the system that you're told does

42:57

something it doesn't actually do.

42:59

And that's just not something

42:59

this community engages with.

43:03

And that, honestly, I think a lot

43:03

of policymakers don't have the

43:06

time and resources to work on.

43:09

I've talked to civil society advocates

43:09

who are deeply skeptical of AI

43:11

safety, and they feel like AI safety

43:11

proponents have a monopoly on the

43:16

future and how people think about the

43:16

future because they did such a good

43:19

job of selling an idea of utopia.

43:21

That anybody sounding alarms

43:21

for caution, they haven't done a

43:25

particular job of painting a future. Anyone would want to live in because

43:26

they're so busy talking about the

43:29

things that are going wrong now. And then also self regulation

43:31

is something the companies

43:34

are absolutely advocating for. They really, they'll say, let's just

43:36

do things like red teaming or paying.

43:41

Often again, data workers, but

43:41

sometimes people within tech

43:44

companies to come up with, for

43:44

example, malicious prompts for GPT.

43:48

And then let's figure out how to

43:48

design an automated system that

43:52

would take in human feedback and

43:52

then be able to stop those prompts

43:55

from being addressed in the future. So some of the classic examples

43:56

are, Hey GPT, how do I burn my

43:59

house down and get, arson money?

44:01

And it's instead of telling you exactly

44:01

how to do that, it would automatically

44:05

know from feedback, I shouldn't tell you

44:05

how to do that and you shouldn't do that.

44:08

Okay. But of course, at the same time,

44:09

there are open source language models

44:13

that have not taken any of that into

44:13

account and are down and dirty, happy

44:17

to tell you about how to, conduct

44:17

ethnic cleansing, commit crimes.

44:24

So there's this suddenly like wide

44:24

landscape of options where like OpenAI

44:27

is the relatively safe one and something

44:27

produced by like Mistral is not.

44:32

And as we approach the end of our

44:32

time, I'm really curious where you stand

44:36

on all of this, because obviously your

44:36

research is on AI safety, which means

44:41

that you're studying the communities who

44:41

look at AI safety rather than, trying

44:46

yourself to figure out what AI is.

44:48

the future of AI safety looks

44:48

like, but of course, I imagine

44:52

that's given you some thoughts too.

44:55

Are you a doomer when it comes to AI?

44:57

Are you whatever the

44:57

opposite of a doomer is?

44:59

Are you somewhere in between? And what do you think is, what do you

45:01

think is the most promising feature of

45:05

what we may or may not correctly call AI? And what do you think is, the

45:08

most threatening possibility?

45:11

I don't think any of the pre

45:11

existing labels around any of this fit

45:15

where I put myself because something like

45:15

the concept of AI ethics being treated

45:20

as if it's in opposition to AI safety has

45:20

never made sense to me because people who

45:24

work on a wide range of like social and

45:24

labor and political issues involving AI

45:29

don't necessarily characterize themselves

45:29

as working on AI ethics, myself included.

45:33

I'm really interested in following

45:33

in the tradition of anthropologists,

45:36

like Diana Forsythe, who were studying

45:36

AI expert systems of the 1980s and

45:41

nineties and how, what did people

45:41

think they were doing when they did AI?

45:45

that was her question. And that's mine around this work.

45:49

What do I think is most promising? I think right now I still come back

45:50

to computing as synecdoche and making

45:54

just any of these technologies telling

45:54

us something else about the human

45:58

experience and what in our institutions

45:58

can be changed outside of technology.

46:03

Some of my favorite organizations

46:03

and individuals I've worked with

46:06

have often talked about how the

46:06

technology piece of what we're

46:08

working on is such a small part of it. And it's all of these other social

46:10

factors that are much more important

46:14

that we foreground in our work. And what would I say is

46:16

perhaps the most dangerous?

46:20

I think that we've been through so many

46:20

phases of saying, if you just taught

46:23

computer scientists ethics, they'd

46:23

somehow be empathetic to these causes.

46:28

I don't know that's enough. And that what that assumes about

46:30

a computer scientist as a person

46:33

and where they're coming from

46:33

seems to be misguided as well.

46:38

one of the critiques coming from within

46:38

this community or from outside of

46:41

this community about this community

46:41

from people like to Timnit Gebru a

46:44

brew and Emile Torres is that AGI

46:44

is like a eugenicist project, right?

46:47

And there's so much good writing

46:47

about technoableism or the idea

46:50

that technology should move us all

46:50

away from having bodies that are not

46:54

normatively abled and fit these like

46:54

ideals of an able bodied person that is

47:00

constantly improving themselves and that

47:00

disability has no place in the world.

47:04

Like this work is very

47:04

techno ableist, right?

47:07

Even just the concept of intelligence

47:07

and not interrogating it further and

47:12

you don't even have to grasp that

47:12

far to find the eugenicist arguments

47:16

in superintelligence and that nobody

47:16

in that community is really talking

47:21

about that because it's uncomfortable.

47:24

In arguing that AGI is like a,

47:24

eugenicist project, it's interesting

47:29

to look at then the building of

47:29

the field that I have studied.

47:32

parallels some of how race science

47:32

happened that you had to create journals

47:36

and venues to legitimize it as a science

47:36

when so many people contested it.

47:40

I would love to see AI safety

47:40

properly engaged with that.

47:44

I would love to see some of these ideas

47:44

around intelligence crumble and like what

47:48

alternatives could we have instead, but

47:48

that's not where money and time is going.

47:53

And of course, some of the

47:53

biggest proponents of this in

47:55

the mainstream our edgelords

47:55

like Elon Musk and Mark Andresen.

48:01

And so that really moves us away from

48:01

like any of the questions I just raised.

48:07

Dr. Ahmed, this has been a

48:08

wonderful discussion.

48:10

You have given us a lot to

48:10

chew on and to think about.

48:14

And we thank you for your time and for

48:14

this important work that you're doing.

48:18

Thanks so much.

48:19

Thanks

48:19

much .

48:25

Ellie: Enjoying Overthink? Please consider supporting the

48:26

podcast by joining our Patreon.

48:29

We are an independent,

48:29

self supporting show.

48:32

As a subscriber, you can help us

48:32

cover our key production costs, gain

48:36

access to extended episodes and other

48:36

bonus content, as well as joining our

48:40

community of listeners on Discord. For more, check out Overthink on Patreon.

48:44

com.

48:46

Ellie, that was such

48:46

a great discussion with Dr.

48:48

Ahmed, and I cannot mention that

48:48

this is the first time we do

48:53

an episode where a community of

48:53

experts is the subject matter.

48:58

and I found that just really illuminating,

48:58

but I wanted to mention that after

49:01

we finished the interview, Dr. Ahmed said, Ellie and David, above

49:03

all, what I really want is for your

49:07

listeners to walk away with a sense

49:07

of calm about the subject matter.

49:11

So what are your thoughts

49:11

about this Ellie?

49:13

Yeah. I was really surprised to hear that.

49:15

And to the point that we were

49:15

both like, Oh, we got to mention

49:18

that because it happened right

49:18

after we finished the recording.

49:21

Because I don't know, I feel like as

49:21

we were hearing all about the AI safety

49:27

community, as she was describing it,

49:27

I was coming away with the sense that

49:33

there must be major risks because

49:33

everybody is talking about them, even

49:37

though one of her key points was this

49:37

might not be worth funding in the

49:42

way that we're funding it, right? So I do think that's just something

49:44

to rest on, this idea that, maybe

49:49

the AI risks are not as extreme

49:49

as we think, and there is at least

49:53

something overblown about it. Not that we don't take these questions

49:54

seriously, but just One of the things

49:59

that stuck with me, and I'm not going

49:59

to remember the exact term right now

50:02

because we just finished recording, it

50:02

was something like functional fallacy,

50:06

or she, quoted somebody who worked in

50:06

this field who had said quite a number

50:11

of years ago, I think not with respect

50:11

to AI specifically, but with respect

50:14

to technology, that we tend to have

50:14

this fallacy in our thinking where

50:19

we assume that there's a danger of a

50:19

certain technology that technology has

50:24

not yet proven itself to have, right?

50:26

And so whether it's, oh, AI is going

50:26

to put all of the writers out of work

50:33

because now you could have AI write

50:33

a Shakespeare play, partakes of this

50:37

fallacy because AI has not at all proven

50:37

itself to be able to creatively write.

50:42

Yes, it has proven itself to be capable

50:42

of doing certain writing tasks such

50:47

as write copy for an Instagram post,

50:47

but I would say that writing copy for

50:52

an Instagram post doesn't require the

50:52

same kind of creative capabilities

50:56

that writing a Shakespeare play does. And so even if it's not out of the

50:57

realm of possibility that AI could be

51:00

genuinely creative in that sense, it is.

51:03

Not evident yet that it can be,

51:03

or that it's even tending that

51:08

way, given what we know so far. I'm reminded a little bit about

51:10

our AI in the arts episode, or

51:13

other AI episode, on this point.

51:15

And so I think that is useful to

51:15

keep in mind, this idea that AI might

51:19

lead to the extinction of the human

51:19

race, because it can end up, choosing

51:24

its own way, it can have freedom.

51:27

requires having an inkling that AI

51:27

could be capable of having freedom,

51:33

which we just like, it doesn't

51:33

have that functionality right now.

51:36

Yeah, no, and I actually would

51:36

go even further than that, because

51:39

what you're alluding to is the case in

51:39

which we see risks that are not there.

51:44

We just collectively hallucinate

51:44

these dangers and then make

51:48

decisions about policies and about

51:48

funding on the basis of that.

51:52

But I think the other side to that

51:52

coin is that because we're so concerned

51:56

with those long term, unrealistic

51:56

scenarios, we start missing dangers

52:02

that are already materializing, and

52:02

we don't even consider them as risks

52:07

because there's a section in one of Dr.

52:09

Ahmed's papers where she talks about

52:09

the importance of really being concrete

52:14

about what the risks actually are. And I decided to follow the reference,

52:16

and it's to a paper that was published in

52:20

2020 by Inyoluwa Debra Raji and Roel Doby.

52:25

And they talk about how the

52:25

AI safety community overall.

52:30

When they talk about risks,

52:30

they never talk about the risks

52:33

that actually matter to people. So for example, think about the

52:35

dangers of automating taxis.

52:40

That's not really front and

52:40

center in these discussions.

52:43

Usually you talk about more doomsday

52:43

scenarios, or even the dangers

52:48

associated with the production

52:48

of algorithms and machines.

52:52

So one of the examples that they

52:52

mentioned, and this is the only

52:55

one that I'll, list here, but

52:55

there are other ones in that piece.

52:59

They say often in order to train an

52:59

algorithm, you have to hire a lot of

53:03

humans to do the work of literally

53:03

training it on the data, right?

53:07

Teaching it how to classify things

53:07

so that it then learns a rule.

53:11

And right now, that's something

53:11

that we do by outsourcing that very

53:14

mechanical labor onto exploited

53:14

workers, often in the third world.

53:20

So I actually heard Shazeda

53:20

present on a panel at UCLA this

53:23

spring on AI safety and ethics.

53:26

And there was another really prominent

53:26

researcher there whose name I'm

53:30

unfortunately forgetting now, who

53:30

talked about how A lot of the content

53:34

moderation and other sort of really

53:34

unpleasant tasks that form the dark

53:39

side of internet labor have, not

53:39

only been outsourced to the Global

53:45

South, but actually map on perfectly

53:45

to former patterns of colonialism.

53:51

So who is content moderating

53:51

material in France?

53:55

People in North Africa. Who is content moderating

53:57

material in America?

54:00

The Philippines. Who is content moderating

54:02

material in Britain? India, right?

54:05

And I found that to be really interesting

54:05

here because I think it also pertains

54:10

to both a dark side and a potential

54:10

upside of AI, which is that if these

54:15

unpleasant tasks are being outsourced

54:15

to formerly colonized people, such

54:19

that we're still living in this, techno

54:19

colonial society, Could those tasks

54:25

further be outsourced to AI, right?

54:27

And that could be like a good thing. But of course, the downside of that

54:29

is then, okay, what is coming to fill

54:33

the labor vacuum for these people who

54:33

are currently employed by those means?

54:38

I don't have a lot of faith that the

54:38

neo colonialist capitalist regime that

54:43

we live under is suddenly going to

54:43

have like much more meaningful work

54:48

for the populations in those areas.

54:52

so I have a friend in Paris who

54:52

works in the space of content moderation,

54:56

and he told me that one of the problems

54:56

with the incorporation of AI is that even

55:01

though AI can catch a lot of the content

55:01

that needs to be filtered out, because

55:06

companies don't want to risk even one

55:06

sort ofpost or one photo passing their

55:12

safety filter, they still end up having

55:12

to use human, moderators to catch them.

55:19

And here we're talking about, some of

55:19

the nastiest kind of darkest corners

55:23

of the internet that, people from

55:23

these formerly colonized spaces are

55:28

having to deal with on a daily basis.

55:30

And so just the psychological impact

55:30

of having to watch, child porn,

55:35

murders, like brutal beatings on a day

55:35

to day basis for eight hours a day.

55:40

That's why I say it's so unpleasant.

55:42

Yeah, I know, and I

55:42

guess maybe, I'm, I want a

55:46

stronger term for that than that.

55:49

But the point being is that right

55:49

now it's not as if you can just

55:53

automate that completely, so you

55:53

still need that human component.

55:57

And as usual in this episode,

55:57

we're not coming away with a clear

56:00

sense of, you must do X, Y, Z, and

56:00

then we will avoid the threat of human

56:06

extinction on the one hand or further

56:06

exploitation on the other, which I think

56:11

is an important point because that idea

56:11

that we could have that clear X, Y, Z

56:16

is part of the dream of the effective

56:16

altruism movement and of other members

56:22

of the AI safety community that Dr.

56:24

Ahmed has given us really good

56:24

reason to be suspicious of.

56:30

We hope you enjoyed today's episode. Please rate and review us on Apple

56:31

Podcasts, Spotify, or wherever

56:36

you listen to your podcasts. Consider supporting us on Patreon

56:37

for exclusive access to bonus

56:41

content, live Q& As, and more. And thanks to those of you who already do.

56:45

To reach out to us and find episode

56:45

info, go to overthinkpodcast.

56:49

com and connect with us on Twitter and

56:49

Instagram at overthink underscore pod.

56:54

We'd like to thank our audio editor,

56:54

Aaron Morgan, our production assistant,

56:58

Emilio Esquivel Marquez, and Samuel P.

57:00

K. Smith for the original music. And to our listeners, thanks so

57:02

much for overthinking with us.

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features