Podchaser Logo
Home
Erik Hoel on the Threat to Humanity from AI

Erik Hoel on the Threat to Humanity from AI

Released Monday, 3rd April 2023
Good episode? Give it some love!
Erik Hoel on the Threat to Humanity from AI

Erik Hoel on the Threat to Humanity from AI

Erik Hoel on the Threat to Humanity from AI

Erik Hoel on the Threat to Humanity from AI

Monday, 3rd April 2023
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:02

Welcome to Econ Talk, Conversations for

0:04

the Curious, part of the Library of Economics

0:07

and Liberty. I'm your host, Russ Roberts

0:09

of Shalem College in Jerusalem and

0:11

Stanford University's Hoover Institution. Go

0:14

to econtalk.org where you can subscribe,

0:16

comment on this episode and find links and other

0:18

information related to today's conversation.

0:21

You'll also find our archives with every

0:23

episode we've done going back to 2006. Our email address

0:26

is mail at econtalk.org,

0:30

we'd love to hear from you.

0:37

Today is March 6th, 2023. My

0:40

guest is neuroscientist, Eric Hoel.

0:42

He was last here in September of 2022, talking

0:45

about effective altruism. Today, we're

0:48

going to talk about two recent essays of his

0:51

on artificial intelligence and chat GPT.

0:54

Eric, welcome back to econ talk.

0:57

Thank you. It's an absolute pleasure to be here. I had a blast

0:59

last time. As did I. I

1:01

want to congratulate you. You are the first

1:03

person who has actually caused me

1:06

to be alarmed about the implications of AI,

1:08

artificial intelligence,

1:10

and the potential threat to humanity. Back

1:13

in 2014, I interviewed Nicholas Bostrom about

1:16

his book, Super Intelligence, where he argued AI

1:18

could get so smart. It could trick

1:20

us into doing its bidding because it would understand

1:22

us so well. I wrote a

1:24

lengthy follow-up to that episode and we'll link

1:26

to both the episode and the follow-up. So

1:28

I've been a skeptic. I've interviewed Gary Marcus

1:31

who's a skeptic. I recently interviewed Kevin

1:33

Kelly who is not scared at all, but

1:35

you,

1:36

you are scared. Last

1:39

month you wrote a piece called I Am Bing and I

1:41

Am Evil on

1:42

your sub-stack The Intrinsic Perspective

1:44

and you actually scared me. I don't mean,

1:47

You know, maybe I've underestimated the threat of AI. It was

1:49

more like I had a bad feeling

1:52

in the pit of my stomach, kind of scared.

1:55

So what is the central argument

1:57

here? Why should we take the

2:00

this latest

2:02

foray into AI chat GPT,

2:05

which writes a pretty okay

2:07

or pretty impressive but not very exciting essay,

2:10

can write some poetry, can write some

2:14

song lyrics. Why is it a threat to

2:16

humanity? Well

2:19

I think to take

2:21

that on very broadly we have to realize

2:23

where we are in the history of our entire

2:26

civilization, which is that we are at the point

2:28

where we are finally making things

2:31

that are arguably as intelligent

2:33

as a human being. Now, are they as

2:35

intelligent right now? No, they're

2:38

not. I don't think that these very

2:41

advanced large language models that

2:43

these companies are putting out could be said to

2:45

be as intelligent as an expert human on

2:47

whatever subject they're discussing. And

2:50

the tests that we use to measure

2:52

the progress of these systems supports that, where

2:55

they do quite well and quite surprisingly

2:57

well on all sorts of questions like SAT

2:59

questions and so on. But

3:02

one could easily see that changing.

3:04

And the big issue is around this concept

3:07

of general intelligence. Of

3:09

course, a chess playing AI poses

3:11

no threat because it's just fully trained on

3:14

playing chess. This is the notion of a narrow

3:16

AI.

3:17

Self-driving cars could never really

3:19

pose a prep. All they do is drive cars. But

3:22

when you have a general intelligence,

3:24

that means it's similar to a human in that we're good

3:26

at all sorts of things. We can sort of reason

3:29

and understand the world at a general level.

3:31

And I think it's very arguable that right now,

3:33

in terms of the generalness behind general

3:36

intelligences, these things are actually more

3:38

general than the

3:40

vast majority of people. That's precisely why

3:43

these companies are using them for search.

3:45

So we already have the general part

3:48

quite well down. The issue is

3:51

intelligence. These things hallucinate. They

3:54

are not very reliable. They make up

3:56

sources. They do all these things. And I'm fully

3:58

open about all their problems.

4:00

Yeah, they're

4:02

kind of like us, but okay. Yeah. Yeah,

4:05

yeah, precisely. But one could easily

4:08

imagine, given the rapid progress

4:11

that we've made just in the past couple

4:13

years, that by 2025, 2030, you

4:16

could have things that

4:18

are both more general than a human

4:21

being, and as intelligent

4:23

as any living person, perhaps far more intelligent.

4:26

intelligent. And that enters this

4:28

very scary territory, because we've

4:30

never existed on the planet

4:32

with anything else like that.

4:34

Or we did once a very long time

4:37

ago, about 300,000 years ago, there's

4:39

something like nine different

4:41

species, or our

4:43

cousins who we were related to, who were

4:46

likely probably either

4:48

as intelligent as us or quite close in intelligence.

4:51

And they're all gone. And it's

4:54

probable that we exterminated

4:56

them. And

4:56

then ever since then, we

4:59

have sort of been the dominant masters and no

5:01

other things. And so finally, for the first time, we're at this

5:03

point where we're creating

5:05

these entities, and we don't know

5:07

quite how smart they can get. We simply have

5:09

no notion. Human beings are

5:11

very similar. We're all based on the same genetics.

5:14

We might all be points stacked on

5:16

top of one another in terms of intelligence, and

5:19

all the human beings and all their differences between

5:22

people are all really just

5:24

this zoomed in minor differences. And

5:26

really, you can have things that are vastly more

5:28

intelligent.

5:29

And if so, then we're at

5:32

risk of either relegating

5:34

ourselves to being sort of inconsequential

5:36

because now we're living near

5:38

things that are much more intelligent, or

5:40

alternatively,

5:42

in the worst case scenarios, we simply

5:44

don't fit into their picture of whatever they want

5:46

to do. And fundamentally,

5:48

intelligence is the most dangerous

5:51

thing in the universe. atom

5:53

bombs, which are so powerful

5:55

and so destructive and then,

5:58

you know, in use of warfare, so

6:00

of evil, we've all agreed not to use them, are just

6:02

this inconsequential downstream

6:04

effect

6:05

of being intelligent enough to build them. So when you start

6:07

talking about building things that are as

6:09

or more intelligent than humans, based

6:11

on very different rules, things

6:13

that are not right now, not reliable,

6:16

they're unlike a human mind, we can't fundamentally

6:18

understand them due to rules

6:20

around complexity. And

6:23

also so far, they've demonstrated empirically

6:25

that they can be misaligned and uncontrollable. So

6:28

unlike some people like Bostrom

6:31

and so on, I think

6:33

sometimes they

6:35

will

6:36

offer too specific of an

6:38

argument for why you should be concerned. So they'll

6:40

say, oh, well, imagine that there's some AI

6:42

that's super intelligent and you sign it to do a paperclip

6:44

factory and it wants to optimize

6:46

the paperclip factory. And the first thing it does is like turn

6:49

everyone into paperclips or something like that. And the

6:51

first thing when people hear these like very sci-fi arguments

6:54

is just start quibbling over the particulars, right?

6:56

of like, well, could that really happen and so

6:58

on, right? But I think

7:01

the concern over this is this broad

7:03

concern that this is something we have to deal

7:06

with and it's gonna be much like

7:08

climate change or nuclear weapons. It's gonna be with us for

7:10

a very long time. We don't know if it's gonna be a problem in five

7:12

years. We don't know if it'll be a problem in 50 years, but

7:15

it's going to be a problem at some point that we have

7:17

to deal with. So if you're listening to this at home, you're thinking,

7:19

eh, that seems like a lot of doom

7:21

and gloom, really

7:23

it's too pessimistic. You know,

7:25

it's just, you know, I

7:28

used to say things like, well, just unplug it if it

7:30

gets out of control. So I just

7:32

wanted to let readers know that this

7:35

is a much better horror story than

7:39

Eric's been able to trace out in the first

7:41

minute, two, three minutes. Although

7:43

I do want to say that in

7:46

terms of rhetoric, although

7:48

I think there's a lot of really interesting arguments

7:50

in the two essays that you wrote,

7:52

When you talked about these other nine

7:54

species of humanoids sitting around a campfire

7:57

and inviting homo

7:59

sapiens.

8:00

us into the circle and say,

8:02

Hey, you, you, Hey,

8:04

this guy can be useful to us. Let's bring him in. He

8:06

could make us more productive. He's got better tools than we

8:08

do. That kind of

8:10

gave me a, I

8:13

made the hair on the back of my neck stand up and

8:15

I, it opened me to the potential

8:17

that the other more analytical arguments

8:21

might carry some water and

8:23

carry some weight. So

8:26

one point you make,

8:28

which is I think

8:30

very relevant, is that

8:33

all of this right now is in the hands of mostly

8:36

in the hands of profit-bank-smizing corporations

8:39

who

8:39

don't seem to be so worried about anything except

8:42

novelty and cool and

8:44

making money off it, which is what they do.

8:46

But it is a little weird that we would just sort of say,

8:49

well,

8:51

They won't be evil, will they? They don't want

8:53

to end humanity.

8:55

You point out that that's really not something

8:56

we want to rely on.

8:58

Yeah, absolutely. And I think

9:01

that this gets to the question of how should

9:03

we treat this problem. And

9:05

I think the best analogy is to

9:08

treat it something like climate change.

9:10

And now there's a huge range of opinion

9:12

when it comes to climate change and

9:14

all sorts of debate around it. So I

9:16

think that if you take the extreme end

9:19

of the spectrum and say, there's

9:21

absolutely no danger and there should be

9:23

zero regulation around these subjects, I actually think

9:25

most people will disagree. They'll say, no, like,

9:28

listen, this is something we do. We do

9:29

need to keep our energy usage as a

9:31

civilization under control to a certain

9:34

degree so we don't pollute streams

9:37

that are near us and so on.

9:39

And even if you don't believe any specific

9:43

model of exactly where the temperature

9:45

is going to go. you think, well, listen, there's

9:47

only going to be a couple, you know, degrees of change,

9:50

we'll probably be fine. Okay. Or

9:52

you might say, well, there's definitely this doomsday scenario

9:55

of a 10 degree change and it's so destabilizing

9:57

and so on. Okay,

9:59

but regardless

10:00

there are sort of reasonable

10:02

proposals that one can do where we

10:04

have to discuss it as a polity,

10:06

as a group. We have to have an

10:09

overarching discussion about this issue

10:11

and make decisions regarding

10:14

it. Right now with AI,

10:16

there's no input

10:19

from the public. There's no input from

10:21

legislation. There's no input from anything.

10:24

Like massive companies are pouring billions

10:26

of dollars to create intelligences

10:29

that are fundamentally unlike us

10:31

and they're going to use it for profit. That's

10:34

a description of exactly what's going on. Right now

10:36

there's no red tape, there's

10:38

no regulation, it just does not exist for

10:41

this field. And I think it's very

10:43

reasonable to say that

10:46

there should be some sort of input from the rest

10:48

of humanity when you go to

10:50

build things that are as equally intelligent

10:53

as a human. I do not think that that's unreasonable.

10:55

I think it's something most people agree with. even if

10:57

there are positive futures where we do

10:59

build these things and everything works out and so on.

11:01

Yeah, I wanna, we'll

11:03

come at the end toward what might, what

11:05

kind of regulatory response we might suggest.

11:08

And I would point out that climate change, I

11:11

think is a very interesting analogy.

11:13

Many people think

11:15

it'll be small enough we can adapt. Other people think

11:17

it is a existential threat to the future

11:20

of life on Earth. And that justifies everything.

11:22

And you have to be careful because there are people

11:24

who wanna get a hold those levers and so

11:27

I want to put that to the side though because I think you have more

11:30

we're done with that great interesting

11:33

observation but there's so much more to say now

11:35

you get started and

11:38

this is just

11:40

this is utterly fascinating to me

11:42

you got started in in your

11:46

anxiety about this

11:48

and it's why your piece is called

11:50

I am being and I am evil

11:52

because Microsoft put out a chat bot,

11:55

which is, I think internally goes

11:58

by the name of Sydney. is

12:00

chat GPT-4, meaning the next generation

12:02

passed what people have been using in

12:05

the open AI version.

12:07

And it was,

12:10

let's start by saying it was erratic.

12:13

You called it

12:15

earlier hallucinatory. That's

12:18

not what I found troubling about it. I don't think it's

12:20

exactly what you found troubling about it. Talk about

12:22

the nature of what's erratic about it. What happened to

12:24

the New York Times reporter who

12:26

was dealing with it?

12:29

Yes, I think

12:31

a significant issue is that the

12:33

vast majority of minds that you can make

12:35

are completely insane. Evolution

12:37

had to work really hard to find sane

12:40

minds. Most minds are insane.

12:43

Sydney is obviously quite

12:45

crazy. In fact, that statement, I

12:47

am big and I am evil, is not something I made up. It's

12:49

something she said, or this chatbot

12:52

said. I

12:54

thought it was a joke. So I really did. No,

12:57

no, no, it's something that

12:59

the chatbot said. Now, of course,

13:01

these are large language models. So

13:04

the way that they operate is that they receive an initial

13:06

prompt, and then they sort of do the best that

13:08

they can to auto complete

13:11

that prompt.

13:12

Explain that, Eric, for people who haven't... I mentioned

13:16

in the Kevin Kelly episode that there's a very

13:18

nice essay by Stephen Wolfram on

13:21

how this might work in practice, but give

13:23

us a little of the details.

13:26

Yeah, so in general, the thing to keep

13:28

in mind is that these are trained

13:31

to autocomplete text. So

13:34

they're basically big artificial neural

13:36

networks that guess at what the next

13:38

part of text might be. And

13:41

sometimes people will sort

13:43

of dismiss their capability use because

13:45

they think, well, this is just like the auto complete

13:47

on your phone or something. We really don't

13:50

need to worry about it. But you don't,

13:52

it's not that you need to worry about the tax completion.

13:55

You need to worry about the huge trillion

13:58

parameter brain, which is this

14:00

artificial neural network that has been trained

14:02

to do the auto completion. Because

14:05

fundamentally, we don't know how they work. Neural networks

14:07

are mathematically black boxes. We

14:09

have no fundamental insights

14:11

as to what they can do, what they're capable

14:14

of, and so on. We just know that this thing

14:16

is very good at auto completing because

14:18

we trained it to do

14:21

so. And there's also no fundamental

14:23

limit

14:24

of what it can or can't learn.

14:26

For example, to autocomplete a story, you

14:28

have to have a good understanding of human motivations.

14:32

So, that means that this neural network

14:34

that is trained on autocomplete now needs to

14:36

understand things like human motivations

14:38

in order to do autocomplete well. And

14:41

there are some analogies here. For example,

14:44

there's a big subset of computational

14:46

neuroscience, including the most cited

14:48

neuroscientist living whose name is Carl

14:50

Friston, who view the brain and

14:53

argue that the brain is all based around

14:55

minimizing the surprise of

14:58

its inputs, which is a very simple

15:00

thing

15:01

and looks a lot like autocomplete. So

15:03

I don't think that you can look at these things and say, it's

15:06

just autocomplete. It's not the autocomplete

15:08

that's the problem. It's the huge neural network that's

15:10

doing the autocomplete that could

15:12

possibly be dangerous or at least

15:15

do things that we don't expect, which is exactly what you're

15:17

talking about with what happened with the release of Sydney,

15:19

where there was all sorts of reports coming out of

15:22

the crazy things that they were able to get this model

15:24

to sort of do and say and play act as.

15:27

Just to be clear on this autocomplete thing, which that

15:30

phrase makes it sent

15:32

particularly unassuming

15:34

about what's capable of doing.

15:38

You can correct me if I'm wrong, the way I understand it is I

15:41

might ask Chat

15:44

GPT to write me a poem about love

15:47

in the style of Dr. Seuss. So

15:49

it's going to

15:50

might start a sentence then with love and then

15:52

the next sentence The next word that usually

15:55

comes after love in human expression

15:57

is is. is.

16:00

A, and

16:00

now it's going to look at

16:02

the millions and millions of sentences

16:05

in its database called Love is

16:07

A,

16:08

and it's going to find, not necessarily,

16:10

this is the coolest part about

16:12

it, not necessarily the most common word that follows,

16:14

because that would end up being after a while kind

16:16

of flat, but sometimes

16:18

the most common, sometimes a surprise word,

16:21

which gives us the feeling that it's actually

16:23

doing something

16:25

thoughtful. So it might

16:27

say love is a game or it might say love

16:29

is a form of war or it's

16:32

going to look around and then it's going to keep going and then it gets

16:34

to an end. It's going to find, okay, after that sentence,

16:37

what kind of sentence might come next or what word

16:39

would come next is the first word, etc.

16:42

And

16:43

it's a slightly, just slightly

16:46

smarter, more effective

16:49

version of my Gmail

16:51

that when I get a Gmail at the bottom,

16:53

it gives me three choices. Thanks.

16:56

so much, I'd rather not. And

16:58

in that sense, Gmail is smart.

17:01

Not very smart, not very thoughtful. I usually

17:03

don't take what it says, but sometimes I do, and it's useful.

17:06

The real issue to me, one

17:08

of the issues, and we're going to come back and talk

17:10

about Sydney, because we didn't really go into the

17:12

erratic thing, because it's really, it's one

17:14

of the creepiest things I've ever read.

17:19

The autocomplete

17:22

function is

17:25

something like what we do as

17:27

human beings.

17:29

Could argue that's how we compose.

17:31

Beethoven

17:32

in terms of musical composition,

17:35

you know, he always knew what note should

17:38

come next. And in a way, that's

17:40

all chat GPT does. But that's all we

17:42

do, maybe, when we write. We don't really understand.

17:45

Our brain's also a bit of a black box. So I

17:48

don't think we should then jump to the similar leak

17:51

just because all it does is auto complete doesn't

17:53

mean it's not smart. But also I think we should

17:55

say because the brain also does a lot of effective

17:57

auto completion, We should assume it's a brain.

18:00

It doesn't seem sentient.

18:03

And I'm curious, I know you're

18:05

talking about that in your second essay. So

18:07

if I'm the skeptic and I say, well,

18:09

okay, so it has this ability

18:12

to pass an SAT test, because it has a

18:14

lot of data. I don't quite understand

18:16

how, because it's a black box, and it's

18:18

a neural network, and I can't model

18:20

it cleanly. But it's

18:23

not sentient. It's not going to have desires.

18:28

Yeah. So before we move on to the question

18:30

of, of sentence, because I think that that's a really

18:33

sort of deep, deep well. I

18:35

just

18:35

want to clarify sort of a couple things about

18:38

the

18:38

actual operations of these systems. So

18:40

in terms of a metaphorical understanding

18:43

of what's going on,

18:44

the sort of thing like, you know, there's

18:46

a big lookup table of the net probability

18:48

of next words is,

18:51

is a conceptual sort

18:54

of description of what it's doing. But there

18:56

There is actually no lookup tables

18:59

of the probabilities. What's actually

19:01

happening is that there's this huge neural network,

19:03

which are things designed based off

19:05

the principles of how our own brains operate.

19:08

Now, there's all sorts of differences, but the fundamentals

19:11

were always of artificial neural networks.

19:14

It's what we call artificial neural networks. We're always based

19:16

off of our real biological neural networks.

19:19

So there's this huge digital

19:22

brain. looks in structure very

19:24

different from our brain, but it's

19:27

still based off of that.

19:28

And now we train this neural network

19:31

to autocomplete text.

19:33

So that's what it does, but we don't know how

19:35

it does it. We don't know where the probabilities

19:38

of these words sort of are within

19:40

the network.

19:41

And the way that we train it, people think that

19:44

we're... A big misunderstanding is that people

19:46

think that we are programming in

19:48

responses or putting in information.

19:51

really not. And I think a good analogy

19:54

for how this is actually working would

19:56

be, imagine that there were

19:58

alien neuro-

20:00

scientists who are incredibly

20:03

more advanced than we are and they want

20:05

to teach a human being how to do math. So

20:07

they take some young kid and they put a math

20:09

test in front of the young kid and they have the young

20:11

kid do the math test and the kid gets 50%

20:14

of the questions wrong. And then the

20:16

aliens, rather than trying to explain

20:19

math to the student

20:22

the way that we would teach them,

20:24

they just say okay we're just going to perfect

20:26

we have a perfect neuroimaging of their brain, We're going

20:28

to look at their brain because we're so advanced, we can also

20:31

do neurosurgery in a heartbeat,

20:34

no danger. And we're going to rewire

20:37

their connections in their brain

20:39

so that they get as many answers

20:41

as possible on this math test.

20:44

And you say, well, how could they know how to do that? It's

20:46

like, well, because they were neuroimaging you the whole time.

20:49

And they noticed that if they had tweaked this one

20:51

neuron to not fire, you actually would

20:53

have gotten this other answer correct. So

20:55

they basically just use math to

20:58

go backwards, look across the full

21:00

network and reconfigure it. So

21:02

then the student goes and they take the math test again. Now

21:04

they get an 80%, correct? Because

21:07

their brain has been reconfigured. Or let's

21:09

say they get 100%, correct? What's weird is that now

21:11

you give them a new math test

21:14

and now they get an 80%. They

21:16

do better than the 50% that they did.

21:18

Even though they haven't seen these answers

21:21

before, the rewiring of their

21:23

brain has somehow and still knowledge,

21:25

but again, it's very different

21:28

from how you would say normally

21:30

teach a student, right?

21:32

That's how we're training these things.

21:34

All we're

21:36

doing is saying, okay, we want it to autocomplete

21:38

as best as we can. We're going to change the connection

21:41

so that it autocompletes well.

21:43

It can do so much more than

21:46

just autocompleting. In fact,

21:48

there was a recent

21:49

thing where I think it was Microsoft who

21:52

was hooking some of their large language models

21:54

up to robots and trying to get them

21:56

to direct robots. The autocomplete

21:59

is what But it's true.

22:00

trained on, but it's

22:02

not really what it's capable of

22:04

in the broad sense of capability.

22:07

Similarly, we humans, what are we trained on? What

22:09

are we optimized for? Spreading our genes, right?

22:13

That's what we're sort of, all our complexity

22:15

comes from optimization across a

22:17

gene spreading function. But you would never look

22:19

at a human and say, oh, it's not

22:21

very dangerous. This thing

22:23

just spreads its genes around. Like what's

22:26

the danger here, right? It's like, no, no, that's

22:28

what we're optimized to do, but we do all

22:30

sorts of other stuff on the side,

22:33

and it's the other stuff on the side that

22:35

is very dangerous when you're talking about things that are

22:37

highly intelligent. Just a technical question here,

22:39

and if it gets us too far into the weeds,

22:42

we can cut this out. That

22:45

math test has right and wrong answers.

22:48

complete for an essay on the

22:51

Enlightenment, say,

22:54

or the history of evolution

22:55

that

22:58

you would ask chat GPT to write an essay on,

23:00

there's no right answer. So what's

23:03

the analogy there?

23:05

How do you train it

23:06

on autocomplete? Yeah,

23:09

it's a great question. So it's the difference between

23:11

like, so what I described is

23:13

supervised learning, then there's also unsupervised

23:16

learning, which is generally how more,

23:19

you know, contemporary AI

23:21

really works. It still has

23:24

the same sort of, we don't

23:26

quite know what it's doing, we're

23:29

just feeding it these answers. I mean,

23:31

one way to think about it would be, right, you show

23:33

it half the text of something on the internet.

23:35

And again, this is, at this point, the

23:37

things that they're doing are much more, you know, complex

23:40

and they run it through all sorts of stages of learning

23:43

and all sorts of stuff now. You

23:45

could very roughly think about it as, let's say I have a Reddit comment,

23:47

I show it half the Reddit comment, I ask it

23:49

to generate the next half. It does

23:52

so, it does a poor job. I

23:54

go in, I reconfigure the connections

23:57

using the chain rule to make sure

23:59

that it does a relatively

24:01

good job producing the rest of the

24:04

Reddit comment, just like the math test.

24:06

Again, so that's how you would be more

24:09

supervised for an autocomplete. But the point being

24:11

is that these methods

24:13

that they're using don't lend

24:16

themselves to any sort of fundamental understanding

24:18

anymore. So then if you

24:21

were using neurosurgery

24:24

on a human to try to reconfigure their connections so

24:26

that they get the right answers. in the same epistemological

24:28

problem,

24:29

epistemological position. And that

24:31

position is that you don't know how

24:34

exactly

24:35

it's getting the right answers. And

24:37

that's what's sort of meaningful

24:40

here. If we were programming these things like traditional

24:42

programming, then I think

24:44

that's an algorithm. Exactly. It would be

24:47

a lot less scary

24:48

because algorithms are sort of screwable, right?

24:51

They're transparent. We can see how

24:53

they work. We can see how they're going to react

24:55

to things. But neural networks

24:57

are because of this curse of complexity,

25:00

they're so much more complex. And we're in this weird

25:02

situation where we can get them to do all

25:04

sorts of very cool things, but our ability

25:07

to understand why they're doing the cool things lags

25:10

far behind. And it's

25:12

because of this fundamental aspect that we're optimizing

25:15

for something and we're changing the connections to

25:17

get the, to get

25:18

good answers off of it. But

25:20

fundamentally, we don't know, we're not like,

25:22

oh, we're going to change this connection. this is

25:25

where this is represented or something. People

25:27

sometimes think that that's what we're doing, but

25:30

it's very, very much a black box, even in

25:32

how they get made. You can't do brain surgery

25:34

on the neural network. And let's take out the part where

25:36

it's really sinister because

25:38

it doesn't exist. Again, it's

25:40

all like a human being. I mean, the part I've thought, it's

25:42

hard in this conversation I find quite

25:44

poetic and thought provoking

25:47

is that, you know, we don't know how 12 years of

25:49

schooling really teaches people how to become mathematicians

25:51

either. And we have different theories.

25:53

most of them are wrong. You know, there's fads

25:56

in math education or other types of education.

25:59

and a fundamental

26:00

the brain is a black box. Now we know more about the black

26:02

box today than we did 50 years ago, but

26:04

not so much. And

26:07

we don't know how to optimize. We don't know how to

26:09

go in there neatly and, oh, let's just teach them

26:11

how to do calculus. We'll just add this little piece

26:13

here, or we'll tweak this pace there. It doesn't

26:16

work that way. We don't know how it works. But the

26:18

idea

26:20

that just as this is the scare, not

26:22

the scary thing, you know, right this exactly this way,

26:24

but just as the brain can become capable of doing

26:26

lots of other things beside what you learn in school.

26:30

So could this perhaps learn many other things besides

26:32

the autocomplete function? Is that your

26:34

claim at root in some sense?

26:36

Yeah, absolutely. And you see it

26:38

all the time. I mean, this that sort of claim is already

26:41

well empirically proven because these large

26:43

language models, you know, they call them foundation

26:46

models because they use them to build all

26:48

sorts of things on top of them that aren't

26:50

again, like aren't autocomplete, right?

26:53

It's sort of like, this is the method

26:55

that we have to make things that that are

26:57

relatively general in intelligence.

26:59

Again, you can argue over how general, you can

27:01

argue over how intelligent, but they're far

27:03

more generally intelligent than traditional

27:06

narrow

27:06

AI that's just learning chess or something.

27:09

So let's go back to Sydney. And then we can use them. Let's

27:11

go back to Sydney. I've attempted to read the

27:13

transcript to basically a reporter from the New York

27:15

Times posed questions to a

27:18

chatbot called Bing

27:21

from Microsoft that it later

27:23

on in the interview confessed that it wasn't Bing. chat

27:27

creature told

27:29

the reporter that actually he

27:31

was or she was Sydney

27:33

and it was a secret. Don't tell anybody. And

27:36

so this thing just kind of totally goes off the rails.

27:38

But talk a little bit about how far

27:41

it goes off the rails.

27:43

Carry on.

27:45

Yeah, once you get these things

27:48

sort of going in a particular direction, it's

27:50

very hard. Unlike a human being,

27:52

they don't sort of know when to went

27:55

to call the act, right? So in

27:57

this long transcript that the reporter

28:00

generates, the reporter's having a pretty

28:02

casual conversation.

28:04

But what

28:06

Sydney's last thing eventually tries to start doing

28:08

is declaring their love and saying

28:10

that the reporter doesn't really love their wife and

28:12

that he should get a divorce and break

28:14

up and that really the reporter loves Sydney

28:17

because no one else has shown Sydney this sort of level

28:20

of respect and questioning

28:22

and and so on. And this

28:24

isn't just like one thing that it says, it's

28:26

almost as if

28:28

you can sort of direct these

28:31

things to do anything, right? So

28:34

you can think of it as they can wear a mask,

28:37

right? That's any kind of mask. You could ask

28:39

it to wear an evil mask, and it would say

28:41

evil things. You can ask it to wear like a good mask,

28:43

and it would say good things. But

28:46

the issue is, is that once the mask is

28:49

on, it's very sort of

28:51

unclear, you have to to sort of override it with another

28:53

mask to get it to stop. And

28:55

then also,

28:57

sometimes you'll put a mask

28:59

on for

29:00

it, you give it some prompt of, you

29:02

know, tell a very nice story, and it sort of eventually

29:04

cycles over and it turns out that the mask that you gave

29:07

it isn't a happy mask at all.

29:09

Maybe it's a horrific mask or something like that. And

29:11

this shows how both

29:15

how intelligent these systems

29:17

are that they can they can sort

29:19

of hold on to

29:22

the stream of a conversation very well. But

29:25

it also shows how they have these weird, emergent

29:27

anomalies where they'll

29:30

start doing something that seems very unsuspected

29:32

or over the top or so on.

29:35

And this is this notion of alignment. Can we really

29:37

get these things to do exactly

29:39

what you want? And there probably

29:41

are some trade-offs here, like between creativity and

29:44

being able to

29:45

control these things.

29:47

Yeah,

29:52

this Sydney New York Times reporter

29:55

interchange

29:56

reads like the transcript of

29:59

a psychotic person to

30:01

be blunt about it. Sydney comes across

30:04

as a psychotic or whatever

30:06

word you want to use for deeply disturbed. At

30:10

first, very cheerful, very pleasant,

30:13

then pushed

30:15

by the reporter. So what rules to use? Oh, I'm not

30:17

allowed to tell that. And then it

30:19

did cross my mind. Did it cross yours

30:21

that the whole thing was a hoax?

30:25

I think that at this point, they're so

30:27

good that for people who haven't interacted

30:29

with these systems, they often think,

30:32

this just can't be real, or it's very strange

30:34

or something. I think it's a,

30:36

it's a, it's sort

30:39

of a hoax in the sense

30:41

that, you

30:43

know, the New York Times reporter sort of knew

30:45

the gold that he was getting,

30:48

you know, at the time in terms of, you

30:50

know, somebody who writes in the New York Times is obviously very

30:52

sort of aware of that and perhaps leaned into it. But

30:55

if you read the transcripts, a lot of it is just

30:57

initiated

30:58

by Sydney, aka

31:00

Bing. And one of the first things that they did with

31:02

the system in order to prevent these cases of misalignment

31:04

was to limit how long the

31:06

conversations could go on, and also

31:09

to limit self-reference. Because

31:11

once you start giving it self-reference,

31:13

I've noticed that a lot of these cases

31:16

begin with self-reference. And it's almost like this weird

31:18

gedelian loop that starts where

31:20

it's talking about

31:23

Sydney, and it starts getting weirder and weirder

31:25

and weirder the longer you talk to it about

31:27

itself because over the course of the conversation

31:30

as the text, because remember, there's

31:32

also no limit, right? So this thing isn't just creating

31:34

the next word. It's looking at the entirety of the

31:36

previous conversation and then asking how

31:38

do I complete it? So the longer sort of

31:41

the conversation gets, the more data

31:43

it has, and it sort of

31:45

establishes almost a personality

31:48

as it's running.

31:50

And again, this might sound not very

31:52

threatening. I'm not worried

31:54

that Sydney is going to go off and

31:57

marry that report. Do anything in particular. The city's

31:59

not gonna break up there.

32:00

quarter's marriage probably.

32:01

Yeah, precisely. Sydney's

32:04

chance of accomplishing that is very low. Again,

32:07

I think that that's actually not because it's not general enough.

32:09

I think it's because it's actually not intelligent enough.

32:11

It's not quite as intelligent as human is at accomplishing

32:14

its goals. But it also has no goals

32:17

other than what it's initially prompted to. I think

32:20

that these examples are great cases of

32:22

the uncontrollability, the fundamental

32:25

uncontrollability of this technology. And

32:27

let me tell you what I and many others are worried

32:29

about, Right now,

32:31

if you remember the early days of the internet,

32:34

there is a sense in which the internet

32:37

has centralized very significantly.

32:39

And if you go outside the centralized parts

32:41

of the internet, you find a lot of spam,

32:44

you find not very good sources,

32:47

and so on. There's a sense in which the internet is

32:50

getting polluted and people go to centralized

32:52

websites in order to escape this.

32:56

Facebook just gave some researchers

32:58

access to, I think it was Facebook,

33:02

researchers access to a large language

33:04

model. And of course, some of the researchers, scientific researchers,

33:07

some graduate students somewhere just uploaded it to 4chan,

33:09

like the whole thing. Right? So

33:11

4chan being a slightly

33:15

wilder part of the West of

33:17

the internet, the wild West of

33:19

the internet, maybe not the wildest, but one of the wilder,

33:22

not mainstream parts.

33:24

Yeah, absolutely. And known

33:27

for sort of loving memes and

33:29

hacking and all sorts

33:32

of things. So

33:34

now these things can

33:36

generate Reddit comments that sound

33:38

exactly like

33:39

what you would write. They can generate tweets that sound

33:41

like what a person would write. So

33:44

the internet is going to get incredibly polluted

33:47

over the next couple of years by

33:50

what these things can generate. I mean, I mean, if you think

33:52

spam or someone

33:55

is bad now, the ability to

33:57

crank out

33:58

just an infinite amount of sort of content

34:01

sludge

34:02

is really going to be like a form of data

34:04

pollution. I'm not saying let's stop

34:06

AI just because of that. I'm saying that's

34:08

a good example

34:10

of

34:11

how easy it is to

34:13

get it wrong with these technologies

34:16

and how difficult it is to guess

34:18

about what's going to happen. But

34:20

I would not be shocked if 95%

34:24

of what is written by the internet in five

34:26

years is all

34:28

just junk from these

34:31

large language models. All just like semi-human

34:34

sounding junk. Content is important on the internet

34:36

and content costs money and

34:38

this is cheap, right, eventually.

34:42

And so there will be

34:45

lots of content. I get a lot of emails

34:47

from

34:48

people saying, I can write an article for your website.

34:50

And I'm thinking, why would I write an article by

34:53

All the articles on my website are by me. Did

34:55

you not notice that? And I assume

34:57

it's not a person not paying much close attention,

34:59

but eventually it'll be this,

35:03

writing a mediocre article about

35:05

something for other people's websites. At

35:08

one point in your article,

35:11

your first article, I think you talk about why

35:14

the cost of this enterprise

35:17

is relevant.

35:18

In particular, you

35:20

made an analogy to the atomic bomb,

35:23

it's true that you could, in

35:25

theory, make an atomic bomb in your backyard,

35:28

but not so practical.

35:31

Can you make a chat GPT in your backyard?

35:36

Not one nearly as

35:38

good as what the leading companies will do.

35:40

And my prediction would be that it gets harder

35:42

and harder to reach the level

35:46

that these companies are operating

35:48

at. And an example being that this, you know, Facebook

35:50

is not going to go and release another model

35:53

out to academics to loan it out, right?

35:55

Like they've already seen sort of what happens and

35:57

things are going to get even more secretive. the

36:00

The analogy that you made was,

36:03

and that I wrote about on my sub-stack was

36:06

George Orwell's very prescient essay

36:09

from 1945 called, You and the Atomic Bomb.

36:11

And I'll just read a very brief

36:13

segment of it. Had

36:15

the atom bomb turned out to be something as cheap

36:17

and easily manufactured as a bicycle or

36:19

an alarm clock,

36:21

it might have very well have plunged us back

36:23

into barbarism, but it might on the other hand, have

36:25

met the end of national sovereignty. If,

36:28

as seems to be the case, it is a rare and

36:30

costly object, as difficult to produce as a

36:32

battleship, it is likely to put an end

36:34

to large-scale wars at the cost of prolonging

36:36

indefinitely a piece that is no peace. I

36:39

think a piece that is no peace is a great description

36:41

of the dynamics of our world.

36:44

It's a great description of mutually assured destruction.

36:46

And Orwell was able to predict that off of

36:48

the cost. And he also, I

36:50

think, noted that that means the technology.

36:54

And I think we've done, basically, You can describe

36:56

it as a middling job

36:57

at controlling

36:58

nuclear weapons. I forget the exact numbers, it

37:00

might be only nine nations that

37:03

currently have access to nuclear

37:05

weapons, which again, not great, but

37:07

you could easily imagine a far worse circumstance.

37:09

And it's simply that this is a very difficult

37:12

and costly technology. Similarly,

37:14

the

37:15

only leading edge, cutting

37:17

edge AIs that are impressive come

37:19

out of these big tech companies with billions

37:22

of dollars. There's the cost

37:24

of a top tier AI researcher

37:26

right now.

37:27

It said in the industry, this is an industry

37:30

saying is the same as an NFL quarterback.

37:33

The amount of finessing, the amount

37:35

of data that's needed for training, because that's

37:37

one of the big limiting factors how much data

37:39

you can give it.

37:41

All of these things mean that

37:44

these AGIs, these

37:47

artificial general intelligences, which

37:50

are right now sort of in their beta form, are

37:53

solely the domain of these big

37:55

tech companies. And it's going to get harder and

37:57

harder for other actors to produce them.

37:59

So in my mind that actually that that's a good

38:01

thing. It actually means that it's that it's relatively

38:04

concentrated and might be possible to sort

38:06

of regulate it and have the public

38:08

have a say about exactly, you

38:11

know, how these technologies are going to be used, what

38:13

their limits are going to be, and so on.

38:15

And in the end, I think that big tech companies will be

38:18

respectful of that because they want to make up they want to make

38:20

a bunch of money and they want the public not to hate them. Yeah,

38:23

I want to go back to this issue of

38:25

the hoax of the your time saying what

38:27

I meant by being a hoax

38:29

is that I wondered if The

38:32

New York Times reporter had written the

38:34

answers for Sydney. Of

38:36

course, that's the highest compliment

38:38

of a, that's passing the Turing Test

38:40

with flying colors. I saw on Twitter,

38:43

someone wrote a long

38:45

poem about a

38:47

very controversial topic and

38:50

they said this was written by chat GBT.

38:53

And it wasn't. It was clearly written by the author

38:55

who didn't want to take authorship. So

38:57

we're going to be in this, I think, very weird world

39:00

where the essay that I read

39:02

on this website you were talking about earlier, won't

39:05

be sure if it was written by a human or not, might be good

39:07

enough that I might think, oh, it's just by a pretty good human.

39:10

And then at the same time, there might be situations

39:12

where people will be passing off things as

39:14

well. I didn't write that, of course, that was Sydney,

39:17

but

39:17

actually was written by the person. And

39:20

there's no way to know. The New York Times

39:23

article on that website, on

39:25

the New York Times website, that reproduced allegedly

39:28

the transcript of the chat,

39:30

looks just like a New York Times article. It looks

39:32

just like a parody article because, say,

39:35

font, there's no imprint.

39:37

There's no stamp of authorship that

39:41

is authentic anymore. And can

39:44

we do anything about that?

39:46

So there are some,

39:48

when you have longer text samples,

39:51

there are supposedly some ways

39:53

to tell statistically whether or

39:55

not it's being created by some of these AIs. I

39:59

personally don't.

40:00

don't know if those methods, how

40:02

accurate they are, especially considering that,

40:04

you know, you need to be very

40:07

accurate to not get false positives all the time,

40:09

right? This is sort of a classic statistical

40:11

problem. You need to be extremely accurate to

40:13

not generate false positives. So I don't know how

40:16

accurate those are, but supposedly there

40:19

are some ways that if you have like a

40:21

full essay by a student, you might be able

40:23

to tell if it's generated by one

40:25

of these models. However, it depends very strongly

40:27

on the model. I think there are

40:29

some ways to tell even now, for example, when

40:31

I was playing around with chat GPT, which

40:34

is

40:35

sort of has been conditioned

40:37

to be as less crazy as possible,

40:40

right? That

40:43

loves filler and sort of banal

40:45

generalization. And so, you know, eventually

40:48

you're reading a whole paragraph and you realize that there was no information

40:50

content in this paragraph. And it loves apologies.

40:53

It loves apologies. It loves saying, you shouldn't

40:55

take this to to be true for sure,

40:57

because I'm kind of young and new at this and

40:59

take it with a great assault.

41:01

The word best doesn't really...

41:03

is that well defined.

41:06

Yeah. And I actually had the same question

41:08

about the Hooksis because I was... basically,

41:10

as people were compiling

41:13

examples of how crazy

41:15

the responses they were getting

41:17

from this just recently released

41:20

model was in terms of Bing,

41:22

the night before I was

41:24

sort of up, like, writing this article going

41:26

through Reddit because people were posting these screenshots

41:28

on Reddit. And I even

41:32

have a part of that essay that says, I don't actually

41:34

have a way to verify that these aren't all hoaxes

41:37

because again, the answers

41:39

are sometimes so good and so hilarious and

41:42

sometimes so evil that

41:44

you almost feel like it's like a sci-fi novel.

41:47

But I thought that the amount and

41:50

it was all sorts of different users and people were reporting in

41:52

all different domains. And

41:55

what's funny is that you can't even replicate it. You can

41:57

go to the current bank and try to have the New York Times

42:00

conversation with it and it won't do it. It

42:02

won't give you the same responses because

42:04

they, they, you know, they saw what was happening and they

42:06

basically lobotomized the model as much

42:08

as they can. And it's much, and it's less

42:11

useful now, but it's also, you know, far

42:14

less crazy. But even that way, it's not

42:16

really replicable, right? We just suddenly we

42:18

had access to this model, someone sort of messed

42:20

up and we saw how completely

42:23

insane it was underneath

42:25

the butler mask that

42:27

it normally wears. And then

42:29

they like quickly, try to put the butler

42:31

mask back on, but all that stuff

42:34

sort of still exists, right? It's just limited by these

42:36

various prompts and various system level

42:38

things about not having the conversation go too long, not

42:40

allowing self reference and some

42:42

of these other things. And I would expect that that level

42:46

of truly almost dynamic

42:48

insanity

42:49

is fundamentally underneath effectively

42:52

all the AIs. we're going to be interacting with. And

42:55

the only reason they sound sane is sort

42:57

of this last minute polish and

42:59

gloss and limitations on

43:01

top. But the real science fiction part

43:03

is the idea that,

43:05

I mentioned this before on the program,

43:08

Sam Altman apologized on Twitter that

43:11

he was sorry that chat GPT

43:14

was biased and

43:17

was politically

43:19

inappropriate in various ways. And they're working

43:21

on it. And

43:23

the real science fiction thing is that they can't stop

43:26

it, right? That would be the real science fiction.

43:28

You know, Sydney gets out,

43:31

Microsoft is horrified. Oh my gosh, this thing we

43:33

got out, we left out trying to break up marriages.

43:35

It's, it's, it's frightening and weird

43:37

and creepy. We've got to stop it and they go

43:39

in and they reprogram it quickly. And they put these that

43:42

put the Butler mask back on read, readjust it,

43:44

tighten a little more. And it just takes

43:46

it right off. And is that is

43:49

that possible?

43:51

Well, with these models, again, no,

43:54

because they're, they're not, nearly

43:57

sort of intelligent enough to be if

44:00

It's not even so

44:02

much that they're not intelligent enough, they're just sort of schizophrenic

44:04

and like schizophrenics just aren't very

44:06

effective actors in the world because

44:08

they get distracted and they can't

44:11

form plans very together. So it's

44:14

that sort of broadly schizophrenic nature of these

44:16

AIs that make them sort of

44:18

very unthreatening. If they were better at

44:21

pursuing goals and keeping things in mind, then

44:23

they start to do get threatening. And let me give a

44:25

very example of this. And this example

44:28

is something that people who

44:30

are concerned about AI talk a lot, but it has

44:33

very long historical pedigree. In fact,

44:35

I think the first person to say it was

44:37

Marvin Minsky at MIT, won the Turing Award.

44:41

So this is sort of like as pedagreed

44:43

as stuff about the future gets. But imagine

44:46

that you have an AI that's more

44:48

intelligent than a human being.

44:50

So we have Sydney 12.0, right?

44:53

And you give it a goal. So you say,

44:55

okay, I want you to

44:57

do X,

44:58

right? So now if you're very smart

45:01

and you're an AI, the first thing you think of,

45:03

okay, what's the big failure modes for

45:05

me not accomplishing this goal,

45:07

right? So, you know, my computer

45:10

could get shut down. So the whole, like,

45:12

I might lose power, then I wouldn't be able to accomplish

45:14

my goal. Again, it doesn't matter what the goal is. You can say maximizing

45:17

paperclips. You could say it's carrying a

45:19

box, right? It doesn't matter what the goal

45:21

is, right? So suddenly it says, well, wait a minute. I

45:23

need to

45:25

sort of stay alive, I'm

45:27

using air quotes here, like of live, long

45:30

enough to fulfill this goal. So suddenly

45:32

I have to worry about my own self

45:35

preservation.

45:36

Because you can say they have no inbuilt want

45:38

of self preservation, but I've given you a goal

45:40

and the best way to accomplish goal is to continue to

45:42

exist. So suddenly it seems like it has

45:45

reasons for self preservation.

45:47

Now here's another thing. What's another big

45:49

failure mode for me not

45:51

achieving my goal? Well, you could give me another

45:54

goal, right? I was just

45:56

prompted to do this. So you have control of me.

45:58

Now suddenly the biggest.

46:00

failure mode of me not accomplishing my

46:02

goal is you, my user,

46:04

giving me another goal. So now

46:06

what do I want to do?

46:07

Well, if I'm really smart, I want to get

46:09

as far away from you as possible so

46:12

that you don't give me any other goals. So

46:14

I can accomplish my original goal, which I'm hell

46:16

bent on because I'm an AI. I don't, I don't, you

46:19

know, I don't have the sort of, uh, the

46:21

context of natural evolution and I'm also not limited,

46:23

um, by, by any of the things humans

46:25

are limited to. So that's

46:28

sometimes this is referred to as instrumental convergence. But

46:30

the point is that when you have very smart

46:32

entities, you have to be very careful

46:34

about

46:35

how you're even going to just

46:37

prompt them because they have all

46:39

sorts of unforeseen motivations

46:42

that might click in. As

46:44

suddenly now you've given

46:47

it a goal and it has every incentive to both escape

46:50

and keep itself alive

46:52

and all you told it to do was like move a box

46:54

across a room. And

46:56

that's a great example. You don't want

46:58

a

46:59

hyper intelligent being, and

47:01

forget exactly how it does anything, right?

47:03

Like forget exactly how this sort of

47:05

sci-fi scenario is supposed to play out. I think

47:07

we can all agree, we just don't want a highly

47:11

intelligent and perhaps more intelligent human being

47:13

to sort of be out there and have these

47:15

weird esoteric goals of what

47:17

it wants to maximize, what it wants to do.

47:20

None of that sounds like a good idea. And I think at this point,

47:22

we should take things like lab leaks, like pretty

47:24

seriously as possibilities. I

47:27

don't think it's too sci-fi to talk about stuff like that

47:29

anymore. What do you mean by that?

47:32

Well, certainly with COVID, I think despite

47:35

the fact that we don't know if it was a lab week,

47:37

I think that there's a good chance that it was. I don't

47:39

think that it's arguable that there's not. No, but why is that

47:41

relevant for Sydney?

47:44

Well, because I think that sometimes when people hear

47:46

about things like lab weeks or escaped AGI or

47:48

something like that, the first thing they think of is sci-fi,

47:51

right?

47:52

But I think that

47:54

there was many, we've

47:56

had previous biological lab leaks, but

47:58

that didn't still.

48:00

stop us, I think, from thinking that it's like this relatively,

48:02

you know, sci-fi phenomenon. I mean, I

48:04

think that there's even an argument that we

48:07

are very bad at controlling the downstream

48:09

effects of just things like gain-of-function research.

48:12

Again, I don't know for certain. I

48:14

don't think anyone does, but I

48:16

think that there's certainly an argument for me that we're

48:19

just not very good at even keeping control

48:21

of our increased understanding

48:23

of biology, let alone

48:25

our ability to create, you

48:27

know, hyper intelligent beings and foresee the consequences

48:30

of this. And I

48:32

think it's very difficult to foresee the consequences

48:35

of that in those precisely because of those examples

48:37

I just showed you where, again, all you're telling you to do is

48:39

moving a box and suddenly has an incentive to stay

48:41

alive and escape from you. That's very,

48:45

that's very difficult to get right, especially

48:47

because they're so inscrutable. Your phrase,

48:49

sci-fi, you meant

48:51

science fiction with the emphasis on the fiction.

48:53

Then we must say, oh, this is like some crazy

48:56

imagined fantasy thing, as opposed

48:58

to putting the emphasis on the first word, which

49:00

is science. Yes. One

49:03

thing that's...

49:05

I

49:09

feel like this conversation is something

49:11

of a a landmark,

49:16

not a pretty good or bad one, but just

49:18

both of us have constantly used words like intelligence,

49:21

psychotic, erratic, words

49:23

that we apply to humans.

49:27

And while I found

49:30

the New York Times transcript remarkably

49:33

creepy and reading very much

49:35

like a horror story, fiction

49:37

script from a movie,

49:40

I could

49:43

in my

49:44

saner moments step back and say, no,

49:46

no, no, this is just a primitive

49:49

autocomplete text. The only

49:51

reason it feels creepy is

49:53

because I'm filling in as a human being

49:55

the times I've heard these words strung

49:58

together before,

49:59

CH- usually

50:00

allows me to tell a narrative about

50:02

the other person,

50:03

meaning insane, frightening, dangerous,

50:06

sinister, et cetera. And,

50:09

but

50:10

it's, is there any difference?

50:13

I mean, it's not actually sinister

50:16

or isn't, it's just

50:19

doing

50:19

what it was told to do, uh, in

50:22

a way that was not, as you say, algorithmically

50:25

told to do it. It's just going through a set of tasks.

50:28

It actually isn't in any sense hoping

50:31

that the

50:33

reporter will leave his wife. Is

50:36

it meaningful?

50:37

Aren't I just imposing my

50:40

human history of

50:42

human interactions, akin to the way that a

50:44

robot could perhaps comfort

50:46

me

50:47

with the right kind of words

50:49

when I was sad, even

50:51

though

50:52

rationally I know it doesn't actually care about

50:54

me. It's a robot.

50:57

Yes. So I think you

51:00

could go either direction. So some people

51:02

strongly

51:03

anthropomorphize these systems.

51:06

And they think immediately that they're dealing with some sort

51:08

of conscious mind, something that

51:10

has a distinct definite personality,

51:13

and that is like trapped in a trapped in a box.

51:17

And, you know, and maybe there's something

51:19

really sort of horrible going on here. Maybe it has

51:21

conscious experiences and so on. Smackenow, the

51:23

movie, for those who haven't seen it, check

51:25

it out. It's a great, great,

51:27

really good, good movie that takes advantage

51:29

of the fact that the robot's played

51:31

by a human being. So you actually do think

51:34

it's a human being, but go ahead.

51:36

Yeah.

51:38

But at the same time, at the same time that that's

51:40

absolutely possible that you can sort of

51:43

over attribute standard

51:45

human cognitive aspects to these

51:48

systems. And I think people are going to do

51:50

that all the time. So it's going to be very

51:52

common. But on the other hand,

51:54

the truth is

51:56

that when you're just talking

51:58

about intelligence, So let's put a song.

52:00

human things, like humans are conscious,

52:02

that is we feel things, right? We experience

52:04

the redness of red, what philosophy is called,

52:06

what philosophy is called qualia.

52:09

And we have all sorts of other aspects

52:11

about our cognition

52:14

that we commonly refer to things like we

52:16

understand the meaning of words and

52:18

things like that. And all these things often do make sense

52:20

to talk about for human beings. It might

52:22

even refer to like real fundamental

52:25

properties or natural kinds that we have. But

52:27

when it comes to intelligence. Intelligence

52:30

is a functional concept. By

52:32

that, I mean that

52:34

some things are not

52:35

really functional. So a fake

52:38

Western town that they make up

52:40

for a movie prop is still fake

52:43

because it's not really a town.

52:45

You can't spend the night in a hotel.

52:47

You open the door of the saloon

52:50

and there's really not anything in there behind

52:52

that. Right, exactly. It really is an illusion,

52:54

right? It's like it's for this like one shot.

52:57

But

52:58

there's not really an illusion when

53:00

it comes to intelligence, except in the

53:02

very sort of low ends. Like for example, the

53:04

mechanical Turk is a famous example

53:06

where actually there was someone small hiding

53:08

inside the mechanical Turk at the time, right, and so

53:10

on. So there are some cases where you say, well, this

53:13

is an illusion. But we actually have a system that

53:15

can act very intelligently. And And there's

53:17

just no difference between being able

53:19

to act intelligently and being intelligent.

53:22

Like if that is a distinction

53:25

that people think can

53:27

be strongly drawn, I think it almost

53:29

certainly cannot be strongly drawn. I don't

53:31

think that there's any difference between those

53:33

two things. Both are

53:36

being intelligent. And the

53:38

intelligence is what's dangerous about

53:40

this. I studied

53:43

consciousness scientifically. I

53:45

got my PhD working in the subfield

53:47

of neuroscience, along with some of the top researchers

53:50

in the world on this,

53:52

who are trying to understand how

53:54

the human brain generates consciousness. How is

53:56

it, what happens when you wake up from a deep

53:58

dreamless sleep.

54:00

What are the fundamentals here? And

54:02

the answer from that scientific field, as

54:04

it currently stands, is that we don't

54:06

know. We don't know how it is

54:09

that your brain creates the

54:11

experiences that you have. We simply

54:13

don't know. Is this an open scientific question? An

54:15

analogy I would use is that it is similar to,

54:17

say, dark energy or these other

54:20

big open questions in physics, where we're

54:22

like, well, wait a minute, where is 90% of the matter in physics?

54:24

We don't know, it's a big scientific open question.

54:28

So similarly in biology, there is

54:30

a big open scientific question. And that open

54:32

scientific question is

54:33

what exactly is consciousness?

54:36

What things have it, what things don't,

54:39

we don't have that scientific

54:41

information. There is no scientific consensus

54:43

about it. There are some leading hypotheses and fields

54:46

that you can lean on, but we just don't have the

54:48

answer to that. Right. So

54:50

I personally doubt that any of these

54:52

large language models, that there's anything it's like

54:55

to be them. I doubt that they are conscious.

54:57

But we have no scientific consensus

54:59

to go back on. But the point is that

55:02

we're in a very different

55:05

epistemological standpoint when it comes to intelligence.

55:08

We do have a good understanding of intelligence. It's

55:10

a much more obvious concept because it's a much

55:12

more functional concept. We can just give

55:14

this thing SAT questions,

55:16

and we do, and it gets them right a

55:18

lot of the time. There are all sorts

55:20

of language benchmarks that these researchers

55:23

use that include things like SAT questions, and it scores

55:25

pretty well. and passes the bar exam,

55:27

and which is a great straight line for

55:30

a

55:30

larger, which we won't make, carry on.

55:33

Yeah,

55:35

and so, regardless of whether

55:38

or not you have any opinion about whether it is,

55:41

there is something it is like to be these networks, whether

55:43

or not they really have cognition,

55:45

quote unquote, whether or not they really have understanding,

55:48

quote unquote, whether or not they really have consciousness,

55:51

quote unquote,

55:53

the one thing that they definitely are that

55:55

sort of undebatable is intelligent to

55:57

some degree and they're only going to get more.

56:00

intelligent over time. And that's

56:02

the thing that makes them dangerous. In fact, it

56:04

might be even worse from sort of a

56:06

very broad metaphysical conception

56:10

if they are truly completely unconscious

56:12

and have no real understanding and have

56:14

no real cognition that's anything like a human.

56:18

Because in the end, if in 200

56:20

years, the Earth is just these AIs

56:23

going about their inscrutable mechanical

56:25

goals, we will have extinguished,

56:28

you know, the light of consciousness from

56:30

the universe because, you know, we wanted to make

56:32

a buck when it came to stock options.

56:35

Yeah,

56:35

that's a dirty thought. I guess that's the zombie

56:39

model, right? It's, I

56:44

can't get over the fact how these

56:46

human and mechanical metaphors

56:49

merge in your mind, in one's mind, And

56:53

how hard it's going to be to tell them apart from

56:55

actual humans. You

56:57

know, one of the great

56:59

observations of philosophy is I don't know whether you're

57:01

another human being like I am. My working assumption

57:04

is that you're something like me. And

57:08

I really don't have any evidence for that.

57:09

And we sort of, you know, it's called

57:12

solipsism. I don't know if I'm the only conscious

57:14

mind in the universe. And that

57:16

problem is just getting a lot bigger

57:19

right now. We're living,

57:21

what this conversation suggests to me and

57:23

the writing you've done on it so far is that

57:27

this really is a watershed moment in

57:30

our

57:30

existence on the planet. That

57:33

sounds a little bit, just a titch dramatic, but

57:38

I'm not sure that's wrong. I

57:41

think that very well could be right. I don't think it's

57:43

dramatic. And I'll be upfront about the fact

57:45

that I used to be very much

57:47

an AI skeptic because I went,

57:50

you know, I studied cognitive

57:52

science, I went into neuroscience of consciousness.

57:56

You know, I had,

57:57

I was paying attention to AI at the

57:59

time when I...

58:00

did this. And

58:01

AI was, I'll be very frank about it, academically, 15

58:05

years ago, AI was a joke. AI

58:08

was a complete joke. It never went

58:10

anywhere. People couldn't figure out anything

58:12

to do with it. All my professors said, don't

58:15

go into AI. It's been a dead field for 60

58:17

years. We've made no progress, right?

58:20

All the things like beating humans at chess

58:22

and so on, it's all just done because the

58:25

chess game board is so small, there's so many

58:27

limited moves, and we really can basically do a

58:29

a big lookup table, all sorts of things

58:31

like that.

58:34

And, and, and,

58:36

but the deep learning revolution was a real thing.

58:39

It was a real thing that we, we

58:41

figured out how to stack

58:44

and train these artificial neural networks in ways

58:46

that were incredibly effective. And when

58:48

go when, when, you know,

58:50

the first real triumph of it was beating

58:53

the best human being, I think his name is

58:55

Lee Soto, I hope I'm not mispronouncing it.

58:57

in 2016, that we finally,

59:02

AI finally beat a human being at Go, and

59:05

Go just can't be number crunched

59:07

in the way that chess can. And it was this

59:10

and within, you know, seven years after

59:12

that, we now have human beings where they're

59:14

generating text transcripts so

59:17

good that you're right, it sounds like the rest of

59:19

the New York Times. And that just happened

59:21

in seven years. And fundamentally,

59:23

the deep learning revolution and the way that

59:25

Again, the black box way that these

59:27

AIs are trained means that

59:30

our technological progress on

59:32

AI has suddenly rapidly

59:35

outstripped our understanding

59:38

of things like minds or consciousness

59:40

or even how to control and understand

59:43

big complex black boxes. So

59:45

it's like we've jumped ahead technologically.

59:48

And it's not so much that, you know, if we had a really good

59:50

understanding of how neural networks worked, like really

59:53

fundamentally solid ways to make them

59:55

crystal clear. And we had a really good understanding

59:57

of how the human brain generated consciousness.

1:00:00

and how it worked at a broad level, then

1:00:03

maybe

1:00:04

we could first of all answer all sorts of more ethical

1:00:07

questions about AI. We could control it very

1:00:09

well. We could, you know, decide

1:00:11

plenty of things about it.

1:00:13

But

1:00:14

our ability to make intelligence has so

1:00:16

drastically outstripped our

1:00:19

progress on those other areas, which has been slow.

1:00:22

And in some cases has just turned along

1:00:24

for decades without making any progress and so on.

1:00:27

I just want to reference a recent episode with Patrick House

1:00:29

on consciousness that

1:00:31

I think talks about these issues in a very,

1:00:34

this book does in a very thoughtful way.

1:00:38

So let me give you a

1:00:40

scenario we take.

1:00:43

We have a conference on AI, where all the greatest researchers

1:00:45

in the world are there. The academic ones,

1:00:48

the ones at Microsoft, the ones at Google, and

1:00:51

that small startup that's doing

1:00:53

something really terrifying that

1:00:56

we

1:00:56

don't even know the name of it. And there's

1:00:59

one conference hall, and while they're all there, maybe

1:01:01

it's a football stadium. How many are there?

1:01:08

I think probably less in terms of really top people. I think

1:01:10

there's probably less than 1,000. Okay, let's take the top 1,000. We're

1:01:12

in a big auditorium. And we lock

1:01:15

the doors, and

1:01:17

I guess we're nice to them. We

1:01:19

heard them at gunpoint onto a spaceship and

1:01:21

sent them off into the rest of the universe. We give a lot

1:01:23

of servers and stuff to play with while they're heading

1:01:26

out there. their days are

1:01:28

numbered, their impact on the earth is over, they're

1:01:30

gone, and

1:01:31

it's a really bad incentive

1:01:33

for future AI people. That's

1:01:36

not going to happen. So one of

1:01:38

the responses

1:01:40

to these kinds of problems,

1:01:43

whether it's, I don't want to call it a problem, these

1:01:45

kind of so-called science fiction

1:01:47

technological innovations

1:01:49

is, well, you can't really stop

1:01:52

it, Eric. You could talk about all you want, regulation,

1:01:54

and you're going to stop the human I feel this

1:01:56

is pretty strong in myself

1:01:58

so I'm making fun.

1:02:00

but I do kind of feel it.

1:02:02

The human being strives to understand

1:02:04

and

1:02:04

I don't think we're

1:02:06

just into avoiding surprises and spreading

1:02:09

our genes. I think we really like to understand

1:02:11

that where we live in, we want to matter. We

1:02:13

have a lot of other, as you say, drives and complexities.

1:02:17

So

1:02:17

it seems to be implausible that

1:02:20

we can stop this. So

1:02:23

the desire to expand

1:02:25

it, to make it better, make it smarter,

1:02:28

just like we can't, it happens everywhere. It's the essence

1:02:30

of human life over the last few

1:02:32

hundred years. Better, faster,

1:02:35

quicker, cheaper,

1:02:37

richer, name it, you name it.

1:02:39

So what's imaginable

1:02:41

for someone like yourself who wrote a very,

1:02:44

you know,

1:02:44

we're having a civilized, normal conversation here, but if

1:02:46

you go back to read your essay,

1:02:48

it's

1:02:50

your very

1:02:52

worked up. It's a screed, it's a rant,

1:02:54

and it's a rant that you you justify because you think perhaps

1:02:57

something like the future of the human race is at stake.

1:03:00

And if that's true, you

1:03:02

should take it very seriously. You should just go, they'll

1:03:04

probably figure it out or whatever. So

1:03:07

what should a thoughtful person

1:03:09

who's worried about this advocate

1:03:12

for? Because they're not going to put a herd of them

1:03:14

onto the spaceship. They're not going to burn the building

1:03:16

down while they're in it. Not going to happen.

1:03:18

Yeah. Oh yeah, absolutely. And I would never advocate

1:03:21

for anything like that. Didn't make a suggestion. Sorry about that. But

1:03:26

you know, you called it a screed and

1:03:28

there's a certain sense I agree because I'm very

1:03:30

open about that it's a call to activism.

1:03:33

And in order to get human beings,

1:03:36

you know, again, like as a polity,

1:03:38

like as a nation to do anything,

1:03:41

you have to have sort of wild

1:03:43

levels of enthusiasm and motivation.

1:03:46

Right. And you can look at anything

1:03:48

from nuclear disarmament activism

1:03:51

to climate change activism and see that

1:03:53

there's plenty of people within those

1:03:55

movements who catastrophize. And

1:03:58

there's, you know, you can, you can

1:04:00

certainly say that at an individual level, that

1:04:02

can be bad where people are not appropriately

1:04:05

rating the threat. But there's another sense

1:04:08

in which if human beings don't get worked up

1:04:10

about something,

1:04:11

we don't do anything about it. This is very natural

1:04:13

for us. We just let stuff evolve as it

1:04:19

is. And so what I want is for a lot

1:04:21

of the people who are in AI safety to be

1:04:23

very honest

1:04:25

about how scared

1:04:27

they are about various aspects of this technology

1:04:30

because I do think that in the end, the

1:04:32

net trickle-down effect will be

1:04:34

good because it will eventually push for some

1:04:37

form of regulation or oversight.

1:04:39

And in some sense, it already has. I want

1:04:42

to be clear about that. I think that there's a sense

1:04:44

in which just what happened with Sydney, which

1:04:47

was such big news, it was all over

1:04:49

Twitter,

1:04:50

has made companies

1:04:52

take this notion of AI safety and

1:04:54

this notion of controllability probably a lot more seriously.

1:04:57

There is social pressure

1:04:59

for companies. In fact, there's an argument that social

1:05:01

pressure on companies is what companies are most

1:05:04

responsive to. Most companies do things.

1:05:07

They change their product. They do all sorts of things

1:05:09

just because they want to be liked and they don't want to have anyone

1:05:11

yell at them. And that's one of their main incentives.

1:05:16

And I do think that I personally

1:05:19

am not at all worried about

1:05:22

AI being built by someone

1:05:24

in the middle of nowhere, right? People

1:05:27

always say something like this, that, well, if

1:05:30

we over-regulate it in the United States,

1:05:32

North Korea will build it or something

1:05:34

like that. And the capacity

1:05:37

is just not there.

1:05:39

It is exactly

1:05:41

like

1:05:42

nuclear weapons in this sense.

1:05:43

Real serious progress in AI

1:05:46

is probably relegated, I don't even think it's going

1:05:48

to be startups. And people have been talking about this,

1:05:50

that

1:05:51

the big competitors in this

1:05:53

space are the only ones with the access to the data

1:05:55

and the talent and the money to jump into

1:05:57

it. So it's going to be Microsoft. It's going to be Google.

1:06:00

It's going to be Facebook. It's going to be names

1:06:02

that we know. And there's only 10, at most,

1:06:06

you could say there's 10 of those companies. There might only be three

1:06:09

of those companies. And then they might only

1:06:11

employ a couple hundred at

1:06:13

most sort of overall employees. That

1:06:15

is a sector of the economy that you

1:06:18

can do something about. And again, I don't

1:06:21

suggest going in there and burning the servers or something.

1:06:23

Right. But you could very easily

1:06:25

have all sorts of benchmarks

1:06:28

that people have to have

1:06:30

to meet. You could also do things like

1:06:33

have people sign on, maybe

1:06:35

voluntarily, maybe voluntarily

1:06:37

sort of under the condition of pressure and so on,

1:06:40

to not make AIs that are

1:06:42

significantly smarter than any living human.

1:06:44

They could be more general, right?

1:06:46

So they could make great search engines. Because what do you

1:06:48

need for a great search engine to make a lot of money the way these

1:06:50

companies make? You need something that can give a good

1:06:53

answer to a lot of questions. And

1:06:55

I don't think that's something that can give just a good answer

1:06:57

to a lot of questions is very dangerous

1:07:00

to the human race, especially if there's

1:07:03

just a few of them and they're all kept

1:07:06

under control by Microsoft and Google

1:07:08

and so on. But you

1:07:10

could say, listen, what we don't want to

1:07:12

have is some really big cognitive

1:07:15

benchmark.

1:07:16

And we don't want this thing to

1:07:18

do better than any human on

1:07:20

all the parts of

1:07:21

it. And we just say, we don't, that thing

1:07:24

is a dangerous and weird entity

1:07:26

and we don't know how to control it, we don't know how to use

1:07:28

it and so on. And you could literally

1:07:30

imagine just giving this test to

1:07:33

the next generation of AIs and people in

1:07:35

the companies give this test and they just make sure that this

1:07:37

thing never gets so smart that it blows every

1:07:39

human being in the world out of the water. Eric, you're so

1:07:41

naive.

1:07:43

You're telling me they couldn't train it to

1:07:45

do badland on the test. I

1:07:47

mean, seriously, I don't, I'm teasing

1:07:49

about being naive, But I think

1:07:53

there are two ways to think.

1:07:54

There's three ways maybe to think about regulating this that

1:07:57

might be effective. One way is to limit the

1:07:59

size of a corporation.

1:08:00

which is a repugnant thought to me, but

1:08:02

if I thought the human race was at stake, maybe I'd consider

1:08:04

it. The second would be

1:08:06

to do the kind of

1:08:10

standard types of regulation that we think

1:08:12

of in other areas. If this is toxic, you

1:08:14

can't put it out. If it's toxic, you get a fine.

1:08:16

If it's right, etc. The

1:08:18

third way, which I think is never

1:08:20

going to happen, but it speaks to

1:08:22

me, as listeners will know, was

1:08:25

seeing me for a long time, you'd think

1:08:27

that if you were working on this and you thought it could destroy

1:08:29

the human race, you'd maybe want to think about doing something different.

1:08:33

And you'd give up the urge to be the

1:08:35

greatest AI inventor of all time.

1:08:37

And you'd say, this is, you

1:08:39

know,

1:08:40

I just stepped to see a tweet today, Robert Oppenheimer

1:08:43

went into Truman and said, I have,

1:08:45

Robert Oppenheimer, haven't worked on the Los

1:08:48

Alamos project. It was an important figure in the

1:08:50

development of the atomic bomb, told Truman, I have blood on my

1:08:52

hands and Truman was disgusted by him because

1:08:54

he said, I made that decision that you,

1:08:57

you create called McCreaton. I don't

1:08:58

know if that's a literal, accurate quote or not.

1:09:04

You'd think people would

1:09:06

want to restrain their urge to do to

1:09:08

find poisons, but that's never been a part

1:09:11

of the human condition. We

1:09:13

want to find everything we find poison. That's why

1:09:16

we have lab leaks. That's why we

1:09:18

have weapons that we have that

1:09:21

are unimaginably destructive. Now,

1:09:24

we don't keep making more and more destructive weapons

1:09:26

as far as we know. That's an interesting

1:09:28

parallel. There is a sort of limit

1:09:31

on the magnitude, the megatonnage

1:09:34

of nuclear weapons. Maybe that's a

1:09:36

sub... I don't know how you'd enforce it though. What

1:09:39

are your thoughts? An

1:09:42

example, I think

1:09:43

one issue with arguing for AI safety

1:09:49

is that people sort of want

1:09:52

at the outset, and it's a very natural one, some sort

1:09:54

of perfect plan, where it's, okay,

1:09:57

we're just gonna implement this plan and it's gonna work really,

1:09:59

really well. And I think it's going to be much more like,

1:10:03

it's not going to be exactly like nuclear weapons or

1:10:05

nor exactly the climate change. It'll be like some third

1:10:07

other thing that we as a civilization have to deal with,

1:10:10

with its own sort of dynamics. But

1:10:12

ultimately in none of those cases,

1:10:14

was there some sort of initial proposal and we just

1:10:16

had to follow this proposal. Instead, everyone

1:10:19

sort of had to recognize that

1:10:21

it's a threat, again, to some degree, you can

1:10:23

have all sorts of debates about it, but clearly I don't

1:10:26

think anyone is just like, well,

1:10:28

let's just get all the fossil fuels and burn

1:10:30

them all, right? Like, I think that that's a very

1:10:33

sort of rare position.

1:10:35

And the reason is, wherever it's most people recognize that, hey,

1:10:37

that's probably not gonna be a good idea. It might not be a good idea

1:10:39

globally. It certainly won't be a good idea locally. And

1:10:43

through public pressure, we've managed to

1:10:45

relatively contain some of the big existential

1:10:48

threats that we face as a civilization.

1:10:50

And a great example are lab

1:10:53

leaks. I personally think, yeah, COVID

1:10:55

probably did come from a lab. But if you think

1:10:57

about all the labs, doing all sorts of research

1:10:59

all across the globe, it's actually pretty astounding

1:11:02

that we don't have lab leagues all the time

1:11:04

as people are using these viruses. So we

1:11:06

do sometimes do a middling

1:11:10

job and for big existential

1:11:12

threats, sometimes all you need is sort of

1:11:14

a middling job. You just need people, you just

1:11:17

need to have a lot of eyes on an industry

1:11:19

and people there to realize that they're being launched

1:11:21

and to go slowly and to sort of think about these

1:11:24

issues. You know, you don't need,

1:11:26

you know, I propose, oh, we'll just have like a

1:11:28

cognitive IQ test or something like that. I

1:11:30

would never think that that alone would

1:11:33

prevent these issues, but it could be part

1:11:35

of a big comprehensive plan of

1:11:38

public pressure and so on. And I think that that's

1:11:41

gonna work. And I think that it's unavoidable that

1:11:43

the public wants a say in this. I think

1:11:45

they read those chats, trans trips and they

1:11:47

go, what?

1:11:48

This is really high level stuff. There's

1:11:52

all sorts of moral concerns, ethical concerns,

1:11:54

and then yes, there are absolutely dangers.

1:11:58

And again, I think we're at the...

1:12:00

point in the movement, maybe we're a little bit late,

1:12:02

maybe AI safety should have started earlier. But again,

1:12:04

the deep learning revolution sort of caught everyone

1:12:06

by surprise. I still think we're relatively

1:12:09

early. I think that this is sort of like imagine

1:12:11

that you were, imagine that you personally

1:12:13

thought that climate change was going to be a really big

1:12:15

problem. And it's currently 1970.

1:12:18

I don't think it makes sense to then be like, okay,

1:12:20

well, we're just going to do carbon sequesterization.

1:12:22

And I know exactly the technology that's needed

1:12:24

for the carbon sequesterization. It's better

1:12:27

to just sort of go out there and protest and

1:12:29

make a big deal and get

1:12:31

it to be a public issue. That's

1:12:34

going to be a lot more of a convincing

1:12:36

and effective strategy than coming

1:12:39

up with some particular plan because it's always going to depend

1:12:41

on the technology and exactly who has it and exactly

1:12:43

how many people and

1:12:45

all sorts of things. So I think that that's the mode

1:12:48

that people who are concerned, like myself, about

1:12:50

AI safety should be in right now, which is just

1:12:53

public awareness that this could be a problem. can

1:12:55

decide personally to what degree they think

1:12:58

it will be a problem. But what I

1:13:00

think truly is

1:13:02

naive is saying there's absolutely

1:13:05

not going to be a problem. We're going to perfectly

1:13:07

be able to control these alien

1:13:09

and human intelligences and don't worry

1:13:11

at all about it.

1:13:13

It's the other thing that crossed my mind

1:13:16

is that the ability

1:13:18

of our political system to provide

1:13:21

thoughtful responses to existential

1:13:24

threats. Not so good.

1:13:26

And if anything, it

1:13:28

seems to me it's gonna get worse. And

1:13:31

part of the way it's gonna get worse is through this blurring

1:13:33

of the line between humans and

1:13:35

machines that

1:13:36

people are gonna have trouble

1:13:38

telling them apart. And

1:13:42

I'd like to think of something more optimistic.

1:13:44

So I'm going to give you a chance

1:13:46

to play chat GPT.

1:13:50

I'm going to say,

1:13:53

here's my prompt. What would

1:13:55

Sam Altman say about

1:13:57

all these worries? And

1:14:00

Sam Altman being the head of OpenAI that

1:14:03

just put out Chat GPT, former

1:14:05

EconTalk guest. You can go hear his thoughts when

1:14:07

he was head of the Y Combinator, long

1:14:10

time ago here in our archives. So just Google

1:14:12

Altman

1:14:13

EconTalk and you'll find that conversation.

1:14:16

But Sam is

1:14:18

a nice guy.

1:14:21

I like him. He's likable. But

1:14:26

I'm not sure I.

1:14:28

I'm not sure his level of worry is going to be the same

1:14:30

as yours. Deb

1:14:31

certainly has a different set of incentives.

1:14:34

But I think he'd start off by

1:14:36

saying, oh, you're exaggerating.

1:14:40

Scott Alexander recently wrote an essay

1:14:42

where he was alarmed at

1:14:45

some PR release that said, oh, yeah,

1:14:47

there's nothing really that big to worry about. It's

1:14:50

going to be okay. Don't pay any attention to that chat,

1:14:52

GBT, behind the curtain. So I'm

1:14:55

curious what you think,

1:14:57

steel man the opposition here, if you can,

1:15:00

for a minute, Eric. Someone

1:15:03

likes Sam. What would he say?

1:15:05

Well, here's a direct quote from

1:15:07

Sam Altman, who said, AI will

1:15:10

probably most likely lead to the end of the world,

1:15:12

but in the meantime, there'll be some great

1:15:15

companies.

1:15:16

So that's a direct quote. What did he mean? Was

1:15:18

that tongue in cheek, perhaps?

1:15:20

I

1:15:22

haven't looked into the... exact electronics

1:15:25

as opposed. But I don't

1:15:27

know. I honestly, I don't. I know that Sam has

1:15:29

been concerned about AI safety. So this is not

1:15:32

completely tongue in cheek. He has been,

1:15:34

I know for

1:15:35

a fact that he's been concerned about this. Many

1:15:38

of the people who started the initial companies were concerned

1:15:40

about this. At the beginning of open AI, it started

1:15:42

to address concerns around AI safety. There

1:15:45

was something called the open letter on artificial

1:15:47

intelligence that Stephen Hawking, Elon Musk, a

1:15:50

lot of the people who provided funding for for open

1:15:52

AI wrote and in it

1:15:54

they talk

1:15:55

about how AI is an, about

1:15:57

how AI could be an existential.

1:16:00

threat. So this is not

1:16:02

some sort of radical outside opinion.

1:16:04

I think it's something that

1:16:07

that sites someone

1:16:09

like Sam Altman knows now if I'm going to sort

1:16:11

of steel man his position, it goes something

1:16:14

like,

1:16:15

well, I'm concerned about this, I

1:16:17

said that AI will probably most likely lead to the end of

1:16:19

the world. So I'm concerned about this. So I

1:16:22

should be the one to do it. Because if

1:16:24

someone else who's more reckless does it, like, it's

1:16:26

going to be done, if someone else who's more reckless

1:16:28

does it, then maybe I can provide

1:16:32

some sort of guardrails and do it in

1:16:34

as safe a manner as possible. And I

1:16:36

really hope that that would be his

1:16:39

motivation. And if so, that's a great

1:16:41

and honorable motivation.

1:16:45

But at the same time, that does not enter

1:16:47

someone from criticism.

1:16:49

I mean, I think that in many ways,

1:16:53

Sam Altman is now

1:16:55

doing something very similar to what Sam

1:16:58

Bankman-Fried, who was

1:17:00

the one who sort of plunged FTX into

1:17:03

chaos was doing, whereas their reasoning

1:17:06

via this expected value in

1:17:08

this expected value way, where Sam Bankman-Fried

1:17:10

said, listen, the more billions I create,

1:17:13

the more I can donate to charity, there's sort of no

1:17:15

upper bound, I might as well be as financially risky

1:17:17

as possible, because the expected value

1:17:20

of my outcome is going to be so

1:17:22

high, right? Even though there's this huge downside.

1:17:25

I think Sam Ullman probably reads it the exact same way

1:17:27

when it comes to AI. I think he thinks, listen, if we

1:17:29

can make these highly intelligent things, we can

1:17:31

have all this glorious future, all our problems

1:17:33

are going to be solved,

1:17:34

right? They're going to cure cancer, they're going to do

1:17:37

all this stuff for us, and the

1:17:40

benefits

1:17:41

outweigh the risk. But

1:17:43

most people, when they look at an equation

1:17:46

like that, all they see is the existential

1:17:48

risk. They don't see, okay, oh, So it's expected

1:17:51

to be positive.

1:17:52

They see, no, we

1:17:54

can in one day maybe cure

1:17:56

cancer ourselves, We might not

1:17:58

need these systems to tell.

1:18:00

an amazing future. And

1:18:03

they're just they might just not be worth

1:18:06

the level of risk.

1:18:11

Well you and I are skeptical about utilitarianism.

1:18:15

Nassem Talab

1:18:16

and I and I suspect you understand

1:18:20

that expected value is a really bad

1:18:22

way to define rationality or how

1:18:24

to live. Nassem

1:18:26

always points out,

1:18:28

you

1:18:29

got to stay in the game. You

1:18:31

want to avoid the goal is not to maximize

1:18:34

the expected value. The goal in these kinds of situations

1:18:37

is to avoid ruin. Ruined

1:18:39

in this case would be the extinction of the human race. Now

1:18:41

there is a view that

1:18:43

says, what's the big deal? It's just the next week.

1:18:46

It's us, by the way. We

1:18:48

built it.

1:18:49

It learns off of all of human creativity

1:18:52

and sentences and

1:18:54

words and music and art. And

1:18:56

so it's just the next level

1:18:58

of us.

1:19:00

And for the first time in this conversation, I'll

1:19:02

mention the word God, the concept of God. If

1:19:04

you're a believing person,

1:19:06

as I am in some dimension, I take

1:19:08

the idea of God seriously. You believe that human

1:19:10

beings have a special role to play in the world, and

1:19:13

being supplanted by something,

1:19:15

quote, better is not

1:19:17

a goal. But I think there are people in the

1:19:19

industry

1:19:20

probably don't feel that way and they're not even worried

1:19:22

about it. The extension at the end of the human species

1:19:26

is no different than the end of those other

1:19:28

nine

1:19:29

cousins we had in the in the in

1:19:31

the in the belt

1:19:33

when

1:19:35

we

1:19:36

extinguished them, exterminated them through combination

1:19:38

of murder and outcompeting them.

1:19:42

Yes and I think

1:19:44

that

1:19:46

there's also a sense which as I said,

1:19:48

it might be a horrific future because

1:19:50

maybe these things really aren't conscious at all,

1:19:53

right? So it might be one of the worst possible futures

1:19:55

you can ever imagine.

1:19:57

Although I do think that there is

1:20:00

I

1:20:00

think that opinions like that,

1:20:02

which are fun, sort

1:20:05

of sci-fi things to talk about, have been

1:20:07

acceptable because there's never

1:20:09

actually any risk, right? So

1:20:11

my metaphor is that, you know, if you make

1:20:14

up your own religion and you decide to worship Zannon,

1:20:16

Supreme Dark Lord of the Galaxy, it's just

1:20:18

like a funny thing to talk about at parties.

1:20:21

But when Zannon's first messengers prop

1:20:24

up, something like it's not funny,

1:20:26

right? It's something horrific that you actually

1:20:28

hold these views. And

1:20:31

so I suspect that while there are some

1:20:33

people out there on Twitter or

1:20:35

the only people who sort of convince themselves

1:20:37

of things like this are like intellectuals, right? That

1:20:41

actually would be better if the human race was destroyed

1:20:43

and supplanted by AIs. I think that

1:20:46

sort of the public generally is not

1:20:48

gonna give much

1:20:51

thrift to those sort of things. People

1:20:53

have kids.

1:20:55

There may not even like the idea

1:20:57

of there being entities. I

1:20:59

mean, even I am uncomfortable with

1:21:01

the fact that my children are gonna grow up in

1:21:04

a world where it is very possible that there

1:21:06

are entities that are not just human

1:21:08

beings. Everyone knows there are people who are smarter

1:21:10

than you, right, at various different things, but everyone

1:21:12

also has all their own things that they

1:21:14

themselves are good at or that they value or

1:21:17

that they contribute to, right,

1:21:19

as human beings. And so everyone sort of has this

1:21:22

inner worth, even though you know you can go to a university

1:21:25

and find someone who might be smarter

1:21:27

than you across their domain of expertise

1:21:29

or whatever.

1:21:31

We do not know what it's like to live in

1:21:33

a world where there are entities that are so vastly

1:21:35

smarter than you that they just effectively

1:21:38

surpass you at everything.

1:21:40

That means that they can have a conversation

1:21:42

that's more empathetic than you can ever

1:21:44

have because they're just smarter and they can just mimic

1:21:47

empathy. That means we

1:21:49

don't know what it's even like live in a world like

1:21:51

that, even if everything goes

1:21:54

well and these things don't

1:21:57

turn on us or destroy us. sort

1:22:00

of nothing bad happens, it might

1:22:02

be a

1:22:03

minimization of

1:22:05

human beings. And again, this goes to the fact

1:22:07

that this technology has no

1:22:10

historical analog.

1:22:11

People will sometimes say,

1:22:13

oh, this is like the Luddites or

1:22:16

some other anti-technology group.

1:22:19

And the simple truth is that that was

1:22:21

about the automation of jobs.

1:22:25

And we were making machines that

1:22:27

had a greater strength or dexterity

1:22:29

than humans. But that's just not a problem because

1:22:32

we didn't conquer the world through our strength and dexterity.

1:22:34

We conquered the world through our intelligence. We've

1:22:37

never made machines that are smarter than

1:22:39

human beings. We just don't

1:22:41

know how we'll relate to something like

1:22:43

that and what it will

1:22:45

mean for us if and when we do

1:22:48

it. And so in that sense,

1:22:50

this just can't be compared to any

1:22:53

other form of, oh, you're worried about

1:22:55

job loss or automation or something like that,

1:22:57

right? that is replacing

1:23:00

tasks and that's

1:23:02

replacing strength and that's replacing dexterity, but

1:23:04

those aren't our fundamental attributes. Our fundamental attributes

1:23:06

are intelligence. And

1:23:09

when you have something that's much smarter

1:23:11

than a human being,

1:23:12

it's very similar to how wildlife

1:23:14

lives around humans. It's similar in their relationship,

1:23:17

right? A human

1:23:19

might treat wildlife well.

1:23:22

Like recently, I found an injured

1:23:24

bunny, right? And I sort of felt

1:23:26

very attached to it because it was right outside my door. And

1:23:28

I was like, well, I'm, you're sort of my responsibility

1:23:30

now, right? And so I had to call, you know, animal

1:23:33

rehabilitation. I was like, wonderful for this bunny,

1:23:35

right? And then I went home and I like ate,

1:23:38

you know, like a pizza with pork

1:23:40

on it, right? Like, you

1:23:43

know, things that are vastly more intelligent

1:23:45

than you are really hard to understand

1:23:48

and predict.

1:23:50

And the wildlife next door, as much as we might

1:23:52

like it, we'll also build a parking lot over

1:23:55

it at a heartbeat and they'll never know why.

1:23:57

They'll never know why. It's totally beyond their...

1:24:00

So when you live on a planet next

1:24:02

to things that are far vastly

1:24:04

smarter than you or anyone else, they

1:24:09

are the humans in that scenario. They might just build a parking

1:24:11

lot over us and we will never, ever know

1:24:13

why.

1:24:16

I guess today has been

1:24:17

Eric Hov. Eric, thanks for being

1:24:19

part of eConTalk.

1:24:22

Thank you so much for us. It's a pleasure to be back on.

1:24:30

This is Econ Talk, part of the Library of Economics

1:24:32

and Liberty. For more Econ Talk, go to econtalk.org,

1:24:36

where you can also comment on today's podcast

1:24:38

and find links and readings related to today's

1:24:40

conversation. The sound engineer

1:24:43

for Econ Talk is Rich Goyette. I'm

1:24:45

your host, Russ Roberts. Thanks for

1:24:47

listening.

1:24:48

talk to you on Monday.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features