Podchaser Logo
Home
Peter Railton on AI and Ethics

Peter Railton on AI and Ethics

Released Friday, 1st July 2022
Good episode? Give it some love!
Peter Railton on AI and Ethics

Peter Railton on AI and Ethics

Peter Railton on AI and Ethics

Peter Railton on AI and Ethics

Friday, 1st July 2022
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:02

this is philosophy bites with me,

0:04

not to wolverton & me, david

0:06

edmonds if you enjoyed, please,

0:09

purchase with conium

0:11

funded through donations, would be gratefully

0:13

received for for details go to www dot philosophy bites dot-com humans

0:19

are creating artificial intelligence machines

0:22

and systems that use a i already

0:24

making decisions that our lives in

0:26

ways so,

0:27

how do we build ethics

0:29

into machines and

0:31

mike we learn something about ethics from

0:33

a ethics peter wales

0:35

and is a philosopher at the university of michigan

0:38

rosen welcome to foresee bites

0:41

thank you very much very glad to be the

0:43

topic we're going to focus on is a

0:45

i and ethics i'm

0:47

not actually sure about ethics bullets

0:50

get clear about ai first what are we talking about

0:52

a eyes a very general term

0:54

well for thing for me to say is i'm not an expert

0:57

on a i i'm a philosopher

0:59

who's worked in moral philosophy and through

1:01

believes that there are many ethical challenges

1:03

raised by a i and that they're

1:06

coming onto us much quicker than we expected

1:08

they would and so it's important

1:10

it's think for philosophers who art expert naias

1:13

to try to do what we can to scramble

1:15

the

1:16

the i literally artificial intelligence

1:18

what kind of intelligence is relevant

1:21

most people i think would say that the most

1:23

general idea of intelligence is this capacity

1:26

to solve problems that , mean

1:28

for example capacity to take what

1:30

has been learned in the past and solved new

1:32

problems problems now i think a lot

1:34

of people would include learning and intuitions

1:37

a data banks just sitting there is not by

1:39

itself into isn't because is not thinking

1:41

of thinking and it's not seeing

1:43

how they work out when learning from whether

1:45

they did workouts so that systems

1:48

that were mostly thinking about now are ai

1:50

systems that do learn in fact learning

1:52

is their fundamental principle their operation

1:55

and when i first started to see such

1:57

systems a couple of years ago i

1:59

was

1:59

the only excited because i think

2:02

that these actually have a chance of

2:04

being something like that human intuitive

2:07

knowledge system obviously there are

2:09

modeled on neurons in

2:11

neural networks in the mind but they aren't strictly

2:13

like those and no one would say

2:16

that were anything like level of reaching something

2:18

like mind but it does

2:20

seem to me that compared to the old

2:22

method the i which

2:24

was a method that was based on sort of logical

2:26

processing and essentially

2:29

required human experts to

2:31

program that knowledge into the machine

2:34

and then you'd rely on the machines power

2:36

and memory and so on to crank through

2:38

possibilities but the machine

2:40

itself was problem solving only and that

2:42

instrumental sense what's exciting

2:45

now is that this machines are

2:47

learning things things that

2:49

we didn't know

2:51

and are able to advance

2:53

the way that we think about a range of issues

2:56

could you give an example of the kind of thing yourself most

2:58

i guess most famous example is the distinction

3:01

between something like deep blue

3:03

the chess playing program that allegedly

3:05

be test for an alpha

3:07

go or alpha zero

3:09

now or muse euro for says next

3:12

phase deep blue was

3:14

programmed by experts just

3:16

experts just experts thus systems

3:18

and so on over many years

3:21

it was, in fact a machine that didn't

3:23

itself beat kasparov would

3:25

be casper, obviously human information and

3:28

knowledge that was put into the machine and

3:30

it turn the crank gave moves

3:32

it wouldn't really be fair to say that

3:35

mushy intelligence defeated

3:37

a human a system like alpha go

3:40

starts with no knowledge of go strategy

3:42

only the rules of the game

3:44

and users simulation playing

3:46

against itself

3:48

in developing it's own method of evaluating

3:51

simulations the only learning signal

3:53

it gets is whether it wins again or not

3:55

and through that developed not only

3:58

discovered some known strike it is

4:00

ago but invented new strategies and

4:03

was able to defeat of world champion mortuary

4:06

fast but he was able to do this not

4:08

with years and years of training and expertise

4:10

that through this process or reinforcement

4:13

learning this kind of learning machine

4:16

of as he has huge potential but

4:18

it also has also quality which seems

4:21

to be getting close to the human mind human

4:23

mind respects and obviously

4:25

we hold individual human

4:27

beings responsible for all kinds of things doing

4:29

this just a world in which will start

4:31

to hold machines responsible

4:33

for thing that's a good question because

4:36

responsibilities else very thorny issue

4:38

there are various kinds of responsibility but

4:40

one question is whether we will

4:43

allow them allow certain amount of autonomy

4:45

in decision making

4:46

that's in effect recognizing that weekend

4:48

interest certain decisions with them and

4:51

that's allowing them allowing certain responsibility

4:54

does that mean we think of them as moral agents

4:56

know because for moral agency

4:58

requires a lot more than just the ability

5:01

to make decisions we think of it as involving

5:03

for example self reflection consciousness

5:06

capacity for morley motion and

5:09

nothing like us to these machines but

5:12

they will be given responsibility

5:14

in the question is

5:15

are we entitled or can we be

5:17

justified in giving them these responsibilities

5:20

what would they have to be like for that me

5:22

result and ,

5:24

speaking the future but they're already been given

5:26

lots of responsibilities responsibilities

5:29

this means that be perfect sarah

5:31

seriously very quickly about

5:33

the question of okay well in what

5:35

sense might these machines be themselves

5:38

capable of responding to a morally relevant

5:40

features of situations i can

5:42

imagine prison for a machine

5:44

that runs an automated heart

5:47

operation that might actually be

5:50

more responsive to subtle changes in the

5:52

body as as impressions going on better

5:54

processing data that's coming three

5:56

more accurate and is cussing than a certain could

5:59

be the you have this machine

6:01

i doing a heart operation

6:03

on a very vulnerable person we've

6:06

, our trust in that if the machine

6:08

does the job better them certain it seems to

6:10

me what's the moral issue that

6:13

good these machines because they're learning

6:15

machines they don't have

6:17

first principles that were given to them their

6:19

first principle so to speak come from the data

6:21

income from experience and so

6:24

a machine that does doing that won't

6:26

necessarily have the knowledge a

6:28

surgeon would have of certain signs

6:31

as the patient is not doing well he

6:33

could make a mistake not because

6:36

it's a poor however or because

6:38

it's for monitoring these vital signals that

6:40

because there are other signals and patience

6:43

or stations body might give that something's

6:45

going wrong we don't know in advance

6:47

which those kinds of false are going to be we've

6:50

already found that there are cases

6:52

for example with self driving vehicles

6:55

where they will see a scene which

6:57

we would clearly identify as involving

6:59

a human being let's say and not see

7:01

the human because of the reflectance of the light

7:03

or something like that and so

7:05

they will make mistakes as

7:08

any learning agent will

7:10

the mistakes want me the same ones perhaps

7:12

that we would make and we won't be able to anticipate

7:14

all of them we can try begin read many

7:16

simulations but the question

7:18

is it in what sense are we

7:21

justified if the kinds of mistakes

7:23

they could make her not just a technical mistakes

7:25

but a ethical mistake the

7:28

subtle difference between a technical mistake

7:30

and unethical mistake consider

7:33

a program that is

7:35

learning to make good decisions

7:37

about the amount of parole

7:40

someone should have

7:41

that machine is going to take in all manner of variables

7:44

demographic variables included some

7:46

, those will concerned things like race and gender

7:48

and so on and we might

7:51

think that it's inadmissible to

7:53

take someone's race or gender into account

7:55

when making account when like this but

7:57

if it's statistically correlated and

8:00

maybe for of complex sociological reason

8:02

statistically correlated with recidivism

8:04

repeat offense then the

8:06

machine will use that information that's

8:09

not a technical mistake

8:11

because it is predicting accurately

8:13

using that information but we might

8:15

think it's immoral mistake it's an injustice so

8:18

the question is given that the machine is harvesting

8:21

all this information and that there will

8:23

be the statistical relevance relations are

8:25

there some which we think it would be impermissible

8:28

to use we have some idea that

8:30

in no law and ethics but

8:32

that isn't something that is automatically discovered

8:35

just having a challenge of making sure that

8:37

you make recommendations that are statistically

8:39

as likely as possible to

8:41

predict who's going to commit a new crime

8:44

or not or how soon they can be trusted so

8:46

does not mean that you have to then find

8:49

a way of building in some kind of ethical

8:51

imitation some kind of cut

8:54

out switch is it worth that doesn't allow

8:56

the machine to act in those sorts of cases

8:59

yes one model of how we might

9:01

respond is well maybe we could build

9:03

in some ethics and

9:06

certainly you could build into machines

9:09

the rhodesians when using certain date now

9:12

because these are highly and capable at

9:14

a probabilistic learning it's not

9:16

all that evident always have the limit

9:18

what information they have because they can follow

9:21

information and project information and

9:23

simulate information and draw inferences

9:26

from that and then see whether those edwards's were

9:28

accurate and then imagine that information

9:30

for example the weather's this hidden variables that

9:32

i wasn't given but that predicts to recidivism

9:35

and it turns out that hidden variable is something

9:37

very much like race so

9:39

did it take race into account not

9:41

clear because what's happening is that the society's

9:44

set up and a discriminatory way and

9:46

, discriminations going to get translated

9:49

into the data and then modeling

9:52

that data as accurate me and pack the

9:54

as possible is going to discovered that there's

9:56

discovered that a latency here we might think

9:58

it's a latency that he can't use of as a know

9:59

are you can't look at that

10:01

but of course if that layton variable is

10:03

highly predictions and machine wealth and

10:06

said have to use it then

10:08

you have a hierarchy of machines we have the moral

10:10

messina the tells the then shuts down

10:12

the ones which are not behaving well

10:14

there's a big it was an interesting literature

10:16

on this question of the off switch for ai

10:19

systems one thing that's often said

10:21

is that as ai systems become agents

10:24

not just capable of modeling data and

10:26

suggesting hypotheses and so on but

10:29

capable of taking actions they will

10:31

have the features of agents bostick

10:33

estimates about the state of the world they

10:36

will have a value function of some

10:38

kind it might be goals it might be

10:40

if it's an animal it's a reward function

10:42

and they will also have a capacity

10:44

to imagine alternative courses

10:46

of action assign expectations

10:48

to them and make decisions correspondingly

10:51

so suppose we give a machine a particular goal

10:54

well one condition of achieving most

10:56

goals is that the machine continue to exist

10:59

and so therefore the machine gets a sub goal

11:02

of protecting his own existence now

11:04

how my to do that if we're tried to shut it off

11:07

well it might anticipate when

11:09

we would shut it off or look for signs of that

11:12

and replicate itself that has access

11:14

to the web let's say and it puts replicas

11:17

of itself out there in there web somewhere and somewhere we flip

11:19

the switch thinking return the machine off

11:21

but we've actually duplicated the machine that's

11:23

man nefarious on the machines part

11:26

it's thinking i can't achieve this goal with i

11:28

don't exist there will be elements

11:30

of what the machine is trying to do that

11:32

we aren't immediately aware of be don't immediately

11:35

understand and anticipating

11:37

those and you know be said well let's build a machine

11:40

that but despite that they're me to

11:42

ah but what about turning that machine off will

11:44

have to build a machine the when anticipates it's

11:46

actions and then him what what's his the top

11:49

what's the ethical master and the

11:51

thing is that ethics has not produced for

11:53

us and axiomatic says

11:55

we have capital series

11:57

which give us general principles but we

11:59

don't have

11:59

the axiomatic system from which you can deduce whether

12:02

given behaviors right or wrong and that's

12:04

true even for what look like the most axiomatic

12:06

system for utilitarianism for example

12:09

many utilitarian now our rule utilitarian

12:12

so the question is what would be that set of rules

12:14

which would have the best acceptance utility bill

12:16

what the heck is answer that question and

12:19

with we don't know the answer to that question that we don't know

12:21

was right or wrong maybe with

12:23

it has a capacity to model

12:25

once he would be like of society follow

12:27

certain rules that we don't have now for

12:30

example machines are capable of modeling financial

12:32

markets and figuring out what's going on and those

12:35

if we had we had that were capable

12:37

of learning not just a

12:39

particular area but more broadly about

12:42

tendencies in human behavior is about social

12:44

interactions and so on we ,

12:47

be able to gain some knowledge about well what might

12:49

this set of rules be like and what might be like if

12:51

they were internalized and foliage a certain

12:53

degree and so even for

12:55

the most dedicated rule utilitarian

12:58

there's going to be a lot about ethics they don't know the

13:00

machine might be able to know that

13:03

to some degree better than an ordinary human

13:05

agent we rely on intuition a great deal

13:07

and that's fine because it's

13:10

more complicated systems more don't think there are axioms

13:12

of human behavior of human there'd be

13:14

axioms of ethics though

13:17

we can't think then that

13:19

we have a an algorithm to write down

13:22

which will say know this actions permissible seconds

13:24

impermissible having said that you

13:27

to terrorism is more plausible

13:29

in that respect many other ethical approaches

13:32

because it uses quantify pull data

13:34

it because it tries to be based on facts

13:37

and consequences and actions

13:40

whereas other types of theory

13:42

maybe theory maybe by theory or

13:44

right space theories it it gets much more complicated

13:47

to think how you could possibly program and

13:49

machine to implement and

13:51

go beyond human understanding

13:54

of essex it wouldn't want to see is

13:56

understanding that x these

13:58

are machines that the attack patterns

14:00

use patterns to form complex models

14:03

but does that mean they have an understanding of it well

14:05

noah alpha though doesn't understand

14:07

that is playing against a human psychology whereas

14:09

any go player would understand turns

14:12

out you can write a very successful go playing

14:14

program that doesn't depend upon knowledge of

14:16

human psychology ethics i don't know

14:18

this has kind of be possible because

14:21

, not a game with a kind of rules and bounded

14:23

as on the other hand there's some ethical series had

14:25

tried it acts as

14:27

acts set of rules and so yes you

14:29

could take an ethics it was a set of rules

14:31

and you could have the machine have

14:33

those rules in the same way that go machines have

14:35

the go rules the tunnels as you wouldn't be

14:38

very happy with it's behavior because

14:40

we don't have a set of rules you could write down we could

14:42

say in this circumstance you should do such and such

14:45

even rules like don't murdering innocent

14:47

person or something like that wealth there are

14:49

lots of cases where that may be the only

14:51

way to prevent others very

14:53

serious harms if the rule just as

14:55

don't do as and be get all these famous problems

14:58

do lie to the dictator who

15:00

comes to your door looking for your our friend

15:02

who's hiding the basement spell if you were told never

15:04

to lie you hidden no

15:07

we could give the machines rules we

15:09

probably wouldn't like the result of and

15:12

one thing that i've been trying to argue is

15:14

that even though we don't think of these agents

15:16

and shouldn't as a stand think of them as having

15:19

ethical understanding they certainly

15:21

don't have consciousness of or as we can tell

15:23

they don't have the kind of moral emotions

15:26

that we house they don't have the kind

15:28

of appreciation of what it is to be

15:30

a person nor respect a person or something that's

15:32

so given their capacities

15:35

the question really is can we make them sensitive

15:37

to ethically relevant features of situations

15:40

those ethically relevant features a situations

15:43

they arise in situations not

15:45

just because we're human and even

15:47

not just because we're conscious

15:49

so if you're a person to a game theory

15:52

you have agents agents make moves

15:54

they have strategies they have interests

15:57

they have estimates and so those

15:59

a that's will

16:01

in certain circumstances either get themselves

16:03

into a collision or a paradox

16:06

unable to solve the coordination problem for example

16:08

even though their goals would require that

16:10

the coordinates or they won't in

16:12

on that view you can say well what

16:14

are the principles that would allow them to do that

16:16

or whether the motivations are goals

16:19

that would allow them to do that and those might

16:21

be quite universal they might not depend

16:23

upon any particular kind of agent

16:25

but they will be true of asians in general the

16:27

you're thing that if you think of ethics in terms of

16:30

human corporation then

16:32

the pope perhaps that machines

16:34

could come up with solutions

16:36

to problems that we haven't solved well

16:39

that seems like one possibility in the immediate

16:41

term i think part of it is can

16:44

we develop communities of

16:46

cooperation with artificial agents

16:49

because they are already out in the world

16:51

in fact for doing something very similar with autonomous

16:54

vehicles all the time there are coordination

16:56

problems there are problems of communication

16:59

of signaling to one another there are problems

17:01

were if we can cooperate

17:04

our able to achieve a result like smooth

17:06

going traffic that are the not and

17:09

those agents out there are learning

17:11

about that from our behavior from

17:13

their behavior and so they're acquiring

17:16

some of this kind of knowledge or

17:18

information that's a sense in which

17:20

we didn't have a program into them the

17:23

way of solving certain sorts of problems because they

17:25

will be trying things perhaps at random events

17:27

unfortunately and so we

17:30

couldn't program ethics into them the question is

17:32

ah good could they have experiences such

17:35

that they learn and people are beginning

17:37

to do some work now with cooperative

17:39

games rather than competitive games like chess

17:42

go and with a multi

17:44

agents settings not just individual

17:46

asian in a good question

17:48

is what kinds of forms

17:51

of behavior can they evolve then

17:53

what would their values function look

17:55

like so that they would have that

17:58

kind of behavior is not just a but

17:59

cuellar solution to this problem

18:01

that at a generalized motivational

18:03

structures that will enable them to go on to new problems

18:07

and solve them in a way that would produce things

18:09

like outcome that were positive some

18:11

have better for them and better for us we've

18:14

been talking a lot about the potential

18:16

of idea skull issues that arise the complexities

18:20

of trying see program

18:22

ai to access actually

18:24

as it were which isn't the contribution

18:27

of philosophers should be or could

18:29

be in this area very good again

18:32

i'm not thinking we program essex into them because

18:34

we don't know how to write doing the

18:36

formula are forgotten how to write down the axioms

18:39

or we don't know how to say well these

18:41

are the following thirty two thousand and ethically

18:43

relevant teachers are situations on

18:45

the other hand if you've machines that

18:47

can evaluate situations and simulate

18:50

out and figure out what what are the things that

18:52

are being affected by this variable in the situation

18:55

they might have that capacity of situational

18:58

responsiveness weaker that moral

19:00

intuition that moral

19:02

intuition i would say is in humans

19:04

actually quite sophisticated and well developed

19:07

and , promise of these machines

19:09

is that they will be much more light

19:11

intuitive knowledge systems than like

19:13

logic machines and so they can

19:15

accomplish the kinds of things that perhaps

19:17

only intuitive intuitive

19:19

can accomplish the way in

19:21

which we actually do

19:23

learn about the world that are it's common

19:26

sense knowledge or even sophisticated physics

19:28

involves a tremendous amount of intuitiveness

19:31

we , have the prospect of machines

19:33

that can have intuitive knowledge in something

19:36

like this sense and that

19:38

therefore could be used know

19:40

just as so tools

19:43

that could function as as agents

19:45

i do think more of loss we can help us

19:47

understand the way that humans learn

19:49

ethics and maybe something about the

19:51

nature of ethical theories what might

19:53

ethical knowledge look like what kinds

19:56

of principles of so decision

19:58

making are learning by be involved not

20:00

necessarily ethical principles it

20:03

might be quite general principles and

20:05

i'm arguing that the same principles

20:07

that enable us to do something like sustain the

20:09

cooperation necessary for a language

20:12

large created episteme a community where

20:14

we share information which humans have managed

20:17

to do in a way that other species of

20:19

know the capacities we have to do

20:21

that are really into row with our capacities for

20:23

morality and so philosophy

20:25

can try to help give us insight into will what

20:27

is that set of capacities

20:31

and then because they can smell this

20:33

smell this of what we would have to see as

20:35

a prospect for these machines the

20:37

runaway looking at these machines might

20:39

actually give us a better understanding as the

20:41

row for instance of moral intuition yes

20:44

i think that's right one of the things i've discovered

20:46

in my own teaching i teach in a way

20:48

that i pose problems to students in

20:51

they're able to respond anonymously using

20:54

little remote and over time

20:56

i've been able to ask them a lot of questions

20:59

other than the usual questions not

21:01

just i should you turn the trolley or something like

21:03

that but how would you feel if

21:05

you during the trolley or how would you feel

21:07

as you learned your roommate had turned the

21:09

trolley or if you were

21:11

to become contrite after

21:14

having pushed someone off a footbridge and

21:16

you went to that person's family to try

21:18

to express your deep

21:21

regret the what you did or thought you had to go

21:23

through you feel how would you

21:25

expect them to feel what would you try

21:27

to come out of that interaction the

21:29

interesting thing is my students don't just as

21:32

are sick questions like said you push them and they have

21:34

answers to those questions and they have

21:36

been very quickly the mother tells

21:38

me about ethical intuition is that it actually embodies

21:40

of very complex set

21:42

of understanding of human relations surprising

21:45

is how robust some of these intuitions are and

21:48

you my she will why are they so robust

21:50

and the answer might be because this because this kind of thing that

21:52

we all learn from our

21:54

human experiences interestingly

21:57

enough there's a have you know about psychopathy

22:00

that is actually a kind of a learning disorder

22:02

though i think that important

22:05

feature ethical intuition it's not some

22:07

kind of a seeing of platonic

22:09

principles neither is it instinctive

22:13

we can have a situation where we

22:15

learn about philosophical

22:17

problems like

22:18

the status of intuitions

22:20

i'm trying to model what's going on

22:22

in intuition and so one thing these

22:25

machines can do is give it a whole body

22:27

of human intuitions try to extract

22:29

from that what kind of implicit theories which

22:32

philosophers have been trying to do but we do it piecemeal

22:35

and very hard to do we

22:37

might find that actually there are these implicit

22:40

series involved in our intuitions and then we can

22:42

look at those critically there

22:44

were no question the measuring

22:47

the world in which they i could

22:49

actually be somewhere where we outsource

22:52

some of ammo thinking what if it gets good at

22:54

modeling moral intuitions why

22:56

don't we need to use intuitions anymore

22:59

isn't there a risk that ah capacity

23:01

for making judgments

23:03

on the fly could atrophy

23:05

as atrophy result of that well yes

23:08

and an example of this would be something

23:10

like the way in which people so sense of direction

23:12

or ability to stay oriented and in physical

23:14

space as and atrophied to some extent thanks

23:17

to the fact that people are using their phones

23:19

new symbol , i expect

23:21

to that some of the social skills of drivers

23:24

may atrophy because they're just counting

23:26

on the car to do this and they wouldn't

23:28

be as good at it if they were put in a position

23:30

of driving one way to think about the idea

23:32

of learning here is that

23:34

learning as learning continuing process

23:36

and to the extent that we do that actually

23:39

there's going to be less effective learning by

23:41

these machines because part

23:44

of what they will learn from is the way that we interact with

23:46

them and we're going to have to interact with them as

23:49

agents that they are needed that we are you

23:51

know you can have a gentle picture of knowledge in

23:53

which okay knowledge is not some unified

23:56

axiomatic structure it benefits

23:58

from the diversity of input that

24:00

it gets added expenses add diversity collapses

24:03

then it's going to be biased heavily by the

24:05

pool no machine can know

24:08

everything so there will always be some

24:10

bias from the amount

24:12

or kind of data is going in in

24:14

humans would be irresponsible if

24:17

save resigned their role because

24:19

just think about the way in which in our lifetimes

24:22

there has been change in the way in which

24:25

gender categories have got

24:27

ethical implications or are connected

24:30

with ethical issues that's a historical thing

24:32

it's not a thing that would be something that them machines

24:35

would be tuned into if

24:37

we turn over there making all these decisions

24:40

then they will reflect twenty

24:42

years ago moral intuition bread ours

24:44

is constantly being updated were constantly

24:46

learning and , is our responsibility

24:49

not to allow that to happen and

24:52

to decide to we do will actually be getting

24:54

less intuitive rosen

24:57

thank you very much for thank you very much

25:00

appreciated wishes

25:06

tomorrow's supposed to be bites go to www

25:09

dot philosophy bites dot com you can

25:11

also find details there a philosophy bites

25:13

books and how to support us

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features