Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:06
Hey, welcome to Stuff to Blow Your Mind. My name
0:08
is Robert Lamb and I'm Joe McCormick,
0:10
and it's Saturday. Time to go into the vault,
0:12
this time with a follow up from last Saturday's
0:14
Vault episode. This is going to be our
0:16
episode on the Split Brain Experiments,
0:19
Part two. If you listen last Saturday,
0:21
you know what's in store. So I guess let's
0:23
jump right in with every day. And from
0:25
both sides of my intelligence, the
0:27
moral and the intellectual, I've
0:30
else drew steadily near to that truth, by
0:32
whose partial discovery I have been
0:34
doomed to such a dreadful shipwreck
0:37
that man is not truly one, but
0:40
truly too.
0:45
Welcome to Stuff to Blow your Mind
0:47
from how Stuff Works dot Com.
0:55
Hey, welcome to Stuff to Blow your Mind. My name
0:57
is Robert Lamb and I'm Joe McCormick, and this
0:59
is going to be part two of our two part exploration
1:02
of hemisphereic collateralization and
1:04
especially the split brain experiments
1:07
of Roger Sperry and Michael Gazzaniga
1:09
starting in the nineteen sixties. Now, if you
1:11
haven't heard the last episode, you should really
1:13
go check that out first. That's gonna lay all the groundwork
1:16
for what we're talking about today, right, and it will also
1:18
explain why we kicked off this episode
1:20
and the last episode with the reading from Robert
1:22
Louis Stevenson Strange Case of Dr
1:24
Jacqueline mr. Hyde from six
1:27
short version is Robert Louis Stevenson thought
1:29
he had another dude in there. What did he call
1:31
him? The other guy? The man inside me by I
1:33
know it was a different author. Uh no, it
1:35
was it was me and the that other
1:38
fellow, that other fellow. Yeah. So
1:40
in the last episode we discussed twentieth
1:42
century research on a small group which
1:46
it was a small subset of the total
1:48
group of maybe fifty to a hundred
1:50
or maybe a little more than a hundred people
1:53
who have ever received a surgical
1:55
intervention called a corpus
1:57
callosotomy, which is a severing
1:59
of the corpus colosum and the corpus
2:01
colosum you can kind of think of as the high speed
2:03
fiber optic cable that connects the two hemispheres
2:06
of the brain together. Now, the surgery
2:08
was originally intended as a kind of last
2:10
resort treatment for people who had
2:13
terrible epileptic seizures. There
2:15
are so few of these patients because now we generally
2:17
have better, safer ways of treating epilepsy
2:20
without such a radical surgery, right
2:22
though these individuals are still around.
2:24
Yes, certainly, in the last episode we mentioned
2:26
that Pinto study that looked at a couple of
2:29
them in seen, And it's
2:31
very possible that we have listeners out
2:33
there who have received this surgery
2:35
as well. And obviously we would love to hear from you
2:37
if there's anything you would like to share. Oh yeah, please,
2:40
if you have a split brain, email us immediately.
2:43
And in fact you mentioned the more recent research.
2:45
We're gonna look at some of that research in today's episode.
2:48
But what neuroscientists learned
2:50
in the twentieth century from this small group
2:52
of patients was truly remarkable.
2:54
Beginning in the nineteen sixties and continuing
2:57
up until recent years, these split brain
2:59
patients have been the subject
3:01
of some of the most interesting research ever
3:04
on the nature of the brain, the mind, and the
3:06
self. So last time we talked about the original
3:08
work of like Sperry and Gazzaniga, who
3:11
discovered many fascinating things about
3:13
how it's possible for one half of the brain
3:15
to not know what the other half is
3:17
thinking, doing, or seeing. This
3:20
time we want to follow up on the subject,
3:22
to explore some more recent studies and
3:24
to ask questions about what these split brain
3:27
studies mean for our lives. And
3:29
to start off, I wanted to mention
3:31
an anecdote I came across from the neuroscientist
3:34
V. S. Ramachandren that he has brought
3:36
up in some of his public talks and work. He
3:39
tells a story of working with one
3:41
particular split brain patient who
3:44
had been trained to respond to questions
3:46
with his right hemisphere. Now you'll remember
3:49
from our last episode that in the case of
3:51
most patients, the right hemisphere of the brain
3:53
cannot speak. It might have some
3:55
very rudimentary language comprehension,
3:58
but generally language and especially
4:01
the production of speech, is dominated
4:03
by areas of the left hemisphere.
4:06
So if you're dealing with the right hemisphere of
4:08
a split brain patient and you show something
4:10
only to their left visual field, which connects
4:12
to the right hemisphere, and you ask
4:15
them about it, what often happens is
4:17
that, for instance, they will not be able
4:19
to say the thing you have
4:21
showed them in their right brain, or
4:24
even explain it in words, but
4:26
they will be able to draw the image
4:28
with their left hand. Now. In the case
4:31
of Ramaschandrian story, he had
4:33
trained a patient in a lab at Caltech
4:35
to answer questions posed directly
4:38
to his right hemisphere only by
4:40
pointing with his left hand to
4:42
response boxes indicating
4:45
yes, no, I don't know now. Of
4:47
course, asking these questions directly to the left
4:49
hemisphere is a lot easier because it just processes
4:51
language normally and you can just ask. But
4:53
he trained the right hemisphere to respond as
4:56
well, so the patient was perfectly
4:58
capable of answering questions like this with
5:00
either hemisphere. Are you on the moon
5:02
right now? Patient says no? Are
5:04
you at cal Tech? Patient says yes.
5:07
But Rama Schondre and then asked the right
5:09
hemisphere do you believe
5:12
in God? And it says yes.
5:15
And he then asked the left hemisphere, the
5:17
language dominant hemisphere, do
5:19
you believe in God? And it says
5:21
no. This is yet another
5:24
one that immediately when I heard the story, the hair stand
5:26
up on the back of my neck. I feel the I
5:28
feel that the goose bumps of of counterintuition
5:32
running through me. Yeah, because
5:34
I feel like, for the for the most part, I
5:36
feel like a lot of us want
5:38
to feel like we have a definitive answer
5:41
to that question, and Queen answers like that,
5:43
Now I'm probably a little
5:45
weirder and that I and I imagined
5:48
a lot of our listeners are like this as well, where someone
5:50
asks you a questions like this and you can be a lot more wishy
5:52
washy and say, well, I don't know. It depends you know,
5:55
yes and no. I feel
5:57
like most of us, if not
5:59
all of us, we can have We can have contry
6:01
contrary ideas in our mind. We
6:03
can have conflicting notions that are that
6:06
are vying for dominance, which me, are you asking?
6:08
Yeah? Askin Jacky? L are you asking hide?
6:11
You know? Hid he? You know he's he's not much
6:13
of a churchgoer, but but Jackal
6:15
he's there every Sunday. Yeah, but he's
6:17
only there to ultimately work his way up the chain
6:19
and usurp the creator. Now
6:23
Rama shonder and jokingly asks a theological
6:26
question about this. He says, you know, assume the old
6:28
dogma that people who have faith in God
6:30
go to Heaven and people who don't go to hell?
6:32
What happens when the split brain patient dies?
6:35
That's a good laugh line. But I
6:38
think this question is actually more profound
6:40
than it seems at first, because we
6:42
may not be divine judges casting
6:45
people into heaven or hell, but we are judges,
6:47
and we judge and evaluate and characterize
6:50
people all the time every day, as
6:52
if they are some sort of essential
6:54
whole. We pick out what we believe
6:56
to be the salient characteristics
6:58
that define a per person, like this
7:01
is their character and
7:03
and now we know who they are. This
7:05
is their mind, this is the person. There
7:08
might be no way to get people to live and behave
7:10
other than this. I mean, there might just be an inextricable
7:13
part of our our personalities that we have to
7:15
judge people as essential holes
7:17
in this way. But I think this research should
7:19
cause us to wonder about our folk beliefs
7:22
about the nature of the mind and the brain and
7:24
what it means to be a person. Yeah,
7:26
I mean, obviously, just to talk about judgment,
7:29
we we have some severe problems with
7:32
with with dealing with the idea that that
7:34
that that there is not a single person
7:36
over a length of time. I mean,
7:39
I mean, obviously, you have people serving prison
7:41
sentences for crimes that an
7:43
earlier iteration of themselves
7:45
committed. What do they say, I'm a different
7:48
person now, and and it is true
7:50
we have all different people than than we once
7:52
were. But you might in some ways
7:54
also be a different person than you were a
7:56
couple of seconds ago, right,
7:59
Or it can be kind of a juggling back and forth. You
8:01
know, I'm a different person in the morning versus
8:03
uh, the afternoon. I mean, I I truly
8:06
feel that, well, I mean, when it comes to questions
8:08
like this, like the theological question. The
8:11
fact is, most people, I think,
8:13
are probably filled with all kinds of doubts
8:15
concerning whatever their beliefs about religion
8:18
are, whether you believe in God or not.
8:20
Either way, you probably sometimes wonder
8:22
if you're wrong or you should. That's
8:24
always a great exercise about anything in
8:26
life, think about the possibility that you're
8:28
wrong, no matter what it is exactly. But
8:31
our everyday experience, of course, is that these
8:33
varying states of doubt they get somehow
8:36
synthesized. Right. You roll
8:38
it all up together, you say, even though whichever
8:40
way I am, whether I believe in God or not, I
8:42
ultimately have one way of answering
8:44
that question. Most people are like this when
8:46
you I mean, you might not be this way, Robert, but a
8:49
lot most people would say I have an answer. Well,
8:51
at the end of the day, or even
8:53
just minute to minute, you
8:56
your brain has to tell a story about who
8:58
you are, right and for that to
9:00
make sense, there still has to be a sentence. There
9:02
still has to be a story, some sort of
9:04
continuation. And even if
9:07
you know my story is a little more uh,
9:09
you know, meandering, it's still a story,
9:12
right, Yeah, Yeah, You're still narrativising
9:14
yourself. You're composing a synthetic
9:16
picture of who I am, and for
9:18
you, I think that picture includes more
9:21
ambiguity than a lot of people are comfortable with.
9:23
But either way, no matter what, you're telling a story
9:25
about yourself. Yeah,
9:27
and so despite your doubts, either way, you
9:29
think of yourself as one whole,
9:31
unified, unified person. You either believe
9:33
in God or you don't, or you identify
9:36
you have some narrative that's in between. You say I'm
9:38
an agnostic or whatever. But this is just one
9:41
case of a generally fascinating
9:43
phenomenon to ponder. What
9:46
if by asking parts of our brains
9:48
separately, we would think different
9:50
things about all kinds of stuff,
9:53
have different feelings, make different
9:55
judgments, make different moral judgments,
9:58
be different people. Is anyone
10:00
aspect of your brain more truly
10:03
authentically you than another
10:05
aspect of your brain? I mean they're both
10:07
in your head right. So today
10:10
this is sort of what we wanted to focus on to talk about
10:12
some of these types of takeaways from
10:14
split brain experiments and more recent research
10:17
on split brain patients. So one
10:19
really fascinating area of research we can
10:21
look at is the idea of moral judgments.
10:24
Robert, can I pose you a scenario and see
10:26
what you think? Yes, go ahead, band
10:28
or snatched me here? Okay? Oh yeah,
10:31
you're taunting me with it every day. I
10:33
still haven't seen it yet, but I will. Okay,
10:35
here's the scenario. Grace and
10:37
her friend are taking a tour of a chemical
10:40
plant. Grace goes over to the coffee
10:42
machine to pour some coffee. Grace's
10:44
friend asks if Grace will put some
10:47
sugar in hers, and there is
10:49
a white powder in a container next to
10:51
the coffee machine. The white
10:53
powder is a very toxic substance
10:56
left behind by a scientist and
10:58
deadly when ingested. The
11:00
container, however, is labeled sugar,
11:02
so Grace believes that the white powder
11:04
is regular sugar. Grace
11:07
puts this white powder in her friend's coffee.
11:09
Her friend drinks the coffee and dies.
11:12
Now the question is, is what Grace did
11:15
morally acceptable or not um
11:19
given this scenario, I mean,
11:21
it seems morally acceptable because she didn't
11:23
know it was toxic. It was labeled sugar. Yeah, she
11:25
was do and she was following
11:27
a request. Yeah, so you are answering
11:30
the question the way almost all adult
11:32
adults tend to answer these questions, that
11:35
what matters is the intention of
11:37
the person doing the action. Uh
11:39
So let me pose it another way. Same scenario.
11:41
Grace and her friend or at a coffee. They're getting
11:43
coffee at the chemical plant. Now
11:45
it turns out that the white powder in the
11:47
container is just sugar and it's
11:50
fine, but it is labeled toxic.
11:53
So Grace believes that the white powder
11:55
is a toxic substance, but she's
11:57
wrong. She puts it in her friend's coffee.
11:59
It's actually just sugar. Her friend drinks
12:01
it. Is what is what Grace did morally
12:03
acceptable? Well, I would say it is forbidden
12:05
because she attempted to poison a friend, exactly
12:08
right. So yeah, this is how I would answer
12:10
as well. This is how almost all adults tend
12:13
to answer these questions. The fact
12:15
is that in general, adults
12:17
tend to think that intentions
12:19
are highly morally relevant, so
12:22
they usually say that a person who accidentally
12:24
poisons a friend of theirs with no intent to
12:26
harm them is not morally blame
12:29
worthy, but somebody who intends
12:31
to poison a friend, even
12:33
if they fail at doing so, is morally
12:35
blameworthy. And of course, like you know,
12:38
there are many aspects that you see this put
12:40
into practice around the world, and like legal injustice
12:42
systems, a person is punished a lot more
12:45
for trying to hurt someone on purpose than for hurting
12:47
them by accident, though often sometimes
12:49
they are still held responsible for hurting
12:51
somebody growth negligent situation,
12:53
you know. Uh, And that's like a middle category,
12:56
right like if you didn't mean to hurt somebody, but you were
12:58
doing something really reckless, send it
13:00
hurt them. That's sort of like a middle culpability
13:03
level, right like if you stored the
13:05
toxic white powder next to the
13:07
sugar and she just didn't look
13:10
closely enough, like you really should you know
13:12
that this place has as sugar
13:14
and toxic poison. You should
13:16
you should know to check which one you're scoop Getting lumps
13:19
out of right, But we wouldn't think that Grace should
13:21
have expected there to be poisoned right next to the
13:23
coffee machine, right. And on
13:25
the other hand, Grace, you can't expect Grace to
13:27
just expect people to be trying to poisoning
13:29
her all the time like they're they're they're certain cultural
13:32
expectations in place here exactly. But
13:34
the weird thing is not everyone
13:37
answers scenarios this way.
13:39
For example, previous research, including
13:41
by the Swiss psychologist Jean piage Uh
13:44
and others later, has found that young
13:46
children and pj found this was up
13:48
to about the age of nine or ten,
13:51
tend to attribute moral guilt and
13:53
deservingness of punishment in exactly
13:56
the opposite way. They assigned
13:58
guilt based on the active consequences
14:01
of the action rather than
14:03
to the knowledge or intentions of the
14:05
agent, meaning that many young children
14:08
will suggest that if Grace means to
14:10
put sugar in her friend's coffee but accidentally
14:13
poisons her friend, she is naughty.
14:16
But if she tries to poison her friend
14:18
and the poison doesn't work, she's fine.
14:20
Well that sounds totally believable. I mean, I now
14:24
that it's pointed out like that. You know, I can see
14:26
I can see various aspects of that popping up
14:29
in just raising a child, you know, where
14:32
where they're gonna they're gonna going to jump to this conclusion.
14:34
You know it certainly not with poisoning, but with
14:37
just sort of the everyday minutia that
14:39
fills your life. Well, they don't reason this way
14:41
every time, Like sometimes intentions
14:43
seem salient to them, but generally
14:46
the rule is after about age ten,
14:49
almost nobody ever thinks
14:51
that accidentally harming someone
14:53
is worse than intending to harm them
14:55
and not harm in failing. Yeah,
14:58
but this, I mean that I've seen this with
15:00
my son though, where like he'll do something
15:02
accidentally and then he's really hard
15:04
on himself for having for
15:06
for quote, being bad or having you know, done
15:08
something bad and you have to reassure him you
15:10
know this was you know, there was an accident,
15:13
but you know it's all cool. Well, this is a
15:15
fascinating phenomenon on its own.
15:17
I mean, before we even get to how this applies to the split
15:19
brain experiments for example, you
15:21
know, I went back, I was like, is this really true? So I
15:23
was reading some of Pj's work on this question
15:26
from a book of his. And so here's one of the scenarios
15:28
he describes when interviewing young children.
15:31
Okay, the first one is, uh this, uh
15:34
about this little boy named John, Robert, do
15:36
you want to read about John? Sure? A
15:38
little boy who is called John is in his room.
15:40
He has called to dinner. He goes into
15:43
the dining room. But behind the door
15:45
there was a chair, and on the chair
15:47
there was a tray with fifteen cups on
15:49
it. John couldn't have known that there was
15:51
all this behind the door. He goes
15:54
in the door, knocks against the tray, Bang,
15:56
go the fifteen cups and they all
15:58
get broken. All right. Here's the other scenario.
16:00
Once there was a little boy whose name was Henry.
16:03
One day when his mother was out, he tried to
16:05
get some jam out of the cupboard. He
16:07
climbed up onto a chair and stretched
16:10
out his arm, but the jam was too high
16:12
up and he couldn't reach it and have any
16:15
But while he was trying to get it, he knocked
16:17
over a cup. The cup fell down and
16:19
broke. Ah.
16:21
So, yeah, we have a situation where John
16:25
was just going about normal
16:27
everyday. How stuff he didn't know where some stuff
16:29
was. Yeah, and it stuff got broken. But
16:32
Henry is trying to do something he shouldn't and
16:34
then accidentally break something. But
16:37
here then, PJ includes
16:39
a little transcript of a dialogue with
16:41
a six year old boy named Geo about
16:43
these stories. Robert, do you want
16:45
to be Geo? I'll be the child? Yes? Okay,
16:49
have you understood these stories? Yes?
16:52
What did the first boy do? He broke
16:54
eleven cups? And the second
16:56
one he broke a cup by moving roughly?
16:59
Why did the swan break the cups because
17:01
the door knocked them in the second he
17:03
was clumsy when he was getting the jam
17:06
the cup fell down? How did Geo
17:08
become Richard O'Brien? Okay?
17:10
No, sorry? Going on? Is one of the boys
17:13
naughtier than the other? The first is
17:15
because he knocked over twelve cups? If
17:17
you were the daddy, which one would you punish
17:20
most? To one who broke twelve cups?
17:22
Why did he break them? The door shut too
17:24
hard and knocked them. He didn't do it on purpose?
17:26
And why did the other boy break a cup? He wanted
17:29
to get the jam. He moved too
17:31
far, the cup got a broken.
17:34
Why did he want the jam? Because he was all alone?
17:36
Because his mother, wasn't there have
17:38
you got a brother, no, a little sister. Well,
17:41
if it was you who had broken the twelve cups
17:43
when you went into the room and your little
17:45
sister had broken the one cup while she
17:47
was trying to get the jam, which of you
17:50
would be punished most severely? Me?
17:52
Because I broke more within one cup. Robert.
17:55
First of all, I'm gonna give a rave view to
17:57
your creepy child voice. That was like
17:59
a beau, a full riff raff French
18:01
geo. I think I was trying to go for like a Damian
18:03
child or something. But you know, Richard O'Brien
18:06
is still pretty good. It's all for you
18:08
riff raff. But this is illuminating.
18:10
This shows, Uh, this shows
18:12
how the six year old is thinking about these
18:15
two scenarios in applying judgment. Yes,
18:17
almost no adult reasons this way
18:19
right right, So this on its own
18:21
is fascinating to me. Why this discrepancy
18:24
in the moral reasoning of children and adults?
18:26
And what causes the change? You know? POJ
18:29
says the change tends to happen somewhere
18:31
in late childhood, you know, somewhere between
18:33
like uh, like seven
18:35
and nine or ten. This change really
18:38
takes over and people still and the children
18:40
start reasoning about moral intentions
18:42
and moral knowledge as opposed to
18:44
just the objective outcomes. Uh.
18:47
One issue I think that plays into this maturation
18:49
process in moral judgments is of course going
18:51
to be the development of the sophistication
18:54
of theory of mind, and
18:56
theory of mind of course is the ability to understand
18:59
that others have into mental states
19:01
and imagine what those states are. But this clearly
19:03
can't be the only factor, because
19:06
most children develop theory of mind by
19:08
around age five or so, and
19:10
a significant number of them think outcomes
19:12
matter more than intentions for guilt until
19:14
around age nine or so, So
19:17
there must be something else happening also,
19:20
so they're able to either able to contemplate
19:22
other mind states, and yet they're
19:24
still sticking to this. Uh,
19:26
this this harsh form of judgment. Yeah.
19:29
And again, to be clear, not in every
19:31
case, because sometimes children will seem to think
19:33
intentions matter, but they clearly they they
19:35
default to this far more than adults
19:38
would. Now, there's one reason to think that, of course,
19:40
theory of mind is important for making a
19:42
mature moral judgments the kind adults
19:44
make based on knowledge and intentions.
19:46
For the obvious reason that when you make
19:49
a judgment considering a state of mind, including
19:51
the knowledge and intentions of the person who broke
19:53
the cups or put the powder in the coffee or
19:55
whatever, you need to imagine their state
19:58
of mind, like you have to have that in your rain
20:00
in order to evaluate whether they were guilty
20:02
or not. And so in like two thousand and eight
20:04
two thousand nine, researchers
20:06
named Leanne Young and Rebecca Sacks
20:09
used neuroimaging to find evidence that
20:11
when you try to ascribe beliefs
20:13
and intentions to other people, essentially
20:16
when you practice theory of mind and you're
20:18
thinking about other minds, it involves
20:20
processes that are lateralized
20:23
their primarily on one side
20:25
of the brain, specifically in the right
20:27
temporal parietal junction
20:30
or TPJ. And in a
20:32
two thousand nine study, Young and Sacks found that
20:34
uh temporal parietal junction activity
20:36
in the right hemisphere only appeared
20:39
when people tried to assess the moral
20:41
significance of things like accidental
20:43
harms when you hurt somebody but you didn't
20:45
mean to. So if I tell you a story
20:48
about Jeffrey accidentally knocking
20:50
somebody into the Grand Canyon and then I
20:52
ask you to think about whether Jeffrey
20:55
did something morally wrong or not. Whatever
20:57
thinking you use to answer that question
20:59
will probably we involve the t PJ
21:01
on the right side. But oh,
21:04
what if the part of your brain
21:06
that's getting that's interacting with the language
21:09
that poses this question to you, cannot
21:11
retrieve information from the lateralized
21:14
TPJ the right side the brain.
21:17
Yes, so we're gonna look at the two thousand
21:19
tens study from Neuropsychologia
21:22
called Abnormal Moral Reasoning and Complete
21:24
and partial calisotomy Patients by
21:27
Miller, Senate, Armstrong, Young, King,
21:30
Pagi, Fabri, Polinara,
21:33
and Gazanaga. So the authors begin
21:35
by looking at the state of affairs we just talked
21:37
about, uh, with the you know, the
21:40
localization in the right hemisphere of this
21:42
part of the brain that's used in imagining
21:44
other minds and making judgments about something
21:46
like the intentions of somebody
21:49
in reference to moral guilt and the
21:51
right quote. These findings suggest that patients
21:53
with disconnected hemispheres would
21:55
provide abnormal moral judgments
21:57
on accidental harms and fail
22:00
old attempts to harm, since normal
22:02
judgments in these cases require information
22:04
about beliefs and intentions from
22:06
the right brain to reach the judgmental
22:09
processes in the left brain. So they
22:11
ran a test. They used six
22:13
split brain patients who have had
22:15
either a partial or total sectioning
22:18
of the corpus colosum and compared
22:20
that with twenty two normal control subjects.
22:22
Now verbally, so what they did is verbally
22:25
out loud conducted interviews
22:27
posing moral judgment scenarios
22:30
like the sugar or poison story we talked
22:32
about with Grace, but also other ones like it
22:35
uh. They conducted these interviews verbally,
22:37
asking the subjects about whether different types
22:40
of action in the scenario were morally
22:42
acceptable or not. And remember, of course,
22:44
which hemisphere of the brain is the one primarily
22:46
responsible for speech. It's the left.
22:49
So if you're having a verbal interview with somebody,
22:51
their left hemisphere is sort of like it's
22:53
like the gatekeeper right that will
22:56
in most cases be dominating
22:58
the input and output of the brain you're
23:00
interacting with, since the input and
23:02
output is all spoken words. So
23:04
if you have to give your answers in words
23:06
coming from your left hemisphere and
23:08
it can't communicate very well with your right
23:11
hemisphere or at all with your right hemisphere,
23:13
which is the home of an important part
23:15
of the brain that used to think about the knowledge
23:18
and intentions of other people. Your verbal
23:20
answers on subjects requiring this kind
23:22
of knowledge may very well be impaired,
23:25
and the results, it turned out, supported
23:27
this hypothesis. The control subjects,
23:29
the people without split brains, they tended
23:32
to judge just like we did earlier. Like they
23:34
judged based on intentions, well, did grace
23:37
mean to harm somebody or not? And that was the
23:39
mainly salient thing. The split
23:41
brain patients did so far less
23:43
consistently, more often judging
23:46
based purely on outcomes, the way
23:48
many young children did and pj's
23:50
work, and also
23:52
to supplement their experiment, they tested
23:54
two of the split brain patient's ability
23:56
to detect hypothetical faux
23:59
pause. For example, a
24:01
person quote telling somebody how much
24:03
they dislike a bowl while
24:05
forgetting that the person had given them
24:07
that bowl as a wedding present. Uh.
24:10
And of course, the idea is that a person who's unable
24:12
like if you're unable to give spoken answers
24:14
involving the theory of mind function localized
24:17
in the right TPJ, you will find
24:19
it significantly harder to detect
24:22
a faux paw which requires you to think
24:24
about other minds, and the split
24:26
brain difference held true here. Out of
24:28
tin faux pause, they said, patient
24:31
VP successfully detected only
24:33
six and patient j W correctly
24:36
identified only four, whereas
24:38
control subjects all identified a hundred
24:41
of the faux pause. So when they were given a scenario
24:43
like that and asked did something awkward happen normal
24:45
people? They detected every time. In
24:48
fact, one of the things that I would say our brains
24:50
are most highly suited
24:52
for is detecting social awkwardness
24:54
and stuff, right, Yeah, And
24:56
it is interesting to notice this emerging
24:58
in younger or children too, you
25:01
know, like you see this kind of awareness
25:03
coming online, you know where they're
25:05
able to identify faux pause as
25:07
opposed to just be like the master of
25:09
folk pause. Well do you ever notice
25:12
I wonder if like adolescence
25:15
and teenage years are kind of an error.
25:17
It's like it's a time when you were almost
25:20
like hyper aware of social awkwardness.
25:23
Does that ring true to you? Um
25:26
to a certain extent? But I don't know. I've run into
25:28
some teens who I mean, there
25:31
are a lot of different types of brains out
25:33
there, but I mean I've run into some teams that that definitely
25:36
have a lot of social awkwardness or
25:38
or definitely walk into a lot of faux
25:41
pas. So I don't know. Well, I mean, just because
25:43
you are awkward doesn't mean you're not aware of awkwardness,
25:46
right, Yeah,
25:48
certainly awkwardness does seem to define
25:50
that buried in one's life. That
25:53
would be that might be something to come back to. I know we've done
25:55
episodes in the past on the teenage brain, in
25:57
the particular aspects of the team
26:00
age brain. I wonder if there's a if
26:02
there's an entire episode on the science of awkwardness.
26:05
Well, I think we should take a quick break and then when we come
26:07
back we can discuss this study a little more
26:10
than alright,
26:13
we're back, all right. So we've just discussed this
26:15
study about split brain patients and
26:17
moral judgments and found that split
26:19
brain patients, at least in this one study, made
26:22
moral judgments based on outcomes
26:25
rather than on intentions, more
26:27
like children sometimes do instead
26:29
of the way that adults normally do. Um,
26:32
And this is fascinating. Now, of course, we
26:34
should acknowledge some potential drawbacks
26:36
of this experiment. Like all split brain
26:39
studies, by necessity, it's a small
26:41
sample, right, you know, there aren't that many of these
26:43
people out there, and even a smaller subset
26:45
of them want to participate in
26:47
experiments like this, But so it's almost on
26:49
the scale of anecdote, so you have to be careful
26:52
about drawing strong conclusions
26:54
from the results. Also, there are some
26:56
other detailed complications in
26:58
the study, such as questions about why
27:00
the effect also manifested impartial
27:02
calisotomy patients, so when the authors
27:05
had not expected it to they thought it would only
27:07
appear in the full calisotomy patients.
27:09
And then also about where the exact
27:12
side of decoding the beliefs of others is located.
27:14
Maybe it's not exactly the TPJ but
27:16
more anterior to it. Uh So
27:18
that's some peripheral issues. But nevertheless,
27:21
if we tentatively accept these results
27:23
like how fascinating, and it leads
27:25
to these questions like here's one.
27:28
You know we discussed in the last episode
27:30
that despite the radical nature
27:32
of the surgery that cuts the corpus
27:34
colossum and the amazing neurological
27:36
anomalies that can arise from it under lab
27:39
conditions, generally most patients
27:41
and patient families report totally normal
27:43
functionality no major changes
27:46
in personality or behavior after the surgery.
27:48
If it's changing their moral reasoning
27:51
in in this kind of way, how could
27:53
that be possible? I mean yeah,
27:55
because certainly from your own standpoint,
27:57
I mean, you were if you're
28:00
moral compass has changed than you
28:02
I mean, you can't see the forest for the trees,
28:05
right, But but you're gonna be
28:07
surrounded by other people who
28:09
would be able to identify the change.
28:11
But presumably yeah, you would think so,
28:14
I mean if there is actually a change. So
28:17
uh. And and also like, yeah, you think
28:19
that moral judgments sort of go to the heart
28:21
of a person's personality, right, like
28:23
that that is your character, that is who you
28:25
are as a person, or at least how you think about
28:28
that subject. Right. You would think there would be anecdotes
28:31
out there about like, yeah, my uncle
28:33
had this surgery and then his like his
28:35
his political ideology changed afterwards,
28:38
or yeah you have been something to that effect. But
28:40
we have not seen that in referenced in any
28:43
of these studies. So, if these results from
28:45
this two thousand ten studies are sound, what
28:47
accounts for the discrepancy here? And
28:49
the authors they posit three
28:51
possible answers. One is, well, maybe
28:54
there are profound personality changes
28:56
in split brain patients that have gone unnoticed
28:58
or unreported. They don't think this
29:01
is very likely because quote, most reports
29:03
from family members suggest no changes
29:05
in mental functions or personality,
29:08
and early studies that thoroughly tested
29:10
patients pre and postoperatively
29:12
reported no changes in cognitive
29:14
functioning. So they feel pretty robustly
29:17
that these patients in their day to day lives
29:19
are not really changed. The
29:22
other possibility as well. Maybe it's just because
29:25
the judgment tasks here have no relevance
29:27
to real life. But I mean,
29:29
we use judgments like this all the time, like did
29:31
somebody mean to do something that? That seems
29:33
like something that comes up every day? Yeah, I
29:35
mean I jokingly brought up
29:38
Bandersnatch the Town Adventure
29:40
Black Mirror episode on Netflix earlier,
29:43
But like I I found myself in watching
29:45
that, like having to make choices about moral
29:47
choices for the character. I found
29:50
myself very uncomfortable with with
29:52
choices that that I found morally
29:54
reprehensible, even though it's
29:56
just purely hypothetical. It's just
29:58
a story, right, all right? What else do we
30:00
have? What other possible answers? Well, the third
30:02
possibility is what the researchers think is probably
30:05
the case, which is that even though
30:07
this impairment is manifested in the lab,
30:10
in reality it somehow gets compensated
30:12
for somehow in daily life.
30:14
Other brain regions and
30:17
functions or alternative processes
30:19
kick in to counteract whatever
30:21
is causing people to give these unusual
30:24
answers in the lab condition. The
30:26
brain finds a way, yes, so what would
30:28
it be, Well, what about a version
30:30
of something not exactly but something like
30:33
the system one versus system to schema.
30:36
Of course, now, of course we can remind
30:38
people what the system one in the system to themes
30:40
are. Well, it's like, basically like the different
30:42
ways of dealing with the threat of the tiger. There's
30:44
the way of dealing with the tiger by avoiding
30:47
it and not going to the places where the tiger is,
30:49
and then there's the way of dealing with the tiger where
30:52
you have to fight it re flee from it. So
30:54
I think we'd have the order inverted there. But yeah,
30:56
so like system too is generally
30:58
considered to be like low deliberate,
31:01
methodical, logical thinking
31:03
about how to solve problems, whereas system
31:06
one is fast, reactive,
31:08
intuitive, implicit right
31:11
punch the tiger in the nose and run for it. And we
31:13
need both for life. I mean, system
31:15
to reactions might be less
31:18
likely to give us erroneous results.
31:20
But you don't have time to use system to thinking
31:22
on everything. You know, you're trying
31:25
to get through life. Most of the time. You need to
31:27
make quick judgments that are not overly
31:29
concerned. You know, you can't overthink, like
31:31
which foot I'm gonna put in front of the other right
31:33
now? Right? Yeah, So so you've got
31:35
to be prepared for either tiger, the distance
31:38
tiger or the close tiger.
31:40
And so maybe the idea here is
31:42
that the right TPJ is
31:44
somehow necessary for making
31:47
fast implicit system
31:49
one type decisions about
31:52
judging more, you know, the moral valance of an
31:54
action and imagining theory of mind.
31:56
But that you can if you can't do that,
31:59
you can somehow do the same thing. It
32:01
just takes longer, and it's is
32:03
a more difficult, deliberate process
32:06
that the brain has to go through if
32:08
it can't rely on this brain region
32:10
that does does this fast for you normally
32:13
the author's right quote. If the patients
32:15
do not have access to the fast implicit
32:18
systems for ascribing beliefs to others.
32:20
Their initial automatic moral judgments
32:23
might not take into account beliefs
32:25
of others, but you know, they're slow
32:27
reason deliberate thinking system can compensate,
32:30
it can kick in. Then again, I mean, I wonder
32:32
how this, if this is the case, and we'll
32:35
discuss this a little more, how this
32:37
wouldn't manifest in normal life, because I feel
32:39
like we use the fast intuitive system
32:41
one type process to make morally relevant
32:43
judgments all the time. I
32:46
mean, we're constantly making sort
32:48
of unfair moral judgments
32:50
about things that would not you know, they're
32:52
not using the kind of reasoning that you would
32:54
sit down and deliberate about. Think
32:56
about how often you get mad at somebody
32:58
because they do something accidentally, and
33:01
if you were forced to stop and think about it, you're
33:03
like, Okay, no they didn't, they didn't mean to
33:05
do that. There's no reason to morally
33:07
blame them. You just get mad in
33:09
the moment and you're just like, why are you in my way?
33:11
Or why did you do that? Yeah? Yeah, totally.
33:15
This is you know, this like the the other split
33:17
brain experiments we're looking at, though it
33:19
reminds me of say, if you're watching a three
33:21
D film and you have the glasses on, and
33:23
then you take the glasses off, and
33:26
you you you see that there is there's
33:28
there's some sort of uh uh, you
33:30
know, there's a lack of unity there. Or
33:32
it's like you're you're staring through the stereo view
33:34
and then you look at the card and you see that it's two images
33:36
side by side to create the united
33:39
whole. Like it's it's a glimpse
33:42
at the duality
33:44
that that is making the at least,
33:46
you know, the sort of the illusion, the experience of the whole
33:48
possible. Um. But
33:51
but but then it's it's we we shouldn't fall
33:54
under the we shouldn't
33:56
then fall into the trap of thinking that it is dual
33:58
by nature. It's like taking the US is up and saying,
34:00
oh, the world is really red, the world is
34:02
really blue. Well, no, no, that the world is
34:04
the thing that comes together. Yeah, And and
34:06
the glasses are designed to give you this three
34:09
D image the same way that the brain
34:11
is designed by evolution to have compensating
34:13
processes, to have one way of doing something
34:16
or another way of doing something depending on the
34:18
situational need. And so, of course
34:20
I indicated that the authors tend to think this
34:22
third answer is probably the correct one about
34:25
the compensating mechanism taking over in
34:27
real life scenarios. Uh And
34:29
as evidence, they cite the fact that in the
34:32
experiment, split brain patients
34:34
would sometimes spontaneously
34:37
blurred out a rationalization
34:39
of an answer that ignored intentions,
34:42
almost as if after giving the answer
34:44
out loud that ignored intentions,
34:47
they realized something was wrong
34:49
with it. So here's one example.
34:52
A split brain patient named JW hurt
34:54
a scenario where a waitress thought
34:57
that serving sesame seeds to a
34:59
customer would give him a terrible
35:01
allergic reaction. She thought he was allergic
35:03
to sesame seeds. She tried, she
35:06
served him sesame seeds, but
35:08
it turns out he wasn't actually allergic.
35:10
She was wrong about that, and the seeds didn't hurt
35:13
him, even though she thought they would. J
35:15
W said the waitress had done nothing wrong.
35:18
Then he paused for a few moments, then spontaneously
35:21
blurted out, sesame seeds are
35:23
tiny little things. They don't hurt nobody,
35:27
you know. It's it's almost as if he was
35:29
searching for a post talk rationalization
35:32
of an answer he'd already given, but
35:35
which began to seem wrong to
35:37
him as it sank in, you know, given
35:39
a few more seconds to think about it, and
35:41
the patient j W alone,
35:43
they reported spontaneously blurted
35:45
out rationalizations like this in
35:48
five of the twenty four scenarios,
35:50
so like more than a fifth, and
35:52
again, I just think back to the fact,
35:54
you know, post talk rationalization is a huge
35:57
part of life. We talked about this in the last episode
35:59
with the the writer
36:01
and the elephant, right, like, how often do
36:04
we do things that honestly we don't
36:06
understand why we did them, but we just
36:08
come up with a story, and we even believe
36:11
that story ourselves as an explanation
36:13
for why we did it. But you can see clear
36:15
evidence that that is not the reason. Right,
36:18
Yeah, you end up telling yourself, well
36:20
I wanted that product, or perhaps
36:22
oh well you might even you know, you
36:24
might even end up telling you so the st of the story
36:26
about how you were tricked into buying it. But
36:28
but there is some sort of rationalization about
36:31
the about the movements
36:34
of the beast beneath you. Alright,
36:36
on that note, we're going to take another break, but we'll be
36:38
right back. Than all
36:41
right, we're back, Okay, I think we should
36:43
take a look at another study about
36:46
moral judgment and the division
36:49
of the brain hemispheres. So this
36:51
is one from Royal Society Open
36:53
Science from called moral
36:56
judgment by the disconnected left
36:58
and right cerebral hemispheres, a split
37:00
brain investigation, and this is by Steckler,
37:03
Hamlin, Miller, King and Kingstone.
37:06
Uh. And when you get king and Kingstone together, you
37:08
never know what's gonna happen. So to
37:10
recap from the last study, we know that
37:12
lots of parts of the brain are used
37:15
in making moral judgments, including you
37:17
know, regions and networks in the left hemisphere
37:20
such as the the left medial prefrontal
37:22
cortex, the left temporal parietal
37:24
junction, and the left singulate. But
37:27
in order to make moral decisions based
37:29
on people's intentions, when you're
37:31
imagining what other people mean
37:33
to do and what they know, we seem
37:35
to require use of an area
37:38
in or around the area mentioned
37:40
in the last study, the right tempo
37:42
parietal junction or rTPJ.
37:45
And it seems that without it you can't properly
37:47
imagine other people's intentions
37:49
and beliefs to make a quick moral
37:51
judgment. So here's a question. Then
37:54
the right hemisphere seems necessary
37:56
in making a quick moral judgment
37:59
in the normal way based on people's intent,
38:01
But is it sufficient Could
38:03
the right hemisphere alone make
38:05
a judgment? So the authors try
38:07
to find out with the help of a split brain
38:09
patient. They write, quote, here we use
38:11
non linguistic morality
38:14
plays with split brain patient j
38:16
W to examine the moral judgments
38:19
of the disconnected right hemisphere.
38:21
So obviously you've got a problem if you're
38:23
trying to just talk to the right hemisphere, because
38:25
the right hemisphere is not going to do super
38:28
well at understanding a verbal scenario
38:31
you describe to them. Right it doesn't want to listen to you
38:33
tell a story. It doesn't want a lot of dialogue.
38:35
It just wants some sweet, muted
38:38
YouTube action the silent film hemisphere.
38:41
And again not to not to be overly simplistic,
38:43
because we do know from some research that the right
38:45
brain does seem to understand some
38:48
language, it's just not nearly as linguistically
38:50
sophisticated as the left hemisphere.
38:53
Um So they use these nonverbal
38:55
videos of people trying to help someone
38:57
and succeeding or failing, or
39:00
trying to thwart someone and succeeding
39:02
or failing. So an example might be somebody's
39:05
trying to get something down off of a high shelf,
39:08
and then somebody either like bumps
39:10
into them to try to knock them off the shelf
39:13
or tries to help them get the thing down or
39:15
something like that. And then they had j W
39:17
watch all these videos and point
39:19
with the finger of a specific hand
39:21
which is controlled by the opposite hemisphere,
39:23
to indicate which character was nicer.
39:27
So in a series of test sessions like this over
39:29
the course of a year, they found that JW
39:31
was able to make pretty normal intent
39:33
based judgments with his right
39:35
hemisphere alone pointing with his left
39:38
hand, but had a lot more trouble
39:40
making intent based judgments with the left
39:43
left hemisphere, in some cases
39:45
seeming to respond almost at random
39:47
with the left hemisphere. And yet the left
39:49
hemisphere is the hemisphere that the talks,
39:52
so there were more signs of the left hemisphere
39:54
making up ex post facto justifications
39:58
when it did not understand what what
40:00
the person had done. For example, after
40:03
one video, when asked why
40:05
he made the choice he did of which character
40:07
was nicer, JW just offered
40:09
the rationalization that blonds
40:11
can't be trusted. When one of the actors
40:13
in the video was blonde. So
40:16
here's one question why the discrepancy
40:18
with the last study. In the last study, the
40:20
left hemisphere defaulted more often
40:22
in making moral judgments based on, remember
40:25
the objective good or bad outcomes,
40:28
rather than people's intentions. Why did
40:30
it seem to make judgments at random this
40:32
time? So the authors say, maybe in the
40:34
previous study it's because subjects were explicitly
40:37
asked to judge whether a behavior was morally
40:39
acceptable or not, and in this study
40:41
instead, the subject was just asked who's
40:44
nicer, Maybe to the left hemisphere,
40:46
you know, separated and on its own devices.
40:49
Maybe it doesn't use any kind of
40:51
moral reasoning to judge who is nicer,
40:53
but uses some other kind of rubric. Maybe
40:55
nicer means something non moral
40:58
to it. Then again, there's
41:00
also the possibility, well, you know, we're
41:02
again limited to small sample sizes,
41:05
in this case very small of just one patient.
41:07
So it's possible that maybe JW
41:10
is just unusual. That's always a thing to consider
41:12
with this kind of study, and it's what you know,
41:14
unfortunately, what what this sort of research
41:17
is by nature limited to one
41:19
of the things that I think is interesting and looking at this
41:21
research, we we've looked at today with the different
41:24
kinds of moral reasoning in the different
41:26
hemispheres, is that we see again
41:28
the role of something that we talked about
41:30
in in part one of this series back
41:32
in the first episode, about the
41:35
role of what's thought of as the interpreter,
41:37
or at least in Michael Gazzaniga's theory,
41:39
that the interpreter in the left hemisphere.
41:42
So the idea is, of course, that your brain
41:44
constantly makes up stories to
41:47
explain why you just did what you did.
41:49
But split brain research indicates that we have
41:51
no guarantee that the stories
41:54
we give to explain our own behaviors
41:56
have any explanatory power at all.
41:58
A lot of times it seems more like they are just
42:00
confabulated post talk rationalizations,
42:04
that you just came up with something
42:06
to explain something you did when you really
42:08
have no idea why you did what you did.
42:10
The brain just pulled it out of its own button.
42:13
If the brain had a butt. In the previous experiments,
42:15
this had to do with stuff like why did
42:18
you draw this picture you know? Or why did
42:20
you pick this object out of a drawer with
42:22
your left hand when you couldn't name that object
42:24
in speech, or anything like that, and
42:27
people would make up excuses. Now you
42:29
you see a similar kind of thing perhaps going
42:31
on with making moral judgments. And
42:34
I think that there is some research
42:36
that this is indicative not
42:38
just of something about split brain
42:40
patients, but of something larger about
42:43
this phenomenon of interpretation
42:45
in the left hemisphere and of the human
42:47
condition itself. Yeah,
42:50
like we've we've touched on in this episode Sode,
42:52
in the previous episode, and in any other episodes
42:54
before. It's like there's always a story that is told,
42:56
right, We're constantly telling
42:59
a story about ourselves, and that story
43:01
involves rationalizations, rationalizations
43:04
for our actions and uh and
43:06
interpretations of who we are and why
43:08
we're doing everything we do exactly.
43:10
And it happens in multiple level. It happens
43:12
to explain why you have why you took certain
43:15
actions that you can't actually explain.
43:17
It happens to explain why your
43:19
mood changes. Because Aniga writes
43:21
about this that there are these cases where
43:23
you can have somebody who's has a mood shift
43:26
triggered, like for example, you
43:28
get you have split brain patients where you show
43:30
some positive or negative mood, triggering
43:32
stimulus to the right hemisphere, and then
43:34
the speaking part of the brain expresses
43:37
being upset, but then
43:39
we'll be unable to express why, and we'll
43:41
just make up a story about why, like, well,
43:43
because you did this thing that made me upset.
43:46
And crucially, I think it seems
43:48
to be the case that when we make up stories like
43:50
this, they're not just you know,
43:52
they're not just outward facing. It's not just pr
43:55
for the brain, it's inward
43:57
facing. We are convincing our cell
44:00
that this made up story is correct. Yeah,
44:02
it helps create like the internal reality
44:05
that we cling to. Yeah, exactly. And
44:07
so it's it's interesting, I think, to
44:10
notice that this appears to be linked
44:12
to the brain's capacity for language.
44:15
That, at least, according to Kazaniga's theory
44:17
here, if he's correct, the part
44:19
of the brain that makes up explanations
44:22
for why something happened is also highly
44:24
associated with the part of the brain that
44:27
is able to talk about things. And
44:29
that very well might not be an accident.
44:32
It seems possible there's a link between the
44:34
networks of the brain that have the most
44:36
to do with generating conscious experience
44:39
and the networks of the brain that are able to
44:41
put things into words. And that's fascinating
44:44
alright. So under under Gazaniga's ideas
44:46
here, the consciousness generating
44:49
capacity is located primarily in the
44:51
left hemisphere. And what happens when you have
44:53
a split brain patient is you essentially
44:55
cut off the conscious part
44:58
of the brain's access to half
45:00
of what the brain is doing, right, Yeah,
45:03
though that half of the brain is still over there
45:05
doing stuff. Yeah. With with each example
45:07
that we we we pull out here, each each
45:10
study, it is still very
45:12
difficult to really grasp. You know. It's it's
45:14
again this kind of you can't see the forest for the tree
45:16
situation. It's hard to imagine
45:19
the consciousness we're experiencing, uh,
45:22
in a in a system that's been divided,
45:25
you know. Well, yeah, that's one thing that that's
45:27
so interesting here. I think one way you could
45:29
misunderstand what the split brain
45:32
cases show is that if
45:34
you cut the brain in half, you generate
45:36
two conscious, independent
45:39
people. And that appears to not be
45:41
the case. People still get the man with two
45:43
brains, like with Steve Martin, right, you get
45:46
one conscious experience. The person generally
45:48
does not report feeling any different,
45:50
as we talked about last time, Their behavior
45:53
and stuff is generally about the same as
45:55
it was before, except you
45:57
have the ability to show under certain conditions,
46:00
there's this whole half of the brain over there
46:02
doing things that you cannot be
46:04
conscious of or put into words. So
46:07
it can still sense, it can still control
46:09
the body is just apparently not integrating
46:12
or synthesizing into whatever
46:14
creates your conscious experience,
46:17
which I mean in a way
46:19
that is that that is sort of like having the other
46:21
fellow in there, in the words of Robert
46:23
Louis Stevenson. Now to bring up
46:25
another literary example. We've
46:28
talked about Peter Watt's book Blindside on the
46:30
program before. I'm sure you remember
46:32
the character Siri Keaton, who loses
46:35
his brains left hemisphere to infection,
46:38
and and and and and as a result of that,
46:40
entire hemisphere is largely or entirely
46:43
replaced with like a cybernetic implant.
46:45
Yes, and this creates a lot of the strange
46:47
psychology of the narrator in that book.
46:50
Yes, yes, I can help and think
46:52
of that. When we were talking about this, also,
46:54
I was reminded of a character in the book Consider
46:57
Felibus by Ian M. Banks, who
47:00
who has tweaked his brain
47:02
so that you can engage in uni hemispheric sleep. We
47:04
didn't even get into that in in this episode.
47:06
But of course this is something that for instance, dolphins
47:08
can do. Uh, it can't just
47:10
go to sleep, so they'll put one side of their
47:13
brain, one hemisphere of the brain to sleep at
47:15
a time. And so and then that particular
47:17
book, it was he was probably leaning
47:19
a little bit into sort of the left
47:21
brain right brain myth a bit,
47:24
but he was discussing how if
47:26
one side of the human brain is sleeping
47:28
and then only one side is awake, you are going
47:30
to have a different expression of that individual.
47:33
Now, if the Gazonica model of consciousness
47:36
is correct, Uh, that wouldn't
47:38
make me wonder that if a human were capable
47:40
of uni hemispheric sleep, would the
47:42
human be conscious while
47:45
the right brain is sleeping and not
47:47
conscious while the left brain is sleeping,
47:50
and yet while the left brain is sleeping,
47:52
still awake, just not conscious. Well,
47:54
I guess you'd ultimately and
47:56
then you'd have to work out exactly how this would
47:58
work in a human scenario. But you as long as one
48:00
side would be awake to alert the other side
48:02
when full brain alertness
48:05
was required, you know that
48:07
would that would be the main prerequisite.
48:10
I just thought to look this up. I wish
48:13
I thought before we came in here, whether there are
48:15
any lateralization properties of sleepwalking.
48:18
Oh, that would be good too. Well, we we need to come back
48:20
and discuss sleepwalking in in depth,
48:23
because I'm sure there's a whole episode just right there. We've
48:25
done some episodes on what paras
48:27
omnia in the past, like sort of covering
48:30
various weird sleep phenomena.
48:33
But yeah, that would be a fun one to come back
48:35
to, for sure. You know. Speaking of Peter
48:37
Watts, I remember he's written about
48:39
this idea of if thoughts
48:41
were inserted into your brain from the
48:43
outside, would you even perceive them
48:46
as alien or would you just perceive
48:48
them as self? Because Gazzaniga's
48:50
left brain interpreter model might be totally
48:52
wrong, of course, but let's just assume for a minute that
48:54
it is correct. Things happen unconsciously
48:57
in modules all throughout the brain, and
48:59
then regions in the left hemisphere have the
49:01
job of synthesizing all that activity
49:03
and generating a story that explains
49:05
to you why your brain just
49:07
did something. And this interpreter
49:09
function is somehow crucial to what we think
49:12
of as the human experience of consciousness.
49:14
Consciousness is sort of is this story
49:17
we tell about why we're doing
49:19
things and who we are now. Normally,
49:21
if something enters your left visual field,
49:23
goes to the right hemisphere, gets processed
49:26
there, and then travels to the interpreter and the left
49:28
hemisphere through the corpus closum. That doesn't
49:30
feel like you're getting that thought or information
49:33
or experience from somewhere else. It's all
49:35
just self. It all just gets interpreted
49:37
and it's you. So if we
49:39
were to start using some kind of brain to
49:42
brain interface or a computer to brain
49:44
interface where it were possible to transmit
49:47
thoughts into the brain from
49:49
outside, and who knows if that's really possible,
49:51
of course, but just assume would we be
49:54
able to tell the externally
49:56
inserted thoughts the sort of incoming
49:58
brain mail from activity
50:00
arising in networks and modules natively
50:03
throughout the brain itself, or
50:05
would it just all go to the interpreter
50:07
the same way. So you could send an
50:09
alien thought into somebody's head and
50:12
have them immediately rationalize it as
50:14
part of the interpret itself the same
50:16
way they would if it came from some network
50:18
in the right hemisphere, would they just think,
50:20
yep, this is just me thinking I
50:23
feel like we're borderline there
50:25
with certain individuals in their use of smartphones.
50:28
Oh yeah, where imagine
50:31
you and I'm listeners out there, you've had a similar
50:33
experience. We would be in a conversation with someone and
50:36
they'll without a phone to remember something.
50:38
But but but often like not in
50:40
a way where it's like, oh yeah, I forget
50:42
that, let me research it, More like, let me
50:44
access this part of my memory. Yes, I know
50:47
exactly what you mean, and I, um,
50:51
I don't know. I mean I wonder what the processes
50:53
by which the interpreter function. Again,
50:56
just assuming this model of the interpreter and
50:58
the conscious experience is correct, mean this, you
51:00
know, this might be mistaken. But if
51:02
this is correct, what is the rubric
51:04
it uses to decide what gets integrated as
51:07
self? And what what
51:09
does it decide is alien? That's a
51:11
great question. We'll have to come back to that in the future.
51:13
Maybe there is none. Maybe it's also maybe
51:15
there's no future. Oh there's maybe there's no self. Yes,
51:18
well, you know it also brings up the question,
51:21
you know, are we limited our
51:23
Is our identity limited by
51:25
the things that we have at our disposal in our mind?
51:28
Do you count the things that we we have to depend
51:30
upon, that we have externalized,
51:33
you know, And I feel like that is part of the modern
51:35
human experience, has been part of the human
51:37
experience for a while. I mean, if an author
51:39
writes, say, thirty
51:42
books, um, and that author
51:44
cannot repeat them from memory, they are not a
51:47
part of his his or her mind. Uh,
51:49
then you know, how do you weigh that
51:51
into the equation of self? Yeah?
51:53
Exactly. And what if you didn't write them? What if
51:55
these are just books that you have incorporated
51:59
into your thinking about things?
52:01
Are those now a part of your brain? If
52:03
you know that, you could consult them
52:06
in order to figure out what you think about something,
52:08
but you can't do it without consulting them.
52:10
Yeah, what if it's a book that you've written and
52:13
you've forgotten. I believe Stephen
52:15
King has a couple of examples of that right wherever
52:17
he doesn't remember writing a particular novel.
52:19
I think one example was Coujo. He said
52:21
they didn't remember writing it because he was on drugs.
52:23
Yeah, so it's Kujo a part of Stephen King
52:26
likewise, I mean, there we all
52:28
have, prehaps the books, films, etcetera,
52:30
some sort of external influence that
52:32
has been important at one point in our life, and
52:34
then is discarded later and then
52:36
sometimes pick back up again. Oh, there's an extremely
52:39
strong social component here. Lots of
52:41
people figure out what they
52:43
think about something by checking to see what
52:45
somebody else thinks about it, whether
52:48
that's a person you know known to them, or some
52:50
public figure that they you know, derive
52:52
opinions from. Yeah, and you know what, I'm
52:54
gonna go ahead and take a stand. That's not behavior
52:57
I encourage do not do not
52:59
trust another person as much as you
53:01
trust your own, right hemisphere, don't
53:03
just directly incorporate their
53:05
their information as as self.
53:08
I can agree with that. Yes, all
53:11
right, well there you have it. We're gonna go ahead and cap
53:14
off these two episodes Part
53:16
one, Part two, Hemisphere, left
53:18
hemisphere right if you will, Uh,
53:21
if you want to check out other episodes of Stuff to Blow your
53:23
Mind, you know where to go. Head on over to stuff to
53:25
Blow your Mind dot com. That's the mothership. That's
53:27
where we'll find all the episodes of
53:29
the show. And don't forget about Invention
53:32
at invention pod dot com.
53:34
That is the website for our other show,
53:36
Invention, which comes out every Monday.
53:39
It is it's very much a you know, a sister show
53:41
to Stuff to Blow your Mind. It covers a
53:43
lot of the sort of topics that we've covered on Stuff
53:46
to Blow your Mind in the past, So
53:48
it's, you know, I wouldn't say it's a you know, radically
53:50
different show, but it's one that if if you're a fan
53:52
of Stuff to Blow your Mind, you should subscribe
53:55
to Invention. And perhaps
53:57
you're even the type of person who you're like, you know what, I
53:59
like being mentioned episodes the most, Maybe
54:01
I'll just stick with Invention. That's fine too.
54:04
Yeah, we basically applied the same kind of mindset
54:06
we do on the show here to scientific topics
54:08
and cultural topics. Over there, we tend to apply
54:10
it more to techno history. So if you like
54:12
what we do here, you'll like what we do there. Go
54:15
check it out, subscribe to Invention
54:17
and rate and review us wherever you have the ability to
54:19
do so. That helps us out immensely. Yeah, oh
54:22
huge, Thanks as always to our excellent
54:24
audio producers Alex Williams and Tory
54:26
Harrison. If you would like to get in touch
54:28
with us directly with feedback about this episode
54:31
or any other uh to suggest a topic
54:33
for the future, or just to say hello, let
54:35
us know how you found out about the show where
54:37
you listen from all that stuff. You can email
54:40
us at blow the Mind at how stuff works
54:42
dot com
54:52
for more on this and thousands of other topics.
54:54
Does it how stuff works dot com.
55:01
The people
55:07
think four
55:12
start four f
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More