Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:00
I'm
0:02
on the Internet to help people learn
0:04
and get curious and do well in school
0:07
and under and more of the knowable
0:09
things of their universe and know more
0:11
about the unknowable things and
0:14
that that's beautiful and it's fun and like I
0:16
get some joy from that.
0:18
You might recognize that voice. Maybe
0:21
you've heard one of his lectures on YouTube
0:23
or seen one of his educational TikTok
0:26
videos.
0:27
So we evolved to have blood production inside
0:29
of our bones, the darkest place
0:31
Evolution could find.
0:34
Hank Green, he's an author.
0:36
A science communicator and a vlogger.
0:39
He got a start on YouTube back
0:41
in two thousand seven. Think
0:44
about that. That was before the term content
0:46
creator was even a thing.
0:49
So he has seen the digital evolution
0:51
play out in real time. You
0:54
know, I I remember sort of it kind of happened
0:56
for me back in like twenty twelve, twenty thirteen when
0:58
the stuff that I started to see getting made on YouTube was
1:00
a lot of sort of cruel pranks
1:03
or, like, manipulative guys
1:06
trying to get. He
1:07
was one of the first people to
1:09
get famous from YouTube.
1:11
He was making vlogs with his brother,
1:13
who's a writer, John Green, and he
1:15
was also doing something that I can really relate
1:18
to. Trying to educate people.
1:20
Along the way, Hank got
1:22
quite a following. A million and half
1:25
followers on Twitter seven million
1:27
on TikTok and tens of
1:29
millions on YouTube channels with
1:32
his
1:32
brother, some of which are shown in schools
1:34
throughout the country. A part of
1:36
the reason I wanted to speak to Hank today
1:38
is because he's also a dad
1:40
and we share some of the same concerns.
1:44
Is my son worse off because
1:46
his dad is an Internet guy and he's gonna
1:48
be on the Internet the whole time all the
1:50
time, like, and there's no way I can be like, you can't
1:52
use social media. He's gonna be like, what do you do?
1:56
So far in the season, I've been asking professors
1:58
and doctors and experts some
2:01
pretty tough questions about screen time and
2:03
social
2:03
media. But now, I wanted
2:05
to hear from someone who makes the content
2:08
that we see when we are scrolling.
2:10
It's hard to feel good about the Internet
2:12
as a
2:13
whole, where back in two thousand seven, it was
2:15
very easy to feel good about the Internet as a whole.
2:18
So how does someone? Whose
2:20
whole job is to be online,
2:23
walk that line. I think now
2:25
we're all kind of comfortable with the idea that the Internet
2:27
is good and bad. And that it's a tool
2:30
and you can build a house with a
2:31
hammer, you can, you know, hit somebody in
2:33
the head with it. Today, my
2:35
conversation with Hank Green
2:38
a content creator and a fellow
2:40
dad about what life is like
2:43
on the other side of the screen.
2:46
I'm Dr. Sanjay Gupta, CNN's
2:48
Chief Medical Correspondent,
2:50
and this is chasing life.
2:57
I gotta tell you, you know, this season
2:59
has been pretty heavy so far. It's
3:02
been very personal. I've had
3:04
conversations with my three teenage
3:06
daughters that have been so illuminating
3:08
and so wonderful and at
3:11
times frightening. To be honest,
3:13
they live another life, their
3:15
digital life, which I simply
3:17
don't know as much about. And
3:19
I think we often fear what we don't
3:22
understand, imagining and anticipating
3:24
all the potential dangers of social
3:26
media and screen time. But, you
3:28
know, I realized something else as well.
3:31
Something that frankly surprised me a bit
3:33
and that is despite the worry
3:35
and the concern I do
3:37
realize that our phones and the internet, it
3:40
can be a sincere source of joy and
3:42
connection and learning for so many
3:44
people. That's part of the reason
3:46
I was so excited to talk to Hank.
3:49
He's someone that kids like my daughters
3:51
look up to. I gotta say I've really
3:53
been looking forward to this one. Hank,
3:55
I I really appreciate your time. Thank
3:58
you so much. I don't
4:00
know why or whether you know who
4:02
I am, but that's cool. I
4:04
do. And and I think that the
4:07
the idea of teaching people
4:09
science in particular, just as a scientist
4:12
myself, I think is super important.
4:14
And I also have three teenagers and
4:16
your videos have actually been shown in their classrooms.
4:18
So when I told them that I was interviewing you,
4:21
they they they also knew
4:23
you. So you cross
4:24
generational, which is pretty cool. That's always
4:26
my that's usually my path and people
4:28
are, like, my children said I should talk
4:30
to you.
4:32
Right? That's good. That's good.
4:33
Yeah. Yeah. Do you have you have children? I
4:35
have one six year old. Boy, six
4:37
year old. So pretty young still. But
4:39
do you have a thought on when you would
4:41
let your child have a
4:44
device and when they might
4:46
be on social media. Howard Bauchner:
4:49
Yeah, I think that those things are really
4:51
important to distinguish that, you
4:53
know, screen screen can
4:55
be anything. But social
4:57
media specifically is
5:00
a tool that we don't
5:02
understand yet. And it changes
5:05
very fast. And we
5:08
we don't we don't know the impact that it has had
5:10
on society yet. So far, like, the
5:13
research is still, like, still very difficult
5:15
to do and it's slow. So
5:17
I I'm
5:20
gonna I think I'm gonna struggle with it because,
5:22
like, already my son, my six year old son will
5:24
look at me and say, dad, stop looking at your phone.
5:26
And I'm like, oh my god. Like, Yeah,
5:29
you're right. You know, and he loves
5:31
a YouTube video, like he'll launch I'm
5:34
very I love YouTube. And
5:36
so I my son watches YouTube, but I watch
5:38
what he watches on YouTube because I also
5:40
know how the YouTube algorithm works and it
5:42
sort of starts to creep away from the
5:45
more educational stuff that I'd like him to be watching.
5:47
Yeah. And I have to say, these are
5:49
tough conversations think as parents. I
5:51
I I've gone through this. Hank and so I have
5:53
seventeen, fifteen, and thirteen. And -- Mhmm.
5:56
-- my seventeen year old the other day said to
5:58
me, I probably wouldn't let
6:00
my kids be on social media as early
6:02
as I
6:03
which was god. It's little bit of a, like, punching
6:06
the gutty here because I thought I was, you know I'm
6:08
gonna be better parent than you, dad.
6:10
Exactly a bit much. But
6:14
I guess the flip side of that is that
6:17
she's very aware of
6:19
of -- Mhmm.
6:20
-- what it's done for her, done
6:23
to her, you know? I mean, and she would she'd
6:25
dial some of that back, which I think in some
6:27
ways may be a little bit optimistic. Like, I I feel
6:29
like a lot of times the the inclination and
6:32
we see this in medicine all the time is to assume
6:34
the worst case scenario. We have to assume
6:36
the worst. Hope for the best
6:38
but assume the worst. And everything that we do
6:40
in terms of our our how we respond
6:43
is is that that worst
6:45
case scenario. But
6:47
that is not the case for most people.
6:50
So how to find the balance
6:51
there? And I I don't know the answer. And I don't
6:53
know that you know the answer, but I'm just curious how you think
6:55
about it. Yeah. I I mean,
6:57
I this
6:59
might be a little bit can
7:02
I can I really believe this about myself? But I
7:04
feel like there, like, you can develop
7:06
expertise in how to use
7:09
a tool well. And so the
7:11
the question that that comes in up
7:14
in my mind is like, is my son
7:16
worse off because his dad is a Internet
7:18
guy and he's gonna be on the Internet the whole
7:20
time all the time, like, and there's no
7:22
way I can be like, you can't do social media. He's gonna be like,
7:24
what do you do all day every day?
7:28
So what I would like is
7:30
and what I would hope is that as sort
7:32
of more people who were who had
7:34
social media as part of their lives, when
7:36
they were younger, have children, that
7:39
we're at least able to have these conversations better
7:42
or or talk about, like, know,
7:44
any any new thing that
7:46
enters into a society is very easy
7:48
to abuse until you sort of develop
7:50
norms and taboos about how to use it. More
7:53
well. But
7:55
I think also just like any other
7:57
sort of thing that can feel
7:59
really good without a lot of it's
8:01
a little like candy. You know, I -- Mhmm.
8:04
-- where where, you know, it it it tastes feel good,
8:07
but it doesn't provide a lot of substance. And
8:09
and so I my hope is
8:12
that we can get to a place where where we know
8:14
how to use it. You know, it's
8:16
okay to have some but you're
8:18
gonna have to have some real food too.
8:22
You know this idea of the metaphor of
8:25
junk food versus real food? I
8:27
think that really resonates with me. I think that's
8:29
an interesting way of framing this.
8:32
Because, yes, there is content
8:34
like Hank's which is educational and
8:36
scientific and well thought out.
8:39
But there's also a lot of junk,
8:41
and therein lies the problem. You know,
8:43
recently, was having a conversation about this
8:45
with my youngest daughter, Solay, and
8:47
she had shown me this meme on Instagram.
8:50
And it was funny. We laughed.
8:53
But it also wasn't true.
8:55
And I asked her, I said, so
8:57
let you know this isn't true, right?
8:59
And she kinda laughed and showed me her
9:01
phone again, yeah, it's on Instagram. And
9:04
I said, right, right, but you know it's
9:06
not true. And what
9:08
she said next is something that really
9:10
kinda stopped me in my tracks. What
9:13
she said is dead, to be honest, I
9:15
really don't think that anything that I see
9:18
anymore online is true.
9:21
Think about that. So much of
9:23
what comes across her feed is garbage.
9:26
So in essence as a result,
9:29
it all becomes garbage. It
9:31
all becomes suspicious. It
9:33
is all lumped together and suddenly
9:35
you find yourself in a really kind
9:37
of scary place where nothing Nothing
9:40
at all can be trusted. Social
9:44
media in a lot of the Internet is just
9:46
a playground for them, not to
9:48
be taken seriously. And that
9:50
drags everyone down. It drags
9:52
everything down. Even for
9:54
someone like Hank, who by all
9:56
accounts makes good, credible,
9:59
fact checked vetted content. We're
10:03
talking about, like, sense making at this point.
10:05
Like like, there's always been structures
10:10
of credibility. You know, if
10:12
you go to the beginning of newspapers, everyone
10:14
knew that every newspaper had a
10:16
a perspective. Like there was a communist newspaper
10:18
and there was a Republican newspaper and there was sort
10:21
of a business newspaper. Like, and we have
10:23
that to some extent now and and we
10:25
have to sort of learn the shape of
10:27
the information landscape. And
10:31
that when I was grown up, in
10:33
the eighties and nineties, that was very
10:35
that was more much more clear. And
10:37
now it is very not clear. And
10:40
so one thing that I try to do
10:42
very hard is to get stuff right
10:44
when the content that I'm making. And when I get
10:46
stuff wrong, talk about the fact
10:49
that I got it wrong and why. And
10:53
and and that's almost for
10:55
a lot of people, like, more instructive to be
10:57
like, oh, so, like, not only
10:59
is, like, is everyone fallible,
11:02
but also you can sort of walk down the path
11:04
of how you what you what
11:06
what assumption you made that led to your wrong
11:09
content in some way. But
11:11
I worry about it being so person focused
11:14
because a person almost
11:16
is, like, has to be less credible than
11:18
a really good organization
11:21
because a person can only do so much
11:24
and a person you can't scale
11:26
them the same way and you can't you
11:28
can't build in fact checking the same way
11:30
as you as you can with, like, a strong,
11:33
robust news organization. And so if
11:35
we end up in situation where we only believe
11:37
individuals, I think that's
11:39
a worse world. This is obviously
11:41
something that you're you're quite good at, and I'm not I'm
11:43
not saying this is to flatter you, but a lot of
11:45
people do pay attention to you and your work
11:47
is also shown in schools, you
11:49
know, and people vet these things out, and they
11:52
wanna obviously educate their kids in a good way.
11:54
But how do you you will admit
11:56
when you've got something wrong. That's one thing
11:58
you said. How much time do you spend
12:01
thinking about the institution of trust with
12:03
your own content? You want to get it
12:05
accurate, but just the idea of trust, everything
12:07
from word choice to, I
12:10
don't know, your background to your your
12:12
presentation. Like, how how much
12:14
do you think about that in this digital
12:16
world? I mean, I think about I
12:18
think about it all the time. It's one of the things I
12:20
worry about it and and
12:23
on a lot for a lot of different good
12:25
reasons. You know, I worry about
12:27
it for my own. I
12:29
think a lot people would be
12:31
pretty devastated if, like, I didn't
12:34
live up to the -- Yeah.
12:36
-- the the sort of, you know, what
12:38
what I what I've been trying to portray
12:40
publicly as, you know, the parts of
12:42
me that are me. The other thing I like
12:44
to say about making content on the Internet is
12:46
that, like, I all of me that you see
12:48
is me. You just don't see all of me. And
12:51
I worry a lot about, you know,
12:53
I've seen people both
12:56
in just mistakes and in, like,
12:58
really intentional ways, do things that
13:01
have really destroyed a lot of their credibility
13:03
And they're not just destroying their own
13:05
thing. They're destroying something else that
13:08
is bigger and it's inside of other people
13:10
and that's the thing that I wanna be most careful
13:12
with.
13:15
Hank interacts online a lot more
13:17
than I do, and that can
13:19
take a toll even for a professional
13:21
like him. When we come back,
13:24
avoiding that urge to
13:26
get into it for your own good.
13:29
I've gotten better over the years at realizing
13:32
that you can't really argue with a professional arguing
13:34
without becoming a professional arguing and that's
13:36
than what I do for a living.
13:44
This message is from Abbott.
13:46
The next potentially dangerous virus
13:48
could be out there right now. We may
13:50
not know what it is or where it will strike,
13:53
but an international group of experts
13:55
is on
13:55
it. The Abbott Pandemic Defense Coalition
13:57
is a global team of scientists hunting
14:00
down the next viral
14:01
threat.
14:01
Doctor Gavin Chlarity is one of them.
14:04
A virus hunter looks for
14:06
emerging viruses that
14:08
may pose a threat, there are millions
14:10
of pathogens capable of jumping into people
14:12
at any
14:13
time. A viral outbreak can become global
14:15
pandemic in days. And once the virus
14:17
hunters uncover a new pathogen, they
14:19
move to the next phase. And the clock
14:21
is
14:22
ticking. We can react very quickly. The
14:24
sooner we find a new virus, the quicker that
14:26
we can start building diagnostic tests,
14:28
making the public health community aware
14:30
but something is out there that could pose a
14:32
threat and potentially prevent an expander.
14:35
Watch the virus hunt, a series by Abbott,
14:37
and learn more about the Abbott Pandemic Defense
14:40
Coalition at abbot Discovery
14:42
dot com.
14:46
Let's face it. There's no shortage when it
14:48
comes to opinions, but when it comes to curating
14:50
the most important stories of the
14:52
day, CNN five Things is one of a kind.
14:54
I'm cohost, Krista Beau. And if you wanna
14:56
know about the top five news headlines happening
14:59
at any given
14:59
moment, in under five minutes, then
15:01
look no further. We're bringing you the facts
15:04
fast. To listen until your device play
15:06
the podcast, CNN five. Thanks.
15:17
And now back to chasing life. And
15:19
my conversation with Hank Green. You
15:25
know, Hank started posting videos at
15:27
a time before YouTube stars were
15:29
much of a thing. Keep in mind, this
15:31
was two thousand seven. And
15:33
because of this, got a pretty unique
15:35
perspective as someone who kind of struck
15:38
gold and gained internet fame,
15:40
a following that so many young people nowadays
15:43
really aspire to. After
15:45
spending so much of his career online,
15:47
Hank decided to write about
15:49
these experiences, sort of,
15:52
Hank published two books that touch
15:54
on Internet fame and the digital
15:56
world. They are called an
15:58
absolutely remarkable thing and
16:01
a beautifully foolish endeavor. I
16:04
love the titles of both of them. And
16:06
again, while these books are fiction, Hanks
16:08
says he was definitively drawing inspiration
16:11
from his real
16:11
life. I mean, I was working through
16:13
it. It's such a it's a weird
16:15
job and not a lot of people have had it
16:17
and was twenty seven when I uploaded my
16:19
first YouTube video. So I was married and
16:21
adult. you know, have
16:24
had a pretty stable set
16:26
of situations That is not
16:28
the true for most people. Like, most of
16:30
the people who were, like, my colleagues
16:33
in that era were were, like, eighteen,
16:35
nineteen, twenty, maybe even younger
16:37
than that. And it was a lot
16:39
easier for them to make worse decisions. And
16:43
because, you know, for
16:45
all the reasons. And and
16:48
and I kind of wanted to walk through, like,
16:50
just sort of let let a
16:52
character make some mistakes for me that
16:55
so that I wouldn't make and also
16:57
explore so sort of a lot of how I how
16:59
I see the Internet
17:00
now, how I think it's gonna be in the future,
17:03
and how good it is at turning absolutely
17:05
anything into a fight. How
17:07
much time do you spend online?
17:10
Obviously, you're spending time online to
17:12
to make your content, but just as a user,
17:15
of of these various platforms. How
17:17
would you characterize your use?
17:20
The parts where it it would be tricky to
17:22
call it work I could probably make
17:24
the case of just to justify it to myself.
17:27
It's probably on the order of
17:29
two to three hours per day. The weird
17:31
thing is that, like, what is online? Everything
17:33
is online to some extent. And so,
17:36
like, Netflix is online. not
17:38
counting that. But
17:40
in terms of, like, using Twitter and TikTok
17:42
and YouTube, consuming content,
17:45
doing social media, social stuff,
17:48
It's hours per day. Probably two or three
17:50
hours per day.
17:51
Have you ever thought that it's too
17:53
much? Have you ever wanted to cut back?
17:56
And if so, how did how did you do it?
17:58
I feel like it's a thing that you have to get good
18:00
at. The times when it's been destructive
18:02
for me are when I have been using
18:05
it, like it's I'm almost looking
18:07
for reasons to be upset. And
18:09
I'm I'm like, I've I've stumbled across
18:11
a piece of the Internet or or like a
18:13
sort of set of content creators who are
18:16
saying things. Specifically, what bugs me
18:18
probably the most of all things is
18:20
when they're taking some, like, nugget of
18:22
scientific truth and then usually
18:25
biological and then applying that to
18:27
a social system or to politics in
18:30
a way that is really appealing, but
18:32
really it's sort of backtracking to the prospect that
18:34
they wanted to have. So it's, like, grabbing
18:36
some science y thing and
18:38
telling, like, doing unscientific things
18:40
with it where I'm, like, I wanna just no.
18:42
I'm so mad. You can't just say things.
18:45
You know? And
18:47
But of course, you can't just say things. And,
18:50
like, I can get tied up in
18:52
in being sort of angry that the world
18:54
is not the way that I wish it were. And
18:57
and my wife kind of can quickly identify when
19:00
I'm in that space. She's like, there's something that you're mad
19:02
about that like has nothing to do with what
19:04
you do. You also have to recognize
19:06
that, like, the present like, you can't
19:08
fix everything. And
19:09
so, like, you have to focus on what you do and
19:11
what you're good at and what you actually like instead
19:14
of you
19:15
know, it's it's harder to be a good dad and
19:17
a good partner and a good leader
19:19
at my businesses if I'm sort
19:22
of caught up in some thing
19:25
that, you know, I think that the
19:27
Internet is somewhat designed to to catch us
19:29
in conflicts because that
19:31
is really good at keeping us on the platforms,
19:33
which is ultimately what the platforms
19:35
are designed to do and what their algorithms are designed
19:38
to do.
19:39
Your your your your wife sounds very wise,
19:42
obviously. I
19:44
I am curious like these these videos
19:47
or content that you're consuming that do
19:49
rile you up in the way that you described. How
19:52
does it get to
19:53
you? Is it is it part of your feed?
19:55
Is it being fed to you by one of these algorithms
19:57
you're describing? Part of my like
19:59
the fun part of my job is that if people will
20:01
ask me questions, And so
20:04
one that I saw just this week, it came
20:06
to me because several people
20:08
on TikTok had seen this video,
20:10
which got millions of views and
20:12
it was talking about some things in a scientific
20:15
frame. And a bunch of people
20:17
had tagged me and they said, Hank Green, is this true?
20:20
And I mean, it was not. Like, you're just
20:22
trying to take something that sounds
20:24
science either lend credibility to your argument.
20:27
But I, you know, I, like, I get in
20:29
I get into that and I get it kinda gets
20:31
my blood pressure
20:32
going. But I've gotten better
20:34
over the years at realizing that
20:38
One of the things I
20:40
realized is that you can't really argue with a professional
20:42
arguing without becoming a professional arguing,
20:44
and that's the one I do for a living. I don't argue
20:46
with people for a living. I'm not on the internet
20:48
to yell at people. I'm on the internet to help people
20:51
learn and get curious and do well in
20:53
school and and
20:55
understand more of the knowable
20:58
things of their universe and know more
21:00
about the unknowable things and that
21:03
that's beautiful and it's fun and, like, I get
21:05
some joy from that. And I also think that
21:07
it does more good than getting enough fight,
21:09
you
21:09
know. It's easy to fire
21:11
people up, you know, to appeal to their their amygdala,
21:14
you know, their emotional centers of the brain, and
21:16
that we'll get lot of provocation
21:18
and probably a lot of views. Do
21:21
you do you resist the urge
21:23
to to go that way, to to just
21:25
provoke because that
21:28
would probably, you know, maybe get even
21:30
more people fired up
21:32
and sharing your videos.
21:34
Yeah. I'm lucky to be in a situation
21:36
where I don't my my goal right now
21:38
isn't to get more attention
21:40
or even
21:43
money or whatever it is. But
21:46
what is pretty clear from research is
21:48
that shouting at people and
21:51
talking about how wrong they
21:53
are and consequently how
21:55
bad they are and, like, creating
21:58
you know, set setting up the dichotomy of
22:00
the battle of the conflict is
22:03
not good for your cause.
22:05
Like, it does not convince people of things.
22:08
It pushes people to
22:09
sides. That's all it does. They're
22:11
riling up people who are already on
22:14
their side, but maybe alienating others.
22:16
Yeah. Alienating others and and, like, rallying
22:18
up the people on the other side,
22:19
they're becoming fuel for the for their
22:21
opponents by being this way. I mean,
22:24
people when they come up to you in real
22:26
life, IRL, it's it's probably very
22:29
complimentary and thankful and grateful for,
22:31
you know, things that you've taught them and taught
22:33
their kids. And I
22:36
imagine that it's the same way online
22:38
in terms of the comments section and
22:40
stuff like that. Do you read comments
22:43
And do you feel like the response
22:45
that you get in the digital world mirrors
22:48
what you get in in in the real world?
22:51
Yeah. Yeah. For the for the most part
22:53
and and, like, there are
22:55
comments that are accrual to
22:57
me sometimes, and there
22:59
are be people who disagree
23:01
with things that I've done in the past in
23:04
in one way or another because, you know, I
23:06
I am not shy about my feelings
23:09
on some controversial topics here and there.
23:12
And the and so
23:14
I'm like, you know,
23:16
I'm kind of fine with that. It's
23:20
take I could think it's taken time to get there
23:23
and to understand the dent
23:25
to which I am
23:28
some and kind of not a person
23:30
to some people. They don't see me
23:32
they see me as a sort of a shell that contains
23:35
a brand or a I
23:38
don't don't know. And and also,
23:41
I have gotten comfortable with the
23:43
both the idea and the reality
23:46
that I have more power than those people and
23:48
so they feel like they can throw a
23:50
punch and I won't feel it. And
23:53
and also if I throw a punch back, they
23:55
will feel like I just,
23:58
like, stepped on them with a
24:00
transformer robot
24:02
foot. Like, the that
24:04
that, you
24:05
know, you you kind of when you when you're
24:07
on the Internet and you have you have
24:10
following and you have, you you
24:13
have, you know, a status for lack
24:15
of a better
24:16
term. You kinda have to understand that
24:18
you you wield a lot more power than
24:20
you feel like you wield. And I watch people
24:22
all the time, and I really bugs me people.
24:24
Like, I'm not gonna censor myself just because
24:27
I have an army of angry
24:29
people who will attack anyone I attack
24:32
and that we will come into their lives and make
24:34
their themselves absolutely miserable to the point where they
24:36
have to delete their Internet history. I'm not gonna censor
24:38
myself just because of that. And I'm like, well, it's not
24:40
really censoring yourself.
24:42
To not to to, like, recognize that you
24:44
have more power than you once had.
24:47
Right.
24:48
And but that's you
24:50
know, that's the whole separate conversation. But
24:52
it's a good self awareness though. I mean,
24:54
to to recognize that you'd be punching down,
24:57
so to speak, and with a pretty heavy blow
24:59
if you decided to engage in that way,
25:01
whereas the flipside is that people
25:04
may not recognize that they could leave a mark on
25:06
you. You
25:06
know, just because you are They have no idea.
25:09
They have no idea. And people have been,
25:11
you know, we've John and I and my brother have
25:13
made this content And we we have been through times
25:15
where there's been pretty large groups of people
25:17
who, you know, just kind of for fun don't
25:20
like us. And They have
25:22
they think it's just fun and that we're having
25:24
a good time
25:25
too, and we aren't, but, like, we can't say anything.
25:27
These have to, like, live through it.
25:29
You know? It's it's a bad part of
25:31
a good thing. You seem to
25:33
feel like it's gonna get worse before
25:36
it gets better. Yeah. You
25:38
did think it was gonna get better. I mean, it seemed like you
25:40
did anyways. Is that how you still is
25:42
that what you still think?
25:45
Yeah. Absolutely. I don't know
25:47
how it gets better. I think that it's
25:49
gonna be I think it's gonna require lot
25:51
of individual people having a lot of
25:53
good thoughtful thoughts. And a lot of
25:55
good conversations and a lot of time
25:58
and experience with this. think
26:00
a lot about the printing press we
26:02
suddenly had the ability for
26:05
people who disagreed with the
26:07
church, with the Catholic church to
26:09
take it on and say, like, I think that
26:11
you're doing it wrong. And we're gonna share that information.
26:14
We're gonna better at it than you, and we're gonna be more
26:16
nimble than you. And has so many
26:18
parallels to the the sets of conflicts
26:20
that we have now. And it was
26:22
a messy, messy time. It was
26:24
it was very bad and lots of people died and
26:26
it was it was a but like nobody
26:29
thinks we shouldn't have books. We
26:32
figured out how to have books and have them not
26:34
be societally destructive. We figured out
26:36
how to move through that time where it was probably
26:38
for the best that we like, that that like,
26:40
we didn't we shouldn't have lived in a world.
26:43
Where the church had that much power. We needed
26:45
to move into a world where there was more individual
26:47
agency. And I think we're having that
26:49
now. Okay. And that's not a conversation
26:52
about young people and screen
26:53
time. That's conversation about every single
26:55
one of us in the society we exist
26:57
in right now.
27:00
You write these books, and and I'm curious because
27:03
I feel like just reading these books and
27:05
listening to you, it's an opportunity for you
27:07
to get these thoughts out through these
27:09
through these novels, fictional scenarios.
27:12
Do you have an idea or did you go
27:14
through the thinking of what a
27:16
healthy, idealized version
27:19
of social media could look
27:21
like.
27:22
I do think about that The thing that
27:24
I'm working on now, which don't know if it'll ever
27:26
come out, is is about what think the Internet
27:29
could be like if we were. Thoughtful
27:32
and careful about it, but it is very different from
27:34
what we have now. And I think that it's so
27:36
like, it's you never feel like you're in history
27:39
when you but we always are, you know.
27:41
You never feel like you're part like, that this
27:44
moment that we're in right now is gonna be part of
27:46
a much larger story than most of these things
27:48
that we're talking about. And twenty twenty three are
27:50
gonna be entirely forgotten. But
27:53
but we are and we're at the beginning of
27:55
this revolution in
27:57
communications I know people
27:59
who've run big social media companies and
28:02
and they do think about the societal implications
28:05
of what they do and they consider society
28:07
and the world, one of the one of their
28:09
stakeholders along with their employees and their advertisers
28:12
and their investors and whatever. But, like, mostly
28:14
the thing that we're we're fine with is
28:16
like, okay, you've got this technology, use it how
28:18
you can and make as much money as you can because that's
28:21
how, like, that makes sense. We
28:23
haven't really thought about how to do it in a
28:25
way that's really a pro social. We haven't
28:27
thought about how to sort of make the
28:30
tool best
28:32
for a human and best for
28:34
human outcomes. Because that's really complicated
28:37
and it's kind of scary to say, like, oh, I'm gonna use
28:39
the social media platform. It's gonna make me happier. She's
28:41
like, whoa. That
28:43
seems like they shouldn't be able to push those buttons.
28:46
But they can. So
28:48
what is what is that feature look like? And
28:50
and is it dystopian? Or is
28:52
it utopia? Or is it a little bit of
28:54
both because that's sort of what the future always is.
29:00
The future. A little
29:02
dystopian, a little utopia,
29:05
all of it at the same time. It's
29:08
a really kinda beautiful way to look at
29:10
it, an authentic way to look at it. It's true.
29:13
We never know which way
29:15
things are going to unfold. You will be
29:17
surprised. If the impact of these
29:19
technologies will get better, will
29:22
they get worse? Over time, maybe
29:24
it's gonna be both. All that we can
29:26
really try and do is under than
29:28
them. These tools at the
29:30
time we are using them. We can develop
29:33
an expertise in how to use them well like
29:35
Hank said. And at the same time, we
29:37
can sharpen our awareness of how
29:39
they can be detrimental to us. Hank
29:42
mentioned that even for him, someone
29:44
with years now of experience the
29:46
pressures of being online can
29:49
take a toll. And he's probably got
29:51
thicker skin than most. And frankly,
29:53
that troubled me. Because Hank
29:55
is Hank, but I worry about regular
29:57
people. I worry about young people.
30:00
This is something I'm hearing about even from my own
30:02
kids. How people see you,
30:04
I guess, with social media. You wanna
30:06
put out, like, a good picture of yourself, make it
30:08
seem like
30:10
feel like your life is so perfect. Even
30:12
though not everyone's life is perfect. Coming
30:16
up next week, a conversation with
30:18
my daughter, Skye. About
30:21
the pressures of being a teenage girl online.
30:24
This is a conversation I will never forget.
30:27
Plus, I'm gonna sit down with child psychologist about
30:30
the impact this pressure can have
30:32
on young
30:32
people. So we're all just in this like
30:34
feedback loop of looking at perfect pictures
30:36
and perfect photos even though we know that's
30:39
not reality. So we're just comparing our worst
30:41
dates, our worst moments, our worst angles
30:43
to other people's best. And of course, you're
30:45
gonna not feel great when you do that.
30:48
Thanks for listening.
30:57
Chasing Life is a production of CNN
30:59
audio. Our podcast is produced
31:01
by Grace Walker, Xavier Lopez,
31:04
Aaron Matuson, and David Rind.
31:07
Our senior producer is Haley Thomas.
31:10
Andrea Kane is our medical writer and Tommy
31:12
Bozerian is our engineer. Dan
31:14
Dezula is our technical director.
31:17
The executive producer of CNN audio
31:19
is Steve Liqtai, and a special
31:21
thanks to Ben Tinker, Amanda Sealy,
31:24
and Nadia Kanang of CNN Health,
31:26
and Katie Hinman.
31:34
This week on the X Files with David Axelrod,
31:37
former GOP rep Peter Meyer of
31:39
Michigan. First of
31:40
all, can Republicans win that see if Donald
31:42
Trump is the nominee of the party. I
31:46
think they can. Again, I don't think that Donald
31:48
Trump would be the nominee. Would you vote for him
31:50
if you were?
31:52
I do not want to get to that point. Listen to
31:54
the ax files with David AxcelRod on
31:57
Apple Podcasts. Spotify, iHeartRadio,
32:00
or your favorite podcast app.
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More