Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:00
Hey there, it's Chris Gautier, producer on Masters
0:02
of Scale. I've got a
0:04
small but really important request for you. It
0:07
seems that Apple's recent update might
0:09
have paused your podcast feed. But
0:12
don't worry, there's an easy fix. So
0:14
for our Apple listeners, here's what you do. Open
0:17
the podcast app and search for Masters of Scale.
0:20
Then go to our show page, look
0:22
at the top right corner. If you
0:24
see a pause button there, simply tap it
0:26
to continue receiving episodes in your feed. If
0:29
you see a download button, go to
0:31
settings and set it to automatically download
0:33
episodes. And if you
0:35
see a plus button, tap it to follow the
0:37
show. This way you won't
0:39
miss out on any new episodes of Masters of
0:42
Scale. And if you have a minute while
0:44
you're there, we'd be so grateful
0:46
if you'd share a five star rating and
0:48
review for the podcast. This really
0:50
helps new listeners discover the show. In
0:55
San Francisco today, President Biden convened
0:57
a meeting of artificial intelligence experts
1:00
to weigh its risks. The
1:02
White House releasing details of a new
1:04
executive order that would direct
1:06
federal agencies to regulate as well
1:08
as shape the world. This artificial
1:10
intelligence expands the boundary of human
1:13
possibility and tests
1:15
the bounds of human understanding. This
1:18
landmark executive order is a testament to
1:21
what we stand for, safety, security,
1:24
trust, openness. This
1:27
is likely to be the most aggressive
1:29
piece of American regulation over the fast.
1:31
The White House is calling this the strongest
1:33
action that any government has taken on artificial
1:35
intelligence, safety and security. Hi
1:42
everyone, it's Bob Safian, former editor of
1:44
Fast Company, founder of the Flux Group
1:47
and host of Rapid Response. My
1:49
colleague Reed Hoffman met with President Joe
1:52
Biden in San Francisco several weeks back
1:54
with the topic of AI high
1:57
on the agenda in this
1:59
very special episode. episode, Reed gives
2:01
us an exclusive window into
2:03
this closed door meeting. Biden's
2:06
the first American president who's had
2:08
to face the full challenge of
2:10
AI from regulation to its emerging
2:12
impact on the economy, jobs, social
2:15
discourse, and of course its potential
2:17
for misuse by bad actors. Reed
2:20
talks with me about all this and more.
2:23
It's good stuff. So let's get to
2:25
it. All
2:36
right, great. Well, Reed, let's dive in.
2:39
Let's dive in. You
2:42
had an opportunity recently in San Francisco
2:44
to meet with President Biden. You talked
2:46
about AI, among other things. You haven't
2:49
spoken about that elsewhere. So
2:51
what can you tell us about it? You
2:54
know, I think people would be
2:56
surprised because they think, oh, Biden,
2:58
wonderful president, gentleman of an older
3:00
generation, older than both of us.
3:03
So I think a little bit of the surprise
3:05
is that he's actually completely with it. He was
3:07
like, I see a lot of discussion happening here.
3:10
There's obviously a lot of unknowns. It's obviously
3:12
moving fast. And this is one of the
3:14
key things that I think is really, really
3:17
excellent about the man and about the man
3:19
as president was he
3:21
doesn't show up going, well, I read something and I know
3:23
what it is and let me tell you what it is.
3:25
It's the, look, I recognize a lot of stuff happening here
3:28
and I'm not expert. So how
3:30
do I approach it? What do you tell me? Here
3:33
are my questions. Are there more important questions?
3:35
Here are the people that I have working on this. Are they working
3:37
on the right things? Are they talking to the right people? So
3:40
he brings an open mind and asking
3:43
questions to it. And
3:45
his goals are Most
3:47
fundamentally being kind of a man of the
3:49
people. He's not like, okay, look, great to
3:52
hear that the bungee different opportunities for American
3:54
industry and great to hear that startup people
3:56
are generally actually thinking about these ethical questions,
3:58
you know, that we're. Having discourses like the
4:01
one we're doing today and that's really important.
4:03
But like, what does it mean for the
4:05
bulk of Americans? Like what's gonna go on,
4:07
how to their lives become better with as
4:09
what does it mean for their jobs? What
4:11
does it mean for the way they care
4:13
for the families? What does it mean for
4:15
education and that kind of thing was actually
4:17
his top of mine questions is obviously a
4:19
whole set of other questions come into, but
4:21
that was the ones he most wanted to
4:23
go. I don't want to leave the room
4:25
without getting a sensitive. I
4:27
mean you know we have every when answering the question
4:30
about what is the future very I look like to
4:32
them and and serious. Did. You get
4:34
a sense from him whether he's an Ai
4:36
Optimists to earn a i pessimists say you
4:38
know it's there, it's you know the regulatory
4:41
him as part gonna get foot do get
4:43
a feel for that. Well
4:45
I think see as an optimist
4:47
because is an optimist but the
4:49
american people he recognizes it's be
4:52
bad people doing bad things is
4:54
not blind. But. He wants
4:56
to say we can be better, We can
4:58
be the people we aspire to be and
5:00
we can build interesting think so. He is
5:03
inherently optimistic in a view of it is
5:05
possible to shape this positively and I think
5:07
isn't that much because he goes. I know,
5:09
but the technologies that would. I'm a technological
5:12
actress and let me tell you why should
5:14
be optimistic that know that's not What is
5:16
a zoo like Diesel been one of the
5:19
noses into? There's Nobody Goes Look Ice. I
5:21
believe in people caring, what the future, collaborating
5:23
together, regular habit and like, what should I
5:25
do? To be helpful to that
5:27
which can include the kids would try
5:30
to stop the things that are bad.
5:32
And here's the things we do to
5:34
support a goodbye Seats More than a
5:36
I optimist, but not because he goes
5:38
I'm an expert in psychology, but more
5:40
because she's a believer in human beings
5:42
being able to do good things. Are
5:45
so me? Take it to the other
5:47
side a little bit. Did you get
5:50
a sense for their biden? wants to
5:52
lead you to wants to be ahead
5:54
of the curve? the in sort of
5:56
setting parameters from a regulatory and government
5:58
framework, or whether. He's sort of
6:01
okay if some of that is done.
6:03
You. Know in other places and see kind of
6:06
fills in around it. I. Think
6:08
your approach has been unusually wise because
6:10
what they first that is, bring a
6:12
whole bunch of companies and innovators into
6:14
the White House and say we're going
6:16
push you want to set of on
6:18
Turks Moments need to be shown irresponsible.
6:20
You need to be making commitments that
6:22
aren't just like slogans, but also like
6:24
things that you're actually navigating risks and
6:26
you're investing in it and you're doing
6:28
it the way of telling, Getting a
6:30
sense of which things are they saying?
6:33
We. See this an issue and we could
6:35
do this and will even make volunteers commitments.
6:37
Then we take those volunteer commitments and
6:39
we start talking to. The. World
6:41
at large university people and say
6:44
okay, well what are we missing?
6:46
What is not yet They're. Ok,
6:49
let's hear that let's hire some people
6:51
specifically about doing it so let's have
6:53
a deputy. The staff was focused on
6:55
this that take the best person that
6:58
we currently have working within the broader
7:00
white house, make them work for the
7:02
deputy to the staff. Let's get the
7:04
various multiple agencies coordinating together and then
7:06
let's get them all engaged in this
7:08
which is one of these offers. You
7:11
end up with a longest ever executive
7:13
order I think you're hundred ten pages
7:15
and let's make sure that we are
7:17
all working on this really important thing
7:19
and we're bringing met together and then
7:22
in what we're releasing we're not going
7:24
and here the final thing as like
7:26
look let's put out some things that
7:28
are amplifications of the voluntary minutes for
7:30
an industry godlike. Let's make sure that
7:33
you're in dialogue with us and let's
7:35
make sure that a number of different
7:37
agencies are paying attention to it. And
7:39
so all of this I think was
7:41
a masterclass In effective Like how do
7:44
we take something dynamic which we don't
7:46
know so fully understand either is industry
7:48
or governments. Understand the future could be
7:50
so much better than the present and that we
7:52
need to get there like even on a safety
7:54
thing the tools for didn't the safety or more
7:56
and more the beach. Now that being said, I.
7:59
Think they want could be
8:01
kind of setting a baseline
8:03
norm. As. Regulatory leaders like
8:05
to your specific questions and what's more
8:08
as I think they're willing to learn
8:10
from other people. So part of the
8:12
Iau they were talking the Uk Uk
8:14
had done a Uk A I say
8:16
to suit the said we should up
8:18
on to the to should be coordinating
8:20
disease seven I think is going to
8:22
be very centered on on a I
8:25
next year and so all of that
8:27
I think is a very good thing
8:29
to be coordinating and I think she
8:31
doesn't want to go. Okay let's find
8:33
other places will drag. Us into a signal
8:35
or we want us. we want to stab us a
8:37
good norm which is both the right amount. Or
8:40
too little, too much rights and the
8:42
thing that gets us into the future.
8:45
That's what I think they are doing
8:47
in a very good way and are
8:49
being leaders. But. Not leaders
8:51
have like no one else needs to do anything.
8:54
We're. Participatory leaders, you know listening to you
8:56
and I'm sure everyone else let's be here
8:59
is say to themselves. Well I understand why
9:01
the President wanted to talk to this guy.
9:03
You know I don't really understand miss all
9:05
of this stuff, but this is someone
9:07
who understands a lot of it's for you
9:10
going into talk to him. With.
9:12
Their goal that you have out of
9:14
it like is there something that you
9:16
took away. That. Was valuable to
9:18
you from that experience. While
9:21
I think there are a number of things
9:23
I mean is obviously stuff that he's got
9:25
decades of experience that I have. No.
9:28
Experience on and I never meet with
9:30
people without having things that I would
9:32
ask questions of them as is Bob
9:35
you know for was birds that are
9:37
gonna bring me I was at least
9:39
our question I would say. The thing
9:41
that I. Probably. Most
9:43
go away from in that was this
9:45
notion of. What's. The
9:48
way that you should
9:50
set. a set of
9:52
expectations and kind of a political compass around the
9:54
everyday person like what's the concern of someone as
9:56
a look i'm i'm in the tech industry i'm
9:59
not doing this I'm trying to
10:01
navigate the experience of my life, things like gas prices
10:03
and other kinds of things. And I'm trying to navigate
10:05
that, because that's one of the things that I think
10:08
Biden is like, okay, what is that
10:10
person's experience of their life, of their
10:12
work, of their family? And
10:14
how can we be making that better year by
10:16
year? And how do we navigate
10:19
crises like the pandemic, which he helped get us out
10:21
of and everything else, in terms of
10:23
making that happen? And so, like, okay, what
10:25
does AI contribute to that? Not, hey,
10:28
look, you got the scaling model, and here's how you know
10:30
what the future is like. Okay, great. You should
10:32
be doing that. That's what expertise should be doing. How
10:34
does that translate to people who
10:36
are, like, living their lives and doing this? They
10:39
care, but they've got their lives to do.
10:41
And I think that kind
10:44
of being grounded
10:47
in how to think about
10:49
it within a political leadership circumstance, and not get
10:51
flustered by the fact that part of when it
10:54
happens, when you're president, is, for example, one of
10:56
the things I find entertaining is these people are
10:58
like, Biden is a socialist. And you know, like,
11:00
he's been around for 50 years. He's never been
11:03
a socialist, ever. Right? And
11:05
so, just going, look, I get it. That's
11:07
the nature of it. I'm not going to
11:09
lose focus on what I care about is
11:12
what happens to the most
11:15
of American people, the middle class and all that. And
11:17
that's the most important thing to think about first, and how
11:19
to talk about it, and how to think about it, how to
11:22
engage. And I thought that was very helpful. Well,
11:25
what's heartening hearing all of that too is,
11:27
yeah, you may be excited about this stuff,
11:29
you folks, in Silicon Valley, but like, how's
11:32
it going to change the way my kid gets taught at school? Yes.
11:35
You know, those are the things that people worry about.
11:37
Yes, exactly. And by the way, with change, it'll always
11:39
have some disruption. There'll always be
11:41
some uncertainty on it. It'll be like, well, what
11:43
should be the policy? Could I write essays with
11:45
it? Does it show your work? Is
11:47
the teacher going to respond positively or negatively? I
11:50
mean, the things that we can accomplish in
11:53
elevating humanity are right here right now. Let's
11:55
really make sure we get to them. And I
11:57
think that's part of the joy of it.
12:00
change things, but that's a change
12:02
that on the other side will be better. Because
12:04
by the way, you go to your average
12:06
American and say, this one's better for your kids, this one's better
12:08
for your family. It'll just be
12:10
some transition and getting there. And
12:13
then people tend to be the, Hey, I'm okay.
12:15
I'm game. I
12:26
want to thank Reed for sharing his
12:28
exclusive experience with the president. Undoubtedly
12:31
there are going to be things happening with AI in
12:33
2024 that no one can
12:35
predict, but it's reassuring
12:37
to know that there are concerted
12:39
efforts at the highest levels for
12:42
an adaptive, realistic and people
12:45
centered approach to oversight. I'm
12:48
Bob Safian. Thanks for listening. Masters
12:53
of scale is a weight. What original are
12:56
executive producer is Chris McLeod. Our
12:59
producers are Chris. Adam
13:01
skews, Alex Morris, Kaka
13:03
Lagerzky and Masha Makutunina.
13:06
Our editor at large is Bob Safian. Our
13:10
music director is Ryan holiday, original
13:13
music and sound design by Eduardo
13:15
Rivera, Ryan holiday, Hayes holiday and
13:17
Nate Casela. Audio
13:19
editing by Keith J. Nelson, Steven
13:21
Davies, Steven Wells, and Andrew Knott.
13:24
Mixing and mastering by Aaron Bastinelli
13:26
and Ryan Pugh. Our
13:29
CEO and chairman of the board is Jeff Berman. Masters
13:32
of scale was created by June Cohen
13:34
and Darren Trist. Special
13:37
thanks to Jodine Dorsay,
13:39
Alfonso Bravo, Kim Cronin,
13:41
Erica Flynn, Sarah Tarter,
13:43
Katie Blasey, Meryl Carrecker,
13:45
Xunimé Essequena, Colin Howard,
13:48
Brendan Klein, Semyo Puta, Kelsey
13:50
Cezon, Louisa Valen, Nikki
13:53
Williams, and Justin Winslow. Visit
13:55
masters of scale.com to find the transcript
13:58
for this episode and to subscribe. or
14:00
email newsletter. I
14:11
produce a podcast called Masters of Scale,
14:14
and I'm looking for a way to get
14:17
listeners to leave us a five-star
14:19
review on Apple Podcasts. Do
14:22
you have any suggestions? Absolutely.
14:25
One effective way is to engage with
14:27
your listeners directly during your episodes. Remind
14:31
them how much a five-star review can support your
14:33
show. What about to
14:35
try to convince listeners to share
14:37
one of their favorite episodes with
14:40
somebody in their life? Do
14:42
you have any suggestions for that? Certainly.
14:45
People love to share content that has
14:48
made an impact on them, so remind
14:50
your listeners how sharing an episode can
14:52
start meaningful conversations with friends or help
14:54
others discover insights they've benefited from. Would
14:57
it be okay if I used maybe our
15:00
conversation as a segment on the show? That
15:02
sounds like a creative approach. Using
15:05
a meta-conversation about engaging
15:07
your audience as a segment could
15:09
be quite engaging. This
15:12
is executive producer Chris McLeod, and
15:15
I urge you to take ChatCBT's advice
15:17
and leave us a review on Apple
15:19
Podcasts and share an episode with a
15:21
friend. It really, really
15:24
helps.
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More