Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:00
a lot of smart people are working on something that's like virtually
0:02
guaranteed that it's gonna happen, it's just
0:04
a question of when.
0:06
Hello, everyone. The A6CZ
0:09
podcast is back, and this is
0:11
your new host, Seth Smith. I'm so
0:13
excited to explore the world of technology with
0:15
you through the lens of the builders shaping it.
0:17
And if you've been listening to the pod for a while,
0:19
I hope you'll stick with me as I take the reins from
0:22
Sonol who by the way is doing really wonderful
0:24
work at our sister podcast web three with a
0:26
sixteen c. Okay. Onto the content.
0:28
Today, we have a very special kickoff episode
0:30
to start our launch series where we
0:33
wanted to explore the age old question. why
0:35
is building the next generation of technologies still
0:38
so important? And perhaps even more important,
0:40
who is gonna build this next generation of technology?
0:43
and what needs to be done to enable those
0:45
founders and builders. And I'm a little
0:47
biased, but who better teacher versus ground than
0:49
a sixteen Z's co founder and general partner,
0:51
Mark Andresen. one who has not only
0:53
built, but also invested in the future time
0:55
and time again, especially when it was not
0:57
the obvious thing to do. So today, together
0:59
with Mark, we explore technology through the lens
1:01
of his history, including the three stages
1:03
of human psychology as we encounter these new
1:05
technologies. And we also talk about how that
1:07
process often ends in regulation, and we include
1:09
a couple examples of that which by the
1:12
way, if you've never heard of red flag laws,
1:14
you'll want to listen in. We also talk about
1:16
when to change your mind. The Cambrian explosion
1:18
of opportunity coming from distributed work the
1:20
importance of founder led companies and so
1:23
much more. And of course, we'll end the
1:25
episode by examining why there's still
1:27
so much reason for optimism. And hopefully,
1:29
this episode will also get you excited about what's to
1:31
come with the a sixty c podcast as we do have
1:33
a lot more coming. That includes coverage
1:35
of major trends like AI, space, carbon
1:37
removal, and you'll soon hear from legends like
1:40
Neil Stephenson, Ballet Chase from your boss and
1:42
even Steve Wozniak, who shared with us as the latest
1:44
venture, Privateer. By the way, that
1:46
episode is already live in the feed if you want
1:48
to give it a listen. Alright. Let's
1:50
dive in. The content
1:52
here is for informational purposes only
1:54
should not be taken as legal business tax
1:56
or investment advice or be used to evaluate
1:59
any investment
1:59
or security and
2:00
is not directed at any investors or
2:02
potential investors in any ASIC sixteen
2:05
z fund.
2:05
For more details, BCA sixteen
2:07
z dot
2:07
com slash disclosures.
2:19
So this is the first
2:20
episode of our launch series. And in the launch
2:22
series, we're going to cover several technologies and
2:24
we're going to cover them with their founders and dive
2:26
really deeply into how these technologies may
2:29
shape our collective future.
2:30
But before we do that, we're going to address an
2:32
important question. And it may sound simple,
2:35
but we need to ask why these
2:37
technologies even matter in the first place.
2:39
Why is it so essential today that we continue
2:41
to build? And today, we have Mark and Jason
2:43
talking to us about this very important topic
2:46
and who better than someone who has built
2:48
the future and invested in the future many
2:50
times especially when it wasn't the
2:52
easy thing to do. So, Mark, why
2:54
don't we start there? And why don't we start by
2:56
attacking this question with a lens
2:58
of
2:58
history.
2:59
So how may be do people view
3:01
the importance of technology today? And
3:04
how has that changed maybe relative to
3:06
how they viewed it in the past? Yeah.
3:07
So thanks. That's a great way to start. And, you
3:10
know, so many of the discussions about today's
3:12
technologies take place in this, you know, kind of a
3:14
historical frame where it's as if there had never
3:16
been any new technologies in the past. Or
3:18
that there were new technologies in the past, but those
3:20
were all the good ones. And today, we just have the bad ones
3:22
or something like that. And so the key question
3:24
to kinda start out asking on all these topics is
3:27
basically, like, why is there something and
3:29
not just nothing? By which I mean, basically,
3:31
like, what was life like before technology?
3:34
Right? Like, what was lifelike in, like, its national
3:36
state before we had all these wonderful technologies, before
3:38
we had steam power, before we had contractors, before
3:40
we had telephones and so forth. And And,
3:42
you know, there's, you know, a lot of historians have talked about
3:44
this, but, you know, the answer is life was
3:46
what was known at the time that, you know, Thomas
3:49
Hobbs famously said it was nasty British
3:51
in short. Right? And so you'll read
3:53
these fantasies from time to time about how people
3:55
kind of been, you know, older periods or historical
3:57
periods, like somehow they were living in the state embrace
3:59
of nature. and kind of everything was wonderful. And
4:01
they were just kinda hanging out and having a good time. And and basically,
4:03
like, that those are all just fantasies of
4:06
reality. But basically, everybody was miserable
4:08
all the time. Everybody was poor. everybody
4:10
was assistance farming just to make enough, you
4:12
know, be able to harvest enough food, be able to eat
4:14
that day, or to be able to hunt enough meat, to be able to
4:16
eat that day. And and for the most part, people died
4:18
young, they sick. They never got anywhere. They never did
4:20
anything. Basically, like, all of
4:22
recorded civilizations, basically,
4:24
only over the course of last four thousand years.
4:26
for the for the many millions of years of humanity
4:29
before that, like, there's basically nothing. And,
4:31
you know, basically, because, like, people had no time to
4:33
do anything other than just try to grow up with
4:35
Okay. So basically, it's a what happened to cause
4:37
life no longer being astro British in short.
4:39
You know, what happened to cause basically
4:41
the reality of human existence to go from what
4:43
it used to be to what it is today. you know,
4:45
it it's not perfect today, but it's a lot better than it
4:47
was. And of course, the answer is technology.
4:49
And in fact, LG is the only answer to that
4:51
question. Right? There's no other answer. Right?
4:53
It's not like people got smarter. It's not like
4:55
people got, you know, I don't know, like, this week.
4:57
You you you you know, DNA is the same. You have
4:59
things haven't changed. It's basically only
5:01
through technology and things have gotten better. And the
5:03
way to kind of think about that is what is technology.
5:06
Technology is tools. Right? Technology basically
5:08
has applied human ingenuity. in
5:10
building tools, and then and then those tools
5:12
basically give human beings leverage on
5:14
the world. That leverage shows up in
5:16
many forms. it shows up in
5:18
forms, you know, for sure, in some cases that are
5:20
bad, it also shows up first in the form
5:22
many cases that are that are also are are very
5:24
good.
5:25
So I
5:26
agree with you that progress continues.
5:28
And if you look at all of the important markers, we
5:31
are doing better, and that has been true
5:33
for a very long time. Why is
5:35
it then that people seem to
5:37
have this view of
5:39
where we are today in history in
5:41
terms of things being so bad? Yeah. So,
5:42
you know, look at obviously, that's a complicated
5:45
question. The technology part of it, I
5:47
think, is very clear, though, which is, you know and
5:49
and again, you've gotta think about this historically.
5:51
So So let's so let's start with
5:53
the story of, like, one of the very original technologies
5:55
that basically him had been kinda discovered
5:57
and mastered, which was fire. Quite
6:00
literally fire. Obviously, today,
6:02
you take fire for granted, you just assume it's something that's
6:04
always been with us. But, like, fire is a
6:06
tool. Fire is a technology with something that
6:08
at some point kind of early human beings had
6:10
to figure out how to master and how to take control
6:12
of. And just to give you a sense of what a
6:14
big deal it was to actually master
6:16
the technology of fire when that happened,
6:18
but the sort of core fundamental kind of myths
6:20
of human existence is is the the myth of Prometheus,
6:22
the god for Prometheus. And the
6:24
god for Prometheus famously is the know, and the
6:26
way the myth goes, you know, Prometheus delivered
6:28
fire to humanity and basically delivered fire as a
6:30
tool. And Prometheus was not
6:32
is not was not beloved as a consequence
6:34
of of delivering fire to people. But
6:36
for from atheists who was condemned. Specifically,
6:39
he was in the Math, he was chained to
6:41
a rock for and for all eternity,
6:43
And according to the math every day, a
6:45
bird comes along and checks out his liver
6:48
very painfully. And then he according
6:50
to the math, he regenerates every night I'd say the bird
6:52
tortures them again. You know, this has been
6:54
happening forever. So so, like, the
6:56
fact that there's that myth with that outcome
6:58
assigned to the introduction of fire as a
7:00
technology to human existence. Right. And and and
7:02
and of course, what assumes the mess is not literally
7:04
true, but it's symbolic. Right? Which is basically
7:06
symbolic of the fact that humanity basically experienced
7:09
the rise of fire technology and said, like,
7:11
wow. This was a big deal
7:13
and maybe not entirely positive. And
7:15
then, of course, that makes sense because if you think about
7:17
how fire was used as a technology. Obviously,
7:19
fire was used for good. Right? Which is like fire
7:21
made it possible to cook meat. Right?
7:23
Fire made it possible to, you know, to basically
7:25
we'll just find the campsite again. like wolves in
7:27
the middle of the night. It made it possible to keep people warm.
7:29
You could keep your baby from freezing to death. Fire
7:31
is just like miracle technology. Of
7:33
course, fire also got used in warfare.
7:35
Right? you know, even up to the modern day, like,
7:37
what's the kind of state of the art weapon of our times?
7:39
It's the nuclear bomb. Of course, what what does that
7:41
bring? It brings fire. And so, you
7:43
know, look, it is a double edged sword. Like,
7:45
would we wanna live in the modern
7:48
world without fire? Because it has, you
7:50
know, both the outsized as well, the the outsized
7:52
we would. But, like, Both of those are
7:54
present. And and, basically, my interpretation is
7:56
if you kind of go through the history of new technologies, basically,
7:58
for every important new technology, there's
8:00
always this argument basically, there's always this argument
8:02
of, like, okay, in theory, there are all these things we can
8:04
do with the new technology. They're very positive.
8:06
In fact, probably, there are things the new
8:08
technology will make possible that we can't even think
8:10
of today. Right? And that and that's been a very common pattern,
8:12
by the way, which is people don't people actually
8:14
have a very hard time anticipating the upside for their
8:16
technologies. But the same argument basically
8:18
always applies us with fire, which is basically like, well,
8:20
what about potential downsides. And then, you know,
8:23
say, human beings are kinda psychologically wired
8:25
to the downside. Like, we're wired to basically
8:27
detect and try to evade threats. And
8:29
so there's basically always the simple to
8:31
say, okay. You know, this is the technology that's gonna
8:33
ruin everything. And if you go through basically a
8:35
history of every new technology, like that, you know,
8:37
that argument of library repeatedly. Give me
8:39
a couple of examples. So one is outdoor
8:41
lighting. Electric lighting actually was first implemented
8:43
outdoors because they didn't know how to do it
8:45
indoors, because get bring places
8:47
down. It took a while to get get it to get it to
8:49
be safe. But the first electric lighting was rolled
8:51
out in, you know, big cities in Paris and London. forget
8:53
what it was, like something like two hundred years ago now.
8:55
And, you know, big deal, like, outdoor
8:58
lighting, all of a sudden, the city is walkable and
9:00
livable and safer right for dark. This is a pretty
9:02
big deal. And the contemporary accounts are
9:04
basically I mean, one was a sense of wonderment
9:06
that this was now, you know, basically, the life can
9:08
now be basically, wait twenty four
9:10
hours. But then there were also all of these kind of
9:12
stories of, like, this is gonna, like, completely destroy
9:14
civilization. This is against the natural order of things.
9:16
You know, this is gonna completely ruin our people
9:18
aren't gonna be able to sleep anymore, you know, all
9:20
kinds things, all kinds of illicit, you know, social activities
9:22
are gonna happen at night. Right? Yeah. In
9:24
the past, everybody just had to go to bed at night. Now
9:26
they're gonna be out of the streets at three in the morning, doing all
9:28
kinds of bad things. by the
9:30
way, there are in every major
9:32
city, they're coming all out of the street at three in the morning
9:34
doing bad things. Like, it's not like the downside
9:36
was was was completely wrong.
9:38
that obviously that the trade off is worth it. My
9:40
favorite story on this is the invention of the bicycle.
9:42
And this one is really is really great because,
9:44
you know, there's this Netflix documentary called
9:46
the pilemma that's kind of, you know, condemned social media as
9:48
this kind of unique threat to civilization. And
9:50
one of the things that one of the guys in in that movie
9:52
says is he's like, you know, social media is the
9:54
first technology like this that's had these
9:56
negative consequences. He's like nobody
9:58
nobody ever complained about the invention of the
10:00
bicycle. And it actually turns out that's not
10:02
true. Like, he just he
10:04
just didn't do enough reading. Like, it turns out people actually
10:06
complained a lot about the invention of the bicycle.
10:08
And this sort of, you know, it's sort of very
10:10
symbolic. So that So the bicycle is a
10:12
big deal. So bicycle rolled out as a consumer
10:14
product about a hundred and fifty years ago, and they kind
10:16
of got it working in manufacture to the point where we got
10:18
cheap enough for kind of normal people can buy
10:20
it. And so the the bicycle starts to roll out kind
10:22
of across the American countryside. And
10:24
and basically, there's this immediate moral
10:26
panic about the bicycle, and this is chronicles in the
10:28
media at the time. If you read the Twitter account,
10:30
pessimists spark. It's which
10:32
maybe we can link to he goes back and reconstructs
10:35
kind of the the moral panic around the bicycle. And
10:37
basically, the argument against the bicycle at the time
10:39
was the bicycle is the first
10:41
transportation tech acknowledge that young
10:43
unmarried women in towns and villages
10:45
across the US could actually afford to be able
10:47
to buy a use. Right? because cars
10:49
didn't exist yet, like, you know, whatever person in wagons are
10:51
too expensive. They couldn't get access to them.
10:53
Whatever. Walking took too long. All of a sudden, like,
10:55
I'm young, I'm married with the bicycle, which means young
10:57
and married women can go from, like, one
10:59
town into the next town, which
11:01
means all of a sudden, young and married women
11:03
can meet you know, boys and men,
11:05
not just in the current time, but any other
11:07
time. And so obviously, to the kind of
11:09
established social order of the time, like this
11:11
would say, profound threat. right, to to kinda
11:13
how things work. The assumption was you're young, a very
11:15
woman in town. You bury one of the men.
11:17
Right? In that town, all of a sudden, your world
11:19
opens up. And so The media at the
11:21
time created completely,
11:23
like, got a whole lot. They they created a new medical
11:25
condition called bicycle face. And
11:28
the idea of bicycle face was you're a young woman,
11:30
you're on the bicycle, you're pedaling to get to the next
11:32
towel, and, you know, pedaling takes exertion. And
11:34
so you're gonna have, like, your face is gonna
11:36
contort, you know, because of the exertion and you're gonna be, like, you
11:38
know, pedaling along like this. And
11:40
bicycle face is the idea that your face is going to
11:42
freeze into that contorted position. And
11:44
then, you know, then then you would never be able
11:46
to get married. Right? Now, fortunately,
11:49
it turned out by a simple face was not actually
11:51
a real thing. it turned out civilization survived
11:53
the infection of the bicycle. But basically,
11:55
like, there's just this, there's this constant
11:57
blowback. Right? And then and then
11:59
basically what you
11:59
find and we'll probably talk about this in some way. basically, what
12:02
you find is the flowback is nominally
12:04
a
12:04
response to the dangers of the technology, but
12:07
the flowback actually is
12:09
in in in almost every case. blowback
12:11
actually is it's basically a
12:13
fear, a statement, an assertion, a
12:15
realization that the introduction of the
12:17
new technology is going to change
12:19
the society. Right? And and and then
12:21
in particular, status and power
12:23
within the society. Right? Who's in
12:25
charge? Who's in power? Who makes
12:27
decisions? Who has status? Who
12:29
gets money? Right? All of a sudden, the ordering of
12:31
society is set for grabs. And that and that's why you get
12:33
this just like, you know, spectacular freak out when
12:35
things show up. Yeah. I'm
12:36
glad you brought up pessimists archive because if people
12:38
are curious, we'll bring it up on the screen. But
12:41
basically, anything that you can think of as a
12:43
prior technology or maybe even things
12:45
that you for granted as not technologies, just
12:47
things embedded in our lives. Had this blowback,
12:49
you mentioned the bicycle, but it's also the
12:51
radio trains. Teddy bears jazz
12:53
music. I know people would view Teddy bears as
12:55
technology, but the point is anything
12:57
that becomes embedded in society
12:59
that as you're alluding to
13:01
can impact society at
13:03
scale. people start to get afraid of
13:05
because they start going through all of these
13:07
scenarios and thinking, oh, who is this gonna impact? Is this
13:09
gonna impact my job? Is this gonna impact the types of
13:11
people I can interact with? it's going to
13:13
impact the power I have in society. So
13:15
Mark, why don't we dive into that? How do
13:17
you see technology and its implementation?
13:19
Or really as it's starting to be
13:21
implemented within society? How do
13:23
people react? And what specifically
13:25
about the power within society is
13:27
being upended? Howard Bauchner: Yeah,
13:28
so there's this incredible book that's it's
13:31
very short and very good. It's actually written
13:33
fifty years ago by a professor at
13:35
MIT at the time. And it's
13:37
the name Elton Morrison. And and it was great about
13:39
this book as it was before the Internet, even before
13:41
personal computers. And so it's got this kind of
13:43
timeless kinda quality to it. And it's it's the
13:45
title of the book is called man machines and
13:47
modern times. It's this topic. It's basically okay.
13:49
What what exactly is the process by which you do
13:51
technology, enter society, and then how does society
13:53
react? Basically, how does the powers that be or the
13:55
status quo of society react? And Morrison
13:57
tells this amazing story. He kinda hangs the whole thing on
13:59
this sort of very amazing story from about a
14:01
hundred hundred twenty years ago now, maybe a
14:03
hundred thirty years ago. around this guy
14:05
named Sims, and this guy named Sims
14:07
at the time working area of naval
14:09
warfare. Right? So, you know, big, big, big, you know,
14:11
battleships, seafaring battleships. you
14:13
know, firing on each other, firing on on
14:15
land targets and so forth. And of course, you know, the
14:17
the world, a hundred and twenty, hundred years hundred and
14:19
thirty years ago, we didn't airplanes, you know, sort
14:21
of military airplanes he added. So see
14:23
warfare, naval warfare was warfare, like, it
14:25
was how invasions happened and it was, you know, and
14:27
and how countries got defended. So this was kind
14:29
of the core aspect of military technology
14:32
at the time. And so since since basically
14:34
worked on basically guns. Like big big guns
14:36
on on big military ships. And then in
14:38
particular, how do you aim big guns on big
14:40
military ships? And the the way the
14:42
story goes is basically before Sims guns
14:45
were fixed in position on a ship. Right? So
14:47
you'd you'd have a big ship. You'd have, you know, some you'll
14:49
see this in movies. You'd have the guns sticking out
14:51
the side of the ship. And then basically, what would
14:53
happen is, and the gun is in like a fixed position.
14:55
Right? because it's, you know,ashing the deck, so it doesn't roll
14:57
around. And so what happens is, you know, the c
14:59
is moving. Right? And so the the
15:01
ship is kinda going back forth like this in the
15:03
water, which means the gun is
15:05
going back and forth like this, which
15:07
means that basically enable battles up
15:09
until Sims, you know, that's accuracy
15:11
rate of of guns being fired off
15:13
these ships was, like, at best ten percent
15:15
and, you know, maybe, you know, quite often below that. And
15:17
they would just miss most of the time and why they
15:19
miss most of the time. the gunner would be
15:21
you know, the gunnery officer would get a fix on
15:23
on the position that he would go to order to fire.
15:25
But by that point, right, the ocean had moved, the
15:27
ship had moved, and then he no longer had a lock, and then the
15:29
thing would fire on the the kind of ball against the other
15:32
ship. And so, you know, a lot of naval battles up
15:34
until that point where the ship's kinda sitting, he's head
15:36
besides, sorry, each other, missing all the time.
15:38
Right? And so Tim said, well, you know,
15:40
g, like, there has to be a better way to do this. And
15:42
basically, what he designed was an mechanical
15:44
mechanism that automatically
15:46
basically work in opposition, right,
15:48
counterbalance to the role of the ship. And
15:50
so if if the if the ship is, you know, if the ship
15:52
basically is going down, Right? The
15:54
the the mechanism for the gun would automatically
15:56
correct. And so the gunnery officer could
15:58
basically get it fixed. And then as long as the
16:00
ship stayed in the same position as it rolled
16:02
around, know, the sort of lock in the target would
16:04
stay intact, then all of a sudden the accuracy rate shoots
16:06
up to, like, ninety percent. Right? And so
16:08
and so what what's so great about this
16:10
example is it's like the most obvious thing in
16:12
the world, which is it's like obviously every country
16:14
in the world would instantly adopt this, every military
16:16
would adopt this, every ship would instantly be
16:18
retrofitted to do this. Right? This is, like, the most
16:20
obvious slam dunk thing you could do. It it's a
16:22
huge advance of warfare. You couldn't imagine
16:24
living without it. Like, that's what you would assume. And
16:26
and, of course, that's not what happened at all.
16:28
what happened basically was since took it took since
16:30
like a full twenty five years to basically
16:32
convince the American and British Navy to adopt this
16:34
technology. It was like a full generational
16:36
shift And and the book goes through kind of in
16:38
detail, basically, what happened? He Sturgo
16:40
Sims ultimately had actually appealed directly
16:42
to president Teddy Roosevelt at the time. like,
16:44
the entire military command structure of that era, basically,
16:47
just told told them that, basically, that's
16:49
off. You know, we don't want your
16:51
new thing. know, we're we're gonna keep doing
16:53
things the old way. And he ultimately appealed to say,
16:55
Roosevelt, and Roosevelt actually ordered the Navy to look at
16:57
it, and I mean, they, you know, they later ended up adapting
16:59
it. And so Morrison basically says, well,
17:01
okay. Like, it's like the ultimate example. Like, if they
17:03
wouldn't adopt even that technology, like,
17:05
okay, this this must be like some very
17:07
primal kind of power force. that's
17:09
happening. And his thesis was, it basically
17:11
is the following, which is every new technology
17:13
as a reordering of of of the
17:15
power and the status in society.
17:17
And then specifically in the form of this
17:19
gun. Right? It was basically like the
17:21
entire
17:21
training,
17:22
basically methods, all of the officer
17:25
promotion methods, like the entire social hierarchy
17:27
of naval vessels and how gunnery officers
17:29
were trained and how guns were,
17:31
you know, built and manufactured and crude
17:34
and managed and how military doctrine
17:36
worked. all the different things about how you
17:38
make decisions, you know, about this. Like, it was
17:40
all based on the old model. And basically, like,
17:42
that skill set became obsolete when this
17:44
gun came out. so those people became obsolete at least they
17:46
worried they become obsolete. This new
17:49
breed of kind of, you know, more advanced innovative
17:51
engineering, you know, kind of mentality
17:53
man and was a profound threat. And by the way, that's what happened,
17:55
which is he ultimately had a generational turnover of
17:57
all of the staff and and officers involved in
17:59
the naval gunnery. And so
18:02
Morrison basically derives he goes through this example, and then
18:04
he derives basically this three part process that he
18:06
says applies to any new technology.
18:08
Basically, as it is
18:11
as it is greeted and fought by
18:13
the status quo by the powers it'd be. And he says,
18:15
basically, it's a three step process. So
18:17
step one is just to completely
18:19
ignore it. Like, so just pretend it doesn't exist,
18:21
refuse to acknowledge it, don't talk about it,
18:23
don't even engage in conversations like, we we're
18:25
just not gonna do this. At
18:27
some point though, at this, you know, at some point, these
18:29
things become too obvious and they have to engage. He
18:31
said step two is
18:32
rational counterargument. Right?
18:34
and
18:34
rational kind of argument is this can't possibly work
18:36
because, you know, it's gonna be too expensive. It's not
18:38
gonna be fast enough. It's not gonna scale. People don't
18:40
know how to use it. Right?
18:42
All all the different kind of rational arguments
18:45
that you can come up with to to to oppose
18:47
something. And then ultimately, when those don't work
18:49
anymore, because you know, people are still watching this
18:51
and being like, okay, this still seems like a good
18:53
idea. Then he says, stage three. He says, stage three
18:55
is when the main calling begins. Right?
18:57
And so stage three is basically just like
18:59
a full out power status political
19:02
fight where all of a sudden, basically, it's like,
19:04
okay, these are, you know, these people who are bringing through
19:06
technology. They're bad people. They have bad morals.
19:08
They have bad intent. They're gonna ruin everything.
19:10
Right? And if and if you think about it, it's so funny
19:12
because it's like, basically, the Internet followed this
19:14
exact trajectory. like, you know,
19:16
crypto cryptocurrency blockchain, web three
19:18
is following this exact same trajectory. Social
19:20
networking followed this exact same
19:22
trajectory. Like, I I've now seen this pattern, you know, out of fifty
19:24
times in the last thirty years, and
19:26
it keeps playing out the same way.
19:29
Nobody learns anything. Right.
19:31
And it's it's this is literally what happens with every new
19:34
technology, and I I I'd become convinced that,
19:36
basically, I I was involved. Yeah.
19:37
I think something
19:38
that reminded me of is something
19:40
Balaji said recently, which he was reflecting on people's response
19:43
to social and how it's impacted
19:45
society over time. And he
19:47
He noted that maybe ten or twenty years ago social was
19:49
seen as like these silly apps being created
19:51
and like who wants to focus on this? And, you
19:53
know, why are we directing smart attention into
19:55
these types of fields? And
19:57
then now he's saying or
19:59
noting that people are framing it as a threat to
20:01
democracy. So it's kind of interesting that people
20:03
go from one angle or
20:06
one perception And then because technology particular
20:08
tends to be this force, it's very, very
20:10
hard to stop. It then turns
20:12
into something much different. And the same people
20:14
who framed it as a silly app
20:16
are now framing it as a threat to democracy. And to
20:19
your point, this isn't new. Right? We had the
20:21
same kind of dynamics happening with
20:23
the bicycle. or actually I think it would
20:25
actually be fun for us to go into these red
20:27
flag laws that you told me about quickly, and
20:29
we'll return to to the framing that that
20:31
you brought up. But could you speak a little bit
20:33
more to the red flag laws that
20:35
were implemented when cars were coming to
20:37
be? Yeah. And I'd like to I'd like
20:38
to come back to Balaji's thing on social, by the way,
20:40
because I'll interpret it a little bit I'll
20:42
put it the the sense framework. But but, yeah, so
20:44
cars. So this is another great one. So, like, cars.
20:46
It's like, okay. Like, we all live
20:48
with cars. We we all can't live without cars. Like, you
20:50
know, there's still huge fights about
20:52
how, you know, cars should be used in our society, but, like, there's cars everywhere.
20:55
Like, our society doesn't function about cars. We
20:57
just kinda take them totally for
20:59
granted. By the way, we granted that we just, like, repeatedly bail
21:01
out the big card companies. Right?
21:03
Like, at this at this point, like, the taxpayers have kept
21:05
them in business for a long time.
21:08
And so, you know, the car and and so then you're you're just
21:10
like, okay. The car must have been this obvious thing. Like,
21:12
of course, you want the car and who could have bought the car
21:14
at least for any kind of valid reason. And so again, to
21:16
your point, like, turns out, actually, the cars
21:19
were were actually a a profound
21:21
threat to the sort of social order
21:23
of that time of the era. This is
21:25
like going back a hundred hundred and twenty years. it was
21:27
basically this exact same kind of process played out with
21:29
the car. And so the the thing you
21:31
mentioned basically, the thing that happened at the peak of
21:33
kind of the anti carcass area at
21:35
the the moral panic around cars was basically basically,
21:37
what happened was cars were a threat to basically they
21:39
were the threat to, like, the ordering of, like,
21:42
from how cities were laid out. They were gonna upend, you know,
21:44
the the ability to have, like, modern transportation,
21:46
modern shipping. They were gonna upend everything from
21:48
the world of local merchants.
21:50
they were gonna upend, you know, there was an entire industry of black snips.
21:52
You know, the horse was like central to a lot
21:54
of economies. A lot of people made their living off of dealing
21:57
with horses. there were people who were, like,
21:59
trained, carriage drivers, all of a sudden, ran a job.
22:01
And so there there was this, like, all of a sudden, this this
22:03
huge backlash. And so what happened was,
22:05
a bunch of sort of state municipal
22:07
level areas both in the U. S.
22:10
like in around Pennsylvania at the time and then
22:12
also in the UK implemented
22:14
their their legislators implemented at the time what what became
22:16
known as the red flag laws. So
22:18
the red flag law looks works
22:21
as follows. which is, okay,
22:23
mister Carr owner, you've got your fancy
22:25
new car. Congratulations, you know. You're
22:27
you're very proud of yourself. You're probably a a
22:29
pretty, you know, well off person. in the
22:31
community, people probably generally are probably just to
22:33
start with. You've got this fancy new
22:35
automobile. And and by the way, in those days, like cars
22:37
broke down all the time. And so when you would take your
22:39
car up for a ride, you'd have, like, you'd be driving the car, and then
22:41
you'd you'd you'd ultimately get pretty a mechanic
22:43
with you. Right? It
22:45
basically fixed the car when it broke. I mean, we're
22:47
still getting everything to work. And so and
22:49
your mechanic or whatever your family, you'd be you'd be out, you know,
22:51
motoring along in your car and whatever dirt
22:53
road. And the law was that you
22:55
had to employ another guy. You had to
22:57
employ a guy. to walk, you know,
23:00
fifty feet in front of the car carrying
23:02
big red flags.
23:05
Right? Okay. So picture this. You're
23:07
driving along. You're out for a nice drive, get your kids, whatever, and get your
23:09
foot mechanics, and you're you're you're you're going along.
23:11
You know, cars in those days didn't go very fast,
23:13
but they did go faster than you could walk. And so
23:15
you're driving along at whatever ten, twenty miles an
23:17
hour, but According to law, you have to
23:19
have a guy in front of you on
23:21
foot, like, out in advance. And he's got
23:23
these, like, big red red red flags, and you have
23:25
to follow this guy because he has to stay in
23:27
front of And so you can only motor along at
23:29
whatever the three, four miles an hour, you know, that
23:31
this guy can walk. And this guy is, like, waving the red
23:33
flags. Why is this guy waving the red flags to
23:35
warn everybody that a car is coming? Right?
23:38
Why why was the explanation that he used to
23:40
warn people that a car is coming? Well,
23:42
because the car might steer the
23:44
horses. Right? So, like, you know,
23:46
if the car comes along, it's making noise and
23:48
scares the horses, you know, the horses that are most of
23:50
us on the road at that time, they freak out or, you
23:52
know, by bystanders freak out, people get hurt.
23:54
this would be really bad. And so so literally,
23:56
it's like, okay. That that was how the card got rolled out.
23:58
The the most advanced form of
23:59
this law that I've been able to find when
24:02
it steps further. So basically, if
24:05
you're driving along and you actually see a
24:07
horse coming at you, you see somebody in a horse
24:09
coming at you in the direction. You
24:11
have to pull over to the side of the road. You
24:13
have to disassemble the car. You have
24:15
to take
24:15
a part. Right? You and
24:18
your
24:18
mechanic, take the car apart and you have to hide
24:20
the parts of the car so that the horse can't
24:22
see them. Right? because the horse
24:23
might get scared. Right? Get scared by the
24:26
appearance of the car. And then when the horse goes by, you
24:28
can then reassemble your car. Right? And
24:30
and keep going. Right? And of course, you look back
24:32
today and you're just like, okay. This is like incredibly
24:34
comical. Like, how could they ever do then, of
24:36
course, you'd exactly your point of my social network. And you think
24:38
of exactly the technologists. And then you're like, oh, yeah.
24:40
You know, they're putting in place laws that, you know, a
24:42
hundred years from now, you know, the laws that are being put
24:44
in place now on a lot of modern l g top
24:46
I sort of want just a silly thread, flight loss. But since
24:48
nobody ever learns anything, you know,
24:50
it's pretty well. Yes, sir. We'll repeat.
24:52
me just hit real quick with the biology point on social networking.
24:54
So this is a really interesting one because he described
24:57
the overall art or you described him as saying
24:59
the overall art was from harmless. You know, who cares what your cat
25:01
had for breakfast? Kind of these are so trivial things
25:03
too. It's like the fundamental threat to
25:05
democracy. It was actually
25:06
a three step process. Right? It was it was step
25:08
one was ignore. Right? Step two in
25:10
the case of social networking. Actually, we went through the whole
25:13
argument. But there was this other step, maybe call it step
25:15
two and a half, which basically is
25:16
like, this is the best thing ever for democracy.
25:19
Right? And you you may remember around about ten
25:21
years ago, around twenty twelve, basically, two things
25:24
happened. One was Obama got reelected and at
25:26
the time this is, like, referred to us, like, the first
25:28
Facebook election. it was when Facebook kinda hit on
25:30
really mainstream. And so, like, they're amazing
25:32
covers newspaper stories talking about
25:34
literally the story was Facebook saved
25:36
democracy. Right? And was Facebook saved democracy, big data,
25:38
you know, saved democracy, all of a sudden, like,
25:40
you know, politicians gave up their messages, their
25:43
voters. you know, the correct candidates can
25:45
get elected. You know, this is, like, the most wonderful thing for
25:47
democracy ever. By the way, that was
25:49
also during the Arab spring. Right? And and the
25:51
social networking, of course, got a lot of credit. for the
25:53
time the Democratic revolutions of the, you know, of
25:55
the of the of the of the spring. And so there was this, like,
25:57
overwhelming sensation. Like,
25:59
when social networking
25:59
executives, you know, in ten years ago, would go to tender
26:02
Davos or Aspen or any of these places with fancy
26:04
people or with important titles. Like, they just hit
26:06
lavage with praise about how wonderful this new
26:08
technology is for democracy. to
26:09
Balaji's point ten years later in
26:11
the story, has done a complete one eighty. Right?
26:13
Now it's the absolute worst fundamental possible
26:16
threat you could have to proceed
26:18
because, of course, it turns out. Not just
26:20
one side from win elections, it turns out the other side
26:22
can too. And it turns out the other side
26:24
also uses social networking and runs ads social
26:26
network thing. And and and to your point, basically, what's
26:28
happened is if you track
26:29
the people involved, the exact same people
26:31
have held every single position.
26:33
Right? So the exact same
26:36
experts, professors, you know,
26:38
pundits, commentators, analysts,
26:40
think tank people, magazine
26:43
publishers, political
26:43
activists. The exact same people have held every
26:45
single one of these viewpoints all the way through
26:47
this, with no
26:48
attempt at any point to reconcile
26:51
their their previous points of view. And so, you know, Elsie
26:53
Morrison, I think, is no longer with us, but he is smiling
26:55
back from having saying, yep. That's exactly
26:57
what I predicted would happen.
26:59
I
26:59
mean, it is fascinating to
27:01
look at history because you can see
27:03
these things repeating. And I think this three step process
27:05
is something that people should apply
27:07
to technologies of today and ask like where is
27:09
that technology or the way the
27:12
people are reacting to this
27:14
technology? Where are we in that cycle.
27:16
And then I think another fascinating thing for
27:18
people to spend time with is this idea
27:20
of the way that we view red
27:22
flag laws today, what will we view
27:24
or what will people view in a hundred years the
27:26
same way. Right? Like, what are the laws that we're
27:28
applying to social or AI or robotics or
27:30
space that regulation
27:32
has its importance, but what laws
27:34
are we influencing today that will seem
27:36
just as outlandish as someone
27:38
literally walking in front of a car to
27:40
make
27:40
sure that we don't scare
27:41
horses. So I would encourage people to think about
27:44
that. But Mark, I wanna ask you because you
27:46
have had a track record of being
27:48
early and right. And that's a
27:50
combination that not many people
27:52
can say about themselves. And I think it's
27:54
important, again, taking this framing
27:56
of people throughout history, being scared,
27:58
always looking things from this pessimistic
27:59
lens, how
28:01
have you been able to kind of see through that
28:03
noise? Like, what are you paying attention to when
28:05
you are introduced in your technology? It's
28:07
really, really easy to say, Well, these are
28:09
all the ways that this technology can go wrong. It's
28:11
not that easy to say, oh, I see
28:13
this light of how this could go right. So
28:15
how
28:15
have you been able to mold that
28:18
more optimistic lens
28:20
on technology? And how have you
28:22
also maintained that over time? Because I think that's an
28:24
equally hard problem to solve. Yeah.
28:26
So start by saying, like, I have the same instincts as
28:28
everybody else. Right? So when I when I read the scenes book,
28:30
I was like, yep. That's also describing me.
28:33
So I have the same, you know, same instinct. It's like somebody brings something forward. Yeah.
28:35
And by the way, like, right, new technologies in their
28:37
earliest stages, like, they're really half baked. You know, they're
28:39
sort of this concept, but the same to Lab
28:41
talks about tinkering it's usually
28:43
somebody in a garage or a metaphorical garage door door or
28:45
something or a computer lab somewhere who's basically working on
28:48
something. It doesn't quite work. You have to, like,
28:50
really squint and see why this would
28:52
ever be than a normal person could could ever
28:54
even understand, much less use, much less get value
28:56
out of. And so, you know, especially if
28:58
you're early, you know, if you're if you're doing, like, early stage
29:00
investing, like, we are really kind of, really advantaged acknowledges.
29:02
Like, you do see these things when they're early and they're
29:04
just not they're doing not ready yet. And so you have
29:06
to you you you basically the natural impulse is
29:08
very clear, which is the step one of the sim cycle, which
29:10
is basically just ignore. I have the
29:12
same instincts as everybody else. I think based on what happened to me.
29:14
It's basically this whole cycle kinda got beaten
29:16
into me, which
29:17
is basically, like,
29:19
a
29:20
very large number of times
29:22
over the last thirty years now when I have
29:24
had that reaction and that it turns
29:26
out that whatever the new technology was turns out
29:28
to be a really big deal. You know, it's
29:30
got like, if you if you go through that
29:32
loop enough times, like, at some point, I
29:34
don't know. It's, like, the tenth time or the fifteenth
29:36
time or something. Like, at some point, you're just like, okay. Like,
29:39
I need to stop. You know, at some
29:41
point, you realize that, basically, the cycle is a form
29:43
of self harm. Like, if
29:45
you wanna be a leading edge of new technology, you
29:47
basically have to break out of the cycle. You have to basically
29:49
stop holding yourself back.
29:51
Now, There's a
29:52
couple of nuances to that. Right? So one is,
29:54
you do have to swims. Like, you do have to
29:56
look at a new unformed thing.
29:58
Right? And then that and this
29:59
goes to the step two of the sentence. cycle.
30:02
But basically, like, rationally, like, has all
30:04
these problems. Right? And,
30:05
you know and, again, these problems usually in our
30:07
era, these problems are, like, it's too slow. It
30:09
doesn't scale. It's too expensive. you
30:11
know, whatever. It's too confusing, but doesn't whatever work
30:13
with the existing systems, like, whatever it is. Or it's
30:15
like a network effect thing, but nobody's using it yet. So
30:17
it's got the cold start problem. And
30:19
so you've of reasons why these new things get worked. And
30:21
you basically have people in a sprint to kinda look
30:24
through that and say, okay. Basically,
30:26
like, you
30:26
know, what if those things all get fixed?
30:29
Right? And and and, basically, the way that that the tech industry works is
30:31
is very helpful on this. If you spent a lot
30:33
of time with engineers, what you notice basically
30:36
is that list of kind of rational reasons why
30:38
something can't work. That list is also the same list
30:40
of all of the technology and
30:42
business opportunities with that technology.
30:44
Right? So I always felt that it's the
30:46
punch list of all the things for founders
30:48
to do. So, yeah, it's
30:49
a good example. Crypto web three has been going through this for
30:51
the last ten years. Right? Which is, you know, Crypto web
30:53
three bitcoin was was read
30:55
it early on. Ethereum was read it early on as, like, this can't possibly work,
30:57
all these different reasons. And basically, you
31:00
know, incredible engineers,
31:02
entrepreneurs, and the crypto web three space now
31:05
are basically fixing all of those things. By the way, today is
31:07
actually a great day to bring that up, right, because today's
31:09
the day the the sort of Ethereum
31:11
merge, what they call, the merges took place. And so
31:13
Ethereum actually just today
31:15
switch from the old method of proof of work to the new method of
31:17
proof of stake. Right? One one of the old
31:19
arguments against crypto, right, including Ethereum, was
31:21
basically the proof of word, but it's all this energy,
31:23
like that necessarily. proof of state doesn't
31:25
have that problem. And so, like, quite literally, the Ethereum
31:28
developer community has basically taken one of those
31:30
rational objections, like completely off the
31:32
table. And and that's just like a great example of how these things
31:34
actually do come to work. Like, here's the other thing. And
31:36
this this also an important one. Like, you're
31:38
not always right. Like, So
31:40
this way, just
31:41
because every successful new
31:44
technology is greeted initially as a
31:46
joke, right, does not mean
31:48
that every new technology that's created is a joke
31:50
is going to be successful. Right?
31:52
And so
31:53
and so you're not always
31:54
right. Like, sometimes, you bet on an early you know,
31:56
we do this all the time. We're an invention capital. You bet on
31:59
an early stage technology,
31:59
and it actually doesn't
32:01
work. It actually it actually doesn't happen. It
32:04
doesn't take. It doesn't become real. Right? And I'm
32:05
not even talking about, like, I'm not even talking about
32:08
fraud. I'm just talking about, like, we thought we had
32:10
the idea. We thought we had direct working on
32:12
it. You know, it just they just couldn't get the thing to
32:14
work or they just couldn't get the thing to work at the
32:16
price point if the market needed. Or by the way, a
32:18
lot
32:18
of the times, it's just it was ten or twenty or thirty
32:20
years too early. Right?
32:21
So virtual reality is a good example of this. Like, I I remember
32:23
when there was the first VR wave, I actually worked on
32:25
AAA bit when I was in college in the
32:27
in the late eighties. Like, there there was this big thing
32:29
around virtual reality time. And it just turned out that,
32:31
like, you just could build VR headsets that worked
32:34
properly twenty five years ago, thirty years
32:36
ago that the technology wasn't ready yet, and and and,
32:38
of course, today you can. And
32:39
so so a lot of times it's just being early, but, of course, being being
32:41
thirty years
32:41
early is the same as being wrong. It doesn't help you anyway.
32:44
Everything you're working on fails and then you have to wait
32:46
thirty years. And so you
32:48
have to be well, you know, if you're gonna kinda think in these terms, you have
32:50
to be willing to be open to the idea that you're only gonna
32:52
be right part of the time. Right? And
32:54
you have to be willing to take the chances anyway.
32:57
is kind of the way I think about it is. The way I think about it, like, for
32:59
our firm is, for every
33:00
new technology that we're exposed
33:03
to,
33:03
right, It's like, okay. Does it pass, like, the basic sniff test of,
33:06
like, okay. If this work, would it be a big deal?
33:08
Right? And then it's, like, are there really smart
33:10
people working on getting it to work?
33:12
And then, basically, if it passes those two sniff tests, then
33:14
we should probably be betting on it today
33:17
because we will be wrong some of
33:19
the time Right? But we will also
33:21
be right early some of the time. And then, you know,
33:23
the way venture capital works, startups work
33:25
is you if they sit there and make more money after
33:27
winners, then you then you lose on the things that
33:29
don't work. So you have
33:30
to kind of be willing to kind of tilt into the
33:33
risk. By the
33:33
way, some people shouldn't do this. Right? Like,
33:35
you know, there there are people for whom, you know,
33:37
ever being exposed failure is just, like,
33:39
too psychologically damaging to them or there's, like,
33:41
large companies, you know, that maybe shouldn't necessarily
33:44
completely reinvent their entire business on the
33:46
basis of a new improve and start up your calendar or something.
33:48
But, like, this is like a time and placement
33:50
for some people and not for other people. But for those of us
33:52
who want to be on the leading edge of new technologies,
33:54
we have to be very open minded. Howard Bauchner: It's also
33:55
a function of Venture capital. Right?
33:58
The the business or the industry that
33:59
you're in
33:59
naturally requires some level of risk so
34:02
that you get the reward at the end
34:04
of the table. But on that
34:06
note
34:06
of timing. I
34:07
wanted to ask you about this because, of
34:10
course, there are examples of specific
34:12
companies. Like, I'll just use a simple one segue.
34:14
Segue didn't work out. A bunch of people thought
34:16
it would. Okay. So that was truly something that, you know, we probably
34:18
won't see in the future. But when we're talking
34:20
about larger industries, you use
34:22
VR as
34:24
an example, Are there
34:26
really examples of technologies
34:28
that a bunch of extremely talented
34:30
people are working on? Again, really foundational
34:32
technologies, whether it's AI
34:34
or crypto, insert other foundational technology here
34:36
that eventually doesn't work out.
34:38
And I'm asking you this because actually several
34:40
of us on
34:42
editorial team. We're trying to think of an example and
34:44
someone who has been investing in technology
34:46
for so long. I was wondering
34:48
if there has been a case where so
34:50
many people have been
34:52
wrong? Well, the
34:52
big probably famous example was Alchemy,
34:55
which is
34:55
AAA term at two eleven. It's AALCHEMY
34:58
So Alchemy was the technology to transmit
35:00
lead if gold. And there
35:02
was this concept that you'd be able to that you'd be
35:04
able to build some form of machine or discover some material that
35:06
they actually and this is, like, you know, three
35:08
hundred years ago now. they actually had this training called the philosopher's stone.
35:10
And there was and basically, all these smart people
35:12
trying to figure out how to basically venture discover
35:14
the philosopher's stone. By the way, Sorry about
35:16
smart people. Isaac Newton spent twenty years trying to figure this out.
35:19
Right? So, like, maybe the smartest human being
35:21
in history of the planet spent twenty
35:23
years on this. So at at the time, you know, they were
35:25
serious. The and the dream of course was, you know, transmitting light into gold.
35:27
Why do they wanna do that? Of course, let us
35:30
plentiful and and and worthless in
35:32
gold. And scarce and and incredibly
35:34
valuable. And so, you know, they they were searching for
35:36
kind of the magic formula for how it basically essentially,
35:38
right, create wealth, you know, basically kind of turbocharged
35:40
economy. kinda make society better in
35:42
one step, and they never figured it out. By the
35:44
way, it's an interesting question. If you happen to know
35:46
any material scientists, I don't think there
35:48
are any active research programs today. literally
35:50
turning red into gold, but it it would be an interesting question for any material
35:53
scientist, you know, whether, oh,
35:55
there is actually an example of this.
35:57
There are now synthetic diamonds. And
35:59
there's now technology for actually turning carbon
36:02
into diamonds. And so, you know,
36:03
what what one could argue, we didn't we didn't get gold at the
36:06
other end, but maybe maybe they got the diamonds
36:08
hard to work. But I cite
36:10
that example because, like, you know, that was, like, three hundred
36:12
years ago. And, you know, it wasn't usually back now
36:14
and it's, like, that wasn't really science as we understand
36:16
it even even for noon at that point. It
36:18
was kinda there was, like, a lot of religion involved and they they were still, you know, figuring some
36:20
really basic thing. I mean, this is when
36:22
Newton was still working on his, like, three laws of, like, how
36:24
the universe works. And so they were they were still trying to
36:26
get the basics
36:28
figured out. You'll look more recently,
36:30
you know, look, there there's a lot of things people
36:32
talk about today that aren't working yet. And,
36:34
you know, we we could have a very long discussion
36:36
about those. obviously, we can't
36:38
see the future. You mentioned segue as an
36:40
example, like, I would guess that that comes back.
36:42
You know, I I would guess that there will come
36:44
a time when people will realize that that actually was a really good idea. But
36:46
I'll I'll actually spend a second on segue.
36:48
Right? So the the theory of segue
36:50
actually was a two part theory. The third part one was the
36:52
the device
36:54
itself. And, you know, it it greeted it got the backlash right up front for
36:56
a variety of reasons just like the car did. And so it
36:58
it became kind of this running joke, and if you
37:01
if you like the very funny TV show arrested
37:03
development, they take the goofiest, you know, kind of
37:05
biggest asshole character in the show they put him on a
37:07
segue. So it became kind of this running joke at the
37:09
time, but But the device itself was only part it was only part
37:11
one of the theory. Part two of the theory was that cities
37:13
would get redesigned around the device. Right?
37:16
And so the theory basically
37:18
was like why are cities laid out the way they are? Cities are laid out the
37:20
way they are because of the car. Right? I'll
37:22
give you an example. Like, there are I think the
37:24
number is, there are something on the order of two
37:26
billion parking
37:28
spots. in the United States. Right? So there's, like, mass in
37:30
the total amount. If you're parking lots, it's something like there's
37:32
the parking lots. I think if you put them all together in the US,
37:34
it's something like the car they would cover the entire state
37:38
of Connecticut. Right? It's
37:39
just this, like, massive amount of space devoted to basically roads and then
37:41
parking, you know, for cars. And then plus there's all
37:43
the issues, there's all the safety issues of
37:45
cars, there's all the evolution
37:47
issues for cars. You know, there's all the noise issues
37:49
and so forth and so on. So, basically, with
37:51
with the segue guys, the time thought was really
37:53
what should happen a city should get redesigned. A city should get
37:55
redesigned, assuming that there are no cars. But, you know, you don't just
37:57
want people walking around,
37:58
you want people to go move faster
37:59
than that. probably
38:02
don't wanna bring horses back. They have
38:04
other issues. And so if you re if
38:07
you design a city from scratch, you could basically design
38:09
it with sidewalks and pass that
38:11
you could have, like, lots of different segue
38:13
powered style right things, including, like, you
38:15
know, single passenger ones, but also maybe, you know, a little,
38:17
you know, cards be four people, six people.
38:19
you devices and so forth. And so, you know,
38:22
it may be that just what hasn't happened yet
38:24
is nobody's actually trying to build
38:26
that city. Right? May maybe
38:28
there just needs to be a new kind of city. And by the way, maybe
38:30
at some point somebody will do that and all of a sudden it'll be
38:32
like, wow.
38:32
Those segue guys whatever forty years ago were
38:34
actually under the right thing. And so If
38:36
a lot of smart people are working on something, it's like virtually guaranteed that it's
38:38
gonna happen. It's just a question of when. And
38:40
and like I said, the when might be forty
38:42
years out, and so it it might
38:45
not be, like, it might not be those people who get the
38:47
benefit or who kinda harp us the gains from
38:49
doing
38:49
the new thing. But, like, it ultimately will
38:51
happen. And and If you
38:52
go back across many is something we can spend a lot of time on. But if you go
38:54
back across a lot of historical technologies, you know,
38:57
it took, like, fifty years to
38:59
get the TV to work there
39:01
were optical telegraph systems fifty years
39:04
before the telegraph systems that we
39:06
became familiar with.
39:07
They they have, like, optical telegraph systems working in Paris
39:09
in, like, the eighteen twenties, eighteen thirties. The fax machine was
39:11
invented in the
39:12
eighteen seventies. It wasn't commercialized
39:14
until the nineteen seventies. It took a hundred
39:18
years. Right? The computer, you know, the computer took,
39:20
like, fifty years to get into
39:22
consumer form and another twenty years after
39:24
that to get into
39:26
your pocket. Right? And so the history here is these things often just do
39:28
take a long a long amount of
39:30
time. And my conclusion for that is basically, it's all
39:32
gonna happen. It's just this this sort of
39:34
massive question
39:36
of timing. Yeah. And on this idea of timing,
39:37
sometimes all it takes is a change in regulation,
39:39
a change in another technology becoming mass
39:42
market that
39:44
allows you know, a follow on Something heard you about before
39:46
are kind of like these unlocks
39:48
that happen throughout history that again
39:50
have follow on effects. And
39:53
one of them that you've mentioned from way back is the ability for
39:55
people to own land and how those incentives really
39:57
spurred a wave of innovation because people
39:59
had the incentive to build
40:02
on top. Are there other unlocks that you're paying attention to
40:04
today that you've noticed, let's say, in the last five or
40:06
so years, it could be COVID that allowed a
40:08
bunch of people to work online
40:10
and that's the next industrial
40:12
revolution, are there things that you
40:14
think are really, really meaningful from the
40:16
last, again, let's say,
40:18
couple years? Yeah. So
40:19
the big one is sort of the post COVID world. So so the big one is kind of, you know, the
40:21
rise in kind of remote work, virtual work. And the
40:23
reason I
40:23
I say that's the big
40:24
one is because it seems relatively straightforward. Well,
40:28
you know, like you and I are recording this more in different locations where you're coming in across webcams.
40:30
So it seems like it's just like a new way
40:32
to work, but it's actually deeper than that. for
40:35
the following reason,
40:36
which is basically the economic
40:38
role
40:38
of cities for basically all of
40:40
recorded human history, again, going back
40:43
for a thousand years, the
40:44
rule of cities basically is what it's what economists think they call
40:47
it agglomeration. So the rule of cities is
40:49
to basically get a critical mass of people
40:51
in a single place. where
40:53
those
40:53
people are able to come together and basically do things
40:56
that are greater than than those people can do as
40:58
individuals. Right? And that ultimately led to the creation of
41:00
companies and led to the creation of, like, you know,
41:02
all these technology and science and all all
41:04
these other things happen kind of as a consequence of
41:06
kind of culture culture came out of cities
41:08
like almost everything today that we would consider to
41:10
be kind of good about kind of giving kinda came out of the that
41:12
people got into cities. Like, all the invention basically
41:14
takes
41:14
place in cities. And so the rule
41:16
of the
41:16
city was basically, okay. You wanna basically
41:19
track the sort of smartest, most ambitious, most
41:21
innovative, most creative people anywhere in society no
41:23
matter where they grow up, whether it's something like a
41:25
small town or on a farm or, you know, by the way, in
41:27
another country or whatever. and
41:29
you wanna basically bring them to a city.
41:31
The
41:31
economists also have this term they call super star
41:33
cities. Right? And these are the cities that
41:36
basically turn out to be, like, ground zero
41:38
know, a fundamental kind of revolutionary kind of thing. Right? And
41:40
and become kind of a permanent hub. And so,
41:42
historically, the San Francisco Bay area has been
41:44
the superstar city
41:46
of technology. Right? The way that works is if you're a ambitious
41:48
technologist, you want to go to
41:50
the San Francisco Bay area, you want to be
41:54
small fish and a big pond because you wanna be around all the other smart people because collective effect
41:56
is gonna be so powerful. Los Angeles is
41:58
a superstar city
41:59
for film
41:59
and television. New York is a superstar city
42:02
for finance for fashion and, know,
42:04
arts and all kinds of things. London, Paris,
42:06
these big cities have played this kind of outsized role in
42:08
economic history. But they they've all been based on this
42:10
idea that you have to get off
42:12
smart people together in one geographic place so that they can actually meet each
42:14
other and talk to
42:15
each other and work together and do projects
42:17
together, bounce ideas off each other,
42:19
challenge each other. again,
42:21
this is like
42:21
four thousand year history of how progress
42:24
has happened. All of a sudden, for the
42:25
first time in four thousand years, we
42:27
now have both the technology and the form of
42:29
the Internet and
42:31
Zoom and
42:31
webcams and remote work and collaboration tools and Slack and
42:34
all these, you know, amazing technologies we have
42:36
now. And we now have this,
42:38
like, sudden proof right,
42:40
that during the very, you know, kind
42:42
of bad, unpleasant dangerous, COVID
42:44
lockdowns, it turns out basically the
42:46
companies were basically able to just
42:48
keep running. any company with
42:50
knowledge work was able to just keep running, you
42:52
know, all the way through COVID. To to a level
42:54
of success that, like, nobody envisioned was even
42:56
possible before COVID. And so now, of course, you got
42:58
this massive societal changes underway. And by
43:00
the way, I think this societal changes from though that are
43:02
just starting, you've
43:02
got this massive societal changes underway where all
43:04
of a sudden people can say, like, wow,
43:06
don't have to be in the San Francisco Bay area
43:09
or in New York City or in Paris or wherever it
43:11
is in order to be part of the computer
43:13
industry or the music industry or the
43:15
movie industry or, you know, financer,
43:16
like, whatever it is. By the way, my kids are not gonna
43:18
have to be there. Right? So even if I'm raising kids
43:20
and I'm worried about their future, like, they
43:24
don't really have to be there. So all all of a sudden, like, the potential fundamentally
43:26
disconnect where people live, from where people work
43:28
has basically
43:28
been open. And then, of course, if you can
43:30
separate the where, you can also separate the
43:34
how. Right? And
43:34
so what kind of community,
43:37
right, would be the best for for example,
43:39
for kids to grow up in? are
43:41
the communities that we built in the last hundred years
43:43
where it's been assumed that you have to kind of be
43:45
part of these specific locations to fully
43:48
participate in economic life. are these the actual
43:50
kinds of communities you wanna raise kids in? Or is there actually a
43:52
completely different kind of community you would build
43:54
if you didn't have to be on a specific place
43:56
to be able to have great jobs? we all probably know
43:58
people like this, like a lot of people are
44:00
fundamentally reexamining. Like, whether they wanna do
44:02
their lives? Like, do
44:03
they wanna work? Do they wanna work in
44:05
the industry that we're getting before? they wanna
44:07
work at the employer they're working at before. Right?
44:09
Do they wanna move to, you know, another
44:11
country, like, they're, you know,
44:14
completely rethinking Right? How many kids do they wanna have? Like, whether they wanna have kids, like,
44:16
all these things? You know, do they wanna get
44:18
reconnected
44:18
back to the extended families that they were forced
44:21
to move away from? all of like,
44:22
fundamental questions are are being asked. And and
44:24
this, I think I think fifty years from now, I
44:26
think we'll look back. It'll be, like, basically, the Internet
44:29
and then this. we're basically the two big things that happen. Howard Bauchner:
44:31
Yeah, I I have to agree with
44:32
you. I've been working remotely for probably seven
44:34
years now. So a little bit
44:38
before COVID, But something that I'm finding fascinating
44:40
is this idea of remote
44:42
work, in my opinion, being a
44:44
technology or at least there's technologies that
44:46
enable it and similar to
44:48
many other technologies as we've talked about throughout
44:50
this conversation, there is a substantial
44:52
backlash. And also, you know, I
44:54
have to give credit to the people
44:56
who are pushing back on remote work because there are things to be
44:58
fixed just like we talked about. Every technology
45:00
has problems to fix and those are
45:02
business opportunities. But
45:04
I wanna hear from you from the perspective
45:06
that you mentioned before of this power balance
45:08
being messed
45:09
with with every
45:10
new technology and as more people
45:13
start to work remotely and reconsider all of this. How
45:16
would you frame the pushback of
45:18
remote work with this
45:20
idea of a power shift? Or
45:23
society reshuffling? Howard Bauchner: Yeah, so
45:25
this really comes up I mean, there's a whole bunch
45:27
of angles on this, but like the one that's the one that I
45:29
hear about all the time right now is when I talk to
45:31
big company CEOs. Right?
45:32
So you talk to CEOs of, like, big banks or big
45:34
software companies, right, big whatever, you know, companies. You know, this
45:37
basically is the
45:37
conversation. Right? because because what
45:39
happened is basically you're the CEO
45:41
of a big company today, you came up
45:44
in an environment where everybody was in physical
45:46
proximity. Right? The whole way that you came up,
45:48
the whole way that you played politics
45:50
and, like, got yourself in position, that got yourself exposed, and important people, and got
45:52
promoted, and, like, all this stuff, you know, did
45:54
work with your teams. Right? The way that you
45:56
did deals, the way that you had relationships
45:58
with other people industry
45:59
with customers. It was all based on in person
46:02
proximity. Right? And and which meant either in the
46:04
office or by the way, it meant on the road. Right?
46:06
Doesn't travel. And that's true
46:08
basically of, like, the entire management hierarchy of,
46:10
like, every existing big company. Like, they're they're all
46:12
like that. Right? And so all of a sudden, now
46:14
you have this incredible kind of phase
46:16
shift happening And so now it's like, okay. Like,
46:18
a
46:18
whole bunch of questions open up. Like, how does the company
46:20
organize? You know, what is the
46:22
balance between between person and and remote?
46:25
one
46:25
of the implications of that for the org chart.
46:27
Right? You know, job roles change.
46:30
Geographically, worship companies be located. Right? I'll
46:32
give you an example. I talked to the head of one
46:34
one big company that has based on Manhattan,
46:36
they've got about twenty percent of their employees are the
46:38
super advanced knowledge workers. About eighty percent are
46:40
kind of back office, white collar
46:42
clerical workers. know, the SEO is like, look, like, the eighty percent clearly don't need to be in
46:44
Manhattan anymore. Like, we can put them in, like, South
46:46
Dakota, you know, we can pay them, like, half as much.
46:48
They'll be their standard of life will be twice
46:50
as high. that,
46:52
you know, that'd be a lot happier. And, you
46:54
know, we we can do that. Right? Now
46:55
he says,
46:56
I think we still need to keep the twenty percent of the creatives
46:58
together. But, like, by the way, the creatives
47:00
get a vote too. And if they decide to leave the
47:02
in person job and go remote and work for
47:04
another, you know, company that industry that's one hire remotely,
47:07
like, they have the ability to do that.
47:09
So Like,
47:09
a lot
47:10
of these companies, I think, are gonna, like, really dramatically restructure
47:12
over the course of the next five years.
47:14
A lot
47:14
of these CEOs, basically, the big
47:17
company CEOs are like, well, we could never go forward mode. Like, that's impossible. We can't do
47:19
it. You could talk to startup CEOs and they're
47:21
like, well, maybe we can.
47:24
Right? And
47:24
so here, I'll I'll talk our book for a second. I'll kind of you
47:26
know, I'll get as close as I can to kinda say that we
47:28
might have something figured out. Look, it may be
47:30
that the remote work revolution is just really
47:32
bad for big company. Right? It just may
47:34
be that it takes existing systems and models for how big
47:37
companies operate, and it basically breaks them. And it
47:39
may be that remote work is
47:41
the kind that you need to build a new company to be able to know how
47:43
to do properly. Right? because you need to kind
47:44
of build a culture from scratch. You need to build systems
47:47
from scratch. You need to build processes
47:50
from scratch. need to build it may just be remote companies just need to get built
47:52
differently. And it might actually not be possible
47:54
to reconstruct a big old company that's
47:56
done things, why do I have to just restructure it so
47:58
that they can do things a completely different
47:59
way. And so, you know, this may
48:02
accelerate the process of turnover where
48:04
some big companies go way faster and some new
48:06
companies get much bigger faster. Like,
48:08
that's
48:08
a possibility. or
48:10
by the way, you know, maybe
48:12
I'm full of it. Right? And maybe the opposite is true.
48:14
Right? Maybe it turns out remote work just isn't good enough.
48:16
Or maybe maybe it turns out we actually don't have
48:19
remote Maybe we don't have the technologies yet. Right?
48:21
Maybe we need holographs.
48:24
Right? Maybe we need, you know, big
48:26
teleconferencing rooms. Maybe we need I'll give
48:28
you an example. maybe every
48:30
company in a certain site should own it. So, like,
48:32
literally, it's on hotel resort. Right?
48:34
And and so maybe what happens is
48:36
basically, like, every team should be basically in residence that are really cool
48:38
and tell a resort location for, like,
48:40
you know, a month in the spring and a month in
48:42
the fall. in a resort, and maybe
48:44
the company should just be running their teams through that.
48:46
And unless they do that kind of thing to
48:48
be able to have critical mass to bonding, like,
48:50
maybe this won't work. And, like, maybe that's a new kind
48:52
of real estate at if it has to happen,
48:54
you know, to
48:54
be able to have basically sort of think about this
48:56
sort of corporate resorts. In other words, to be
48:58
an attractive enough place to come where you're willing to
49:00
actually be away from home for a month maybe you can even
49:02
bring your family with you. Right? But everybody gets to
49:03
work together. Like, you know, those don't exist today.
49:06
And I'm just speculating. But, like, maybe there are
49:08
things like that that we have yet to
49:10
figure out. I just
49:11
have, like, a very strong sense that on the other side of
49:13
this, the the world's gonna work very differently.
49:15
I
49:15
do too. And what you're saying reminds me
49:17
of something that when I had a chat
49:19
with Neil Stephenson, we we talked about
49:22
which is when electricity was invented, it
49:24
was very easy to think, okay,
49:26
electricity can can be applied to lighting. Okay? That makes sense. People can see
49:28
that line. But the example he gave
49:30
was electric guitars. Not many
49:32
people when electricity was first being
49:34
invented could think, oh, okay. Well,
49:36
we're gonna have electric guitars from this. And I think
49:38
and when you have these really foundational shifts like
49:40
remote work, it's very hard to think about their
49:42
second, third order effects, but you do know
49:45
they're gonna happen you don't you just don't always know exactly how
49:47
they're going to manifest. But I wanted
49:49
to ask you quickly because a sixteen z has
49:51
moved to the cloud recently Was
49:53
there an aha moment or was there a
49:56
specific thing that got you to
49:58
change your mind? Or were you always kind of,
49:59
you know, coming
50:01
around to this idea because a lot
50:03
of people say strong opinion is weekly
50:05
held, but not many people actually operate that way.
50:07
And so I think that it was interesting to see a sixteen
50:10
z actually make a shift of previously being in the office
50:12
to moving to the cloud. So was there
50:14
something that kind of
50:16
unlocked that? Yeah. So I
50:17
was very people who know me will confirm this.
50:20
I was very anti remote work basically
50:22
before COVID. I was probably
50:23
pretty far on the on the
50:25
end of, like, a spectrum of people thinking about this and not not being open minded on it.
50:27
And that was true of us as a firm. Right? And so we historically
50:29
ran as a firm. We ran in actually a single office. We had
50:31
a single office, excuse me, at
50:33
MMO Park. actually even refused to
50:36
open an office in San Francisco for for ten years
50:38
because we just we wanted my critical mass in the
50:40
office. And by the way, people who
50:42
actually worked or visited our office between, call it, two
50:44
thousand ten and two thousand twenty, they will tell you, like, it
50:46
was a high vacuuming. It was a really cool
50:48
place. There was always all this all this amazing stuff
50:50
happening and always amazing. People walking around and, like, you
50:52
know, we wanted that a hot house
50:54
environment, and and we got it. But that man,
50:55
everybody, like, had to be in the office. And then the
50:58
other
50:58
thing I always maintain was I was like, look, like, there's
51:00
this theory remote work but, like, it hasn't yet worked for any company
51:02
at any level of scale. Like, the only companies that
51:04
have made it work are very small. And then I
51:06
said, look, just look at the data. Right? And
51:08
what the data basically says is this
51:11
superstars anything I mentioned is basically
51:13
in full effect. And with the database that they
51:15
said was venture capital, you know, the funding
51:17
for technology was actually concentrating more and
51:19
more into the Bay Area. Right?
51:21
Between between twenty ten and twenty twenty. If you look
51:23
at just like the money was flowing more and
51:25
more into the single place, the people
51:28
were flowing more and more into the
51:30
single place. I
51:30
mean, the the and to clear, bursting the seams
51:32
and housing became crazy expensive and transit
51:34
to disaster, and our politicians hate
51:36
us. Like, there's all these issues.
51:39
But this Hot House environment,
51:41
this agglomeration Hot House, you know,
51:43
kinda thing was real. Right? And it's and
51:45
it's the thing that traditionally made the value so
51:47
special and it's why was always third of
51:49
the hub for, you know, this has been a hub for forty years where almost all the great breakthroughs in in
51:51
in technology that's happening. And so I was
51:54
very much kind of on
51:56
that page. So for me, it
51:58
was literally going through COVID. Right? And again, we were
51:59
looking at this a couple different levels. We were looking to look at
52:02
this level of, like, how's our firm
52:04
gonna operate? you
52:04
know, under lockdown. We were also thinking about, like, how our our company is going
52:06
to operate under lockdown. And then we were also
52:08
thinking about, okay, how is the industry as
52:12
a whole? right, going to operate under lock down. Like, for example, like,
52:14
how will new companies get started? If a new
52:16
company starts under conditions of lock down, like, how will
52:18
that actually
52:20
happen? And
52:20
then basically, we we, you know, on a
52:22
completely involuntary footing, we we basically
52:24
ran this experiment for two years. things
52:28
we did is we did surveys the entire time. So we
52:30
did we did repeated sampling of the of the different constituencies
52:32
and kind of ask, you know, to try to get
52:35
that data kinda handle on what was happening. And so it was
52:37
really interesting because it's, like, six months in, like, people are,
52:39
like, super nervous and it's, like, are these companies
52:41
even gonna continue working? And then it's
52:43
like twelve months in, people are like, well, actually, it turns out remote work. At least for now,
52:45
it works just as well. If anything, maybe even
52:47
better because people don't have anything else to do.
52:49
So they're actually working
52:52
more. Right? So the new problem is burnout, you know, eighteen months later, it
52:54
was, well, you know, actually, we're gonna start hiring
52:56
remote people now that we can do
52:58
remote work. We can now
53:00
people who don't live, you know, near near the campus, so we hire more broadly. Right? And
53:03
and and so forth. And so you can
53:05
actually see in the data. You could actually see
53:07
basically the sort of beliefs moving.
53:10
And then and then basically what what Ben and I decided with our, you know, with
53:12
our colleagues with the firm basically is like, look,
53:14
the firm is built to live in
53:16
the future. The firm is built to
53:18
to fund the best new companies and work with the best entrepreneurs build in the future. We wanna
53:20
be able to lead the industry. This is such a
53:22
big change for the
53:23
industry that, like, we have to limit. Like, it doesn't
53:25
make any sense to try
53:28
to do a model where we're basically finding the new model. Like, we have to live the new model.
53:30
I think it's going very well.
53:31
Like, having said that, like, we're still figuring it
53:33
out. Right? So true you're experiencing this. It's just like,
53:36
we're still figuring
53:38
out, like, So
53:38
for example, in our firm, we we we put a big premium on having off-site.
53:40
Like, we we now take off-site really seriously. We
53:42
really take getting people together on a regular
53:45
basis very seriously. we
53:46
have, like, whole teams inside our firm now that are devoted just, like, orchestrating all that and
53:48
making all those happen. But, you know, do we have
53:50
the right balance of, like, remote versus off-site? Do we
53:52
need to have you know, people have
53:56
more some time. Do we have the right tools? And since it's in place, are we running
53:58
the first and we have the right management practices? And
53:59
so, you know, we're we're certainly still figuring it out,
54:02
but we're definitely on
54:04
that path. And
54:04
like we talked about, that comes with time. But the reason I asked you how
54:06
you made that shift or almost in in some
54:09
way changed your opinion as you
54:11
got new data, is because I wonder how
54:13
perhaps we might be able to that within the framing of how people view
54:16
innovation and technology, kind of returning to what
54:18
we talked
54:20
about at the very beginning of this conversation,
54:22
it does feel like there is this perception of technology. Of
54:24
course, not everyone holds it, that
54:27
we're in a bad place or, again, the world's getting worse
54:30
or or negative thing about our current
54:32
state of affairs. And I
54:34
think
54:34
it's really inspiring at at
54:36
least at a sixteen Z to see that people do hold this very optimistic view.
54:38
And I wonder what you
54:40
think maybe we can do as
54:43
a collective, as a society to maybe
54:45
orient more around this more positive
54:48
view of technology because from
54:50
my understanding and I'd actually love for you to go into
54:52
this history, it doesn't sound like this was
54:54
always the case. That technology I
54:56
mean, specific technologies were
54:58
viewed quite negatively, but a sense
55:00
around innovation, I think, has
55:02
been different in the past. Is that
55:03
right? Yeah. So that's
55:04
a good point. So technology has always kinda gone through
55:07
the cycle, and I I would go so far as
55:09
to say, like, the technology
55:10
adoption cycle resistant cycle, like, I'm
55:12
not sure actually gonna change. Douglas Adams, the the the science fiction novel, Douglas
55:14
Adams, who wrote the hitchhiker's guy to the galaxy,
55:16
had had another take on this. of
55:19
Yeah. both very fun, very funny, very serious. Right?
55:21
So he said, basically, it's, like, any
55:24
technology that existed before you were, like,
55:26
fifteen years old is just, like, the natural order
55:28
of things. any technology that
55:30
gets invented between the time when you're, like, fifteen
55:32
and thirty five is, like, new and exciting and
55:34
cool and cutting edge and, like, maybe you can make a career
55:36
in it. And then
55:37
any new technology that arrives after the age of thirty five is unholy and against, you
55:39
know, against the against the natural order. Right?
55:41
And it's gonna bring doom
55:44
to civilization. another way to kind
55:46
of putting the putting the same cycle. So so,
55:48
like, the the the the most kind of I don't know. The most negative
55:50
thing you could say or something would be like, there's just
55:52
this, like, permanent generational psychological thing
55:54
this cycle where another famous quote
55:56
I think it's a great physicist Max Planck once said
55:58
science advances one pug
55:59
girl at
56:02
a time. Meaning that, like,
56:02
you know, in science, you need, like, the old senior scientists who have,
56:04
like, one paradigm that, like, quite literally die off so
56:07
that the young scientists of the new paradigm
56:09
can actually, like, take over. you know,
56:11
that sort of view be is so deeply baked into the
56:13
mentality of how people operate psychologically, that it's
56:15
it's just like the permanent
56:17
state of affairs.
56:19
By the way, like, I don't know if that's
56:22
optimal for society. I will tell you that state of affairs
56:24
is very good for entrepreneurs. Right?
56:26
Because if more people
56:28
in positions power, we're more open to new technologies. Right? The
56:30
the opportunity, especially for the young entrepreneur
56:32
with a breakthrough idea, would actually diminish. Right?
56:34
because the big gold companies already be doing all the
56:36
new things.
56:37
So it it may be, like, what one
56:40
response to me would be, like, Mark shut
56:42
up. Like, stop talking to
56:44
people about this because we
56:46
actually want everybody in position of power to just assume that all new technologies are stupid
56:48
and evil because we want all the opportunity to be
56:50
available to all the kids who are starting all the
56:52
new companies. So so,
56:54
yeah, so that's one side of things. And then, yeah, the what
56:56
you mentioned, so there's broader societal things. So then
56:58
there's this broader societal things happening, which is
57:00
basically this the
57:01
sort of society's different
57:04
societies have sort of different
57:06
ethics and belief systems around, even
57:08
just like the fundamental idea
57:10
of progress. Right?
57:10
And so I I won't pick on. I won't I won't name names. But let's just
57:12
say there are certain societies in global history
57:15
that, you know, at different
57:18
points in time have decided, like, we're just not gonna do new things.
57:20
Like, we're not gonna talk to outside people,
57:22
we're not gonna adopt new technologies. I mean, there
57:24
are societies on a planner, you know, there I
57:26
mean, to North
57:28
Korea. You know, there are societies like
57:30
that today. Well, another example, you
57:32
know, the Amish, like, have a in entire
57:34
religion and belief system right around the the the fact that they don't have not
57:36
new technologies. By the way, like, I'm a
57:38
I'm a free market, free mind sky. I think people should live
57:40
how they want it. If they wanna live that way, I
57:43
think that's fine. it's a choice. You can choose. Right? To not like,
57:45
think progress is a good thing. To not think that your
57:48
technology should be adopted. I would argue in
57:50
the fullness of time that,
57:52
you know, it's it's it's hard to have your quality of life be at
57:54
the same level as if you're you're more open into
57:56
technologies. But, like, you know, societies do
57:58
decide that
57:58
they don't want
57:59
new technologies. if you kind of
58:01
look historically at the US and more
58:04
broadly kind of the west, you do
58:06
basically see this pattern where there was
58:08
a lot of resistance in
58:10
new you know, call it from the end of the Robin Empire through to, basically, the
58:12
Renaissance since they're called dark ages. Right? So
58:14
there there was, like, a whatever, twelve hundred year stretch
58:16
or something where not
58:18
much happened. And then basically, you know, over the last five hundred years, like, in the west,
58:20
there there's sort of this ethos of progress that kind
58:22
of emerged, you know, since the total
58:24
enlightenment, the scientific revolution, and
58:26
industrial revolution. Right?
58:28
And so forth, and then the rise of protestantism, which was very important because it
58:30
meant that people can seek out the answers to life mysteries on
58:32
their own. Right? And then the rise
58:34
of what was called philosophy that
58:36
became science. And so there was sort of this
58:38
system that was developed to basically go uncover
58:40
basically scientific truths and then
58:43
and then build allergy is that that basically builds, you know, what we
58:45
consider to be like a modern capitalist economy on
58:47
top of that. If you asked
58:49
people in eighteen eighty or
58:51
nineteen hundred or twenty or even
58:54
nineteen sixty. If you ask
58:56
most people if they thought that that whole set of
58:58
things was a good idea, most people would have
59:00
said yes. This is
59:00
a sort of the era known as modernity. Right? So this was, like, basically, this
59:03
is the era of, like, you know, we want
59:05
a progress of civilization. We want higher
59:07
standards of living. wanna be able to
59:09
look back and say, like, yes, we, like, we advanced
59:12
civilization as compared to what our
59:14
forefathers had. And if you
59:16
read, like, get books written in those eras. They, you know, they only take great
59:18
pride in in in all the progress that's been
59:20
made. In the west, in the last fifty
59:21
years, that's kinda
59:24
gone sideways. there's
59:24
an ethos in the west that kind of started in the sixties and into our era, which
59:26
basically says, well, maybe, you know, a lot of the stuff
59:29
is not so good, maybe it's like bad food
59:31
environment, maybe it's bad for whatever
59:33
whatever like,
59:34
there's, you know, there's a whole bunch of different arguments like this
59:36
and, like, maybe we've had enough progress, maybe
59:38
we've had enough science, maybe we've had enough technology.
59:41
the sort of classic example of this is, like, nuclear
59:43
power. Right? Like, you know, we invented a way to
59:45
basically have a limited clean energy, and then we just,
59:47
like, decided we don't we don't wanna
59:49
QTL people in nineteen ten that they could have
59:52
nuclear power. Like, there would be, like, ten thousand
59:54
nuclear plants, like, in the United States
59:56
running today. but you tell people in
59:58
nineteen seventy five or nineteen, you
59:59
know, ninety eight or, you know, whatever, twenty
1:00:02
twenty two, they can have nuclear power. They're like, yeah. Let's not
1:00:04
have any of that. And so
1:00:05
I I think there there has been this sort
1:00:07
of negative cultural shift. Tyler Cohen
1:00:09
calls this the complacency, raw
1:00:11
raw stat, that calls this a
1:00:13
decadence there's shift to, like,
1:00:15
things are good enough. We
1:00:16
don't need more of this. You know, or or the
1:00:19
more extreme form, which is, like, all
1:00:21
progress bad. Like, all these new technology. Like, they're just flat out bad. It's all
1:00:23
bad. It's all bad. It should all stop. You know, kind
1:00:25
of the the Univar kind of
1:00:27
argument. Right? Yeah. But
1:00:28
people weren't born
1:00:30
with that mentality. Right? Like, people weren't
1:00:32
born with progress as bad. Right? They've they've learned
1:00:34
that. And I do wonder, is it just because things
1:00:36
have been so good? Right? Like, that we've we've done
1:00:38
so well. You mentioned specifically this mentality within
1:00:41
the west that people
1:00:44
no longer have something
1:00:46
to strive for even though they certainly do. But
1:00:48
is that just a reality that things
1:00:50
have gone so well that people have obtained
1:00:53
this mentality? So that's a
1:00:53
theory. That's a theory. Some people are calling out in this sort of
1:00:56
upper income trap. So, basically, it's like
1:00:57
the theory basically is, yeah, once people hit, like,
1:00:59
an upper middle class
1:01:02
standard living, where they're
1:01:02
kinda like, you know, they got a house, they got a car, you know, they got whatever, they got, you
1:01:04
know, college, they got, you know, a hospital nearby. And,
1:01:07
like, they got Netflix. got,
1:01:10
like, whatever, you know, the good restaurants
1:01:12
nearby, whatever. Like, at some point, they're just like,
1:01:14
damn. Like, it's fine.
1:01:15
You know, it's it's good enough or
1:01:17
anything beyond this. It's probably excessive. some
1:01:19
people are joking right now that this explains what's
1:01:21
called the Fermi Paradox. So the
1:01:23
the the Fermi Paradox is this question of,
1:01:25
like, why do we not know of other civilizations
1:01:27
in the galaxy? like,
1:01:28
why do we not know of any, like, alien civilizations? Like, you know, billions billions of
1:01:31
planets, like, throughout the universe and, like, we should be
1:01:33
able to, like, pick up signals be
1:01:36
able to get, like, TV broadcast or whatever for the civilizations traveling
1:01:38
across space and, like, basically, why why is it
1:01:40
still after all this time with all these radio telescopes? It's
1:01:42
just still, you know, humanity on planet Earth.
1:01:45
so this this argument basically goes, it's basically the
1:01:47
the same thing. It's just like, yeah, other
1:01:49
species of aliens develop the same level of
1:01:51
upper middle class lifestyle that we have, and then they all
1:01:53
just kinda shrunk its you
1:01:55
know, well, you know,
1:01:58
look, you
1:01:58
know, here we sit, like, so mankind, you
1:01:59
know, we went to the moon. Like, we went
1:02:01
to the freaking moon. We went to
1:02:03
the moon a lot. We have to move a bunch of times. I think the last time we went to the moon, I
1:02:05
think the last thing humanity went to the moon was in,
1:02:07
like, nineteen seventy two or
1:02:10
seventy three.
1:02:11
Right? And so I'm pretty sure that's nineteen seventy two.
1:02:14
Nineteen seventy two. Right? So it's been fifty
1:02:16
years since we went to the moon. And, like, our
1:02:18
answer to
1:02:20
that is, Right? And, like, you know, Elon comes out and he's like, let's go to Mars and
1:02:22
everybody's like, wow. You know, I I guess we
1:02:24
could. Like, nobody
1:02:26
ever really
1:02:27
you know, nobody's you
1:02:29
know, like, in eighteen seventy two, it was, like,
1:02:31
the obvious next step is, of course, you go to Mars. Right?
1:02:33
because first, you go to the moon, then you go to Mars. Right? Then
1:02:35
you go to Jupiter. Right? then at some
1:02:37
point, you're going other places. Like, you're going other solar systems. Like, of course, this is like the arc that you
1:02:39
wanna be on. And, you know, we
1:02:42
in
1:02:42
our advanced era just decided,
1:02:46
yeah. like, let's
1:02:46
not do it. So, yeah, I think there's I think there's something to
1:02:48
that. You know, this this gets into
1:02:51
theories of social change
1:02:53
and societal forces, political forces,
1:02:56
religious forces. I mean, I've got I've
1:02:58
got theories for hours on kinda how we we got
1:03:00
here and what the problem is. But it it it
1:03:02
say very rapidly gets political. It's
1:03:04
hard to talk about kind of societal structure without
1:03:06
getting into politics. And so I try not to
1:03:08
get kind
1:03:08
of too kind of kinda kinda
1:03:11
fully say what I think on some of these topics. But
1:03:13
it is just
1:03:14
this, like, prevailing ethos. So, like, one way to
1:03:16
think about it is that if you think about socially, it's like
1:03:18
there basically, there are three errors in human access
1:03:20
there was like pre modern, which was basically like caves
1:03:22
all the way up through like kings. Right? And then
1:03:24
there was modern, which is like
1:03:27
science technology, democracy, you
1:03:30
know, capitalism and then in progress. And then now
1:03:32
you could argue, like, we're in this post modern phase
1:03:34
where we're just gonna, like, sit around and argue
1:03:36
all the time instead of actually doing anything.
1:03:39
you know
1:03:39
Yeah. May may
1:03:42
maybe, like, III would like
1:03:44
to believe that this is not you know, I would like to believe
1:03:46
Netflix and chill is not the terminal point of humanity.
1:03:48
Well,
1:03:49
I I think this this picture that
1:03:51
you're painting of a bunch of other alien
1:03:54
species getting so complacent
1:03:56
that they just decided never to venture out
1:03:58
or maybe they discovered VR or just went into their screens the way that a lot
1:04:00
of people are going today. But I I
1:04:03
wanna just quickly ask you
1:04:06
about what you're
1:04:08
excited about because you are someone
1:04:10
amongst there still is a
1:04:12
large cohort of people who are excited to
1:04:14
go to Mars or to take things
1:04:16
further than we ever have before. So I think that's exciting. And I
1:04:19
think many people take inspiration from that.
1:04:21
So within that
1:04:22
dimension, like, what
1:04:23
are you excited about?
1:04:26
We already talked about this like kind of next industrial revolution through
1:04:28
remote work, but are there other things where you're
1:04:30
like, wow, this is really game changing and
1:04:32
this is really exciting. And I'm glad that
1:04:34
us as a society smart
1:04:36
people working on this? Yeah. There's a
1:04:38
bunch of
1:04:38
things. I mean, there's a bunch of, like, specific technologies. And
1:04:40
so, I mean, the easy ones to bring
1:04:42
up kind of the cuff. It's it's, you know, AI. Like, it's
1:04:44
just like, it's amazing what's happening with machine learning
1:04:47
right now. And then biotech, like, the
1:04:49
genomic revolution is, like, a really big
1:04:51
deal. there's all kinds of questions. I mean, we're just saying
1:04:53
mind blowing concepts now on that front. I won't reveal it, but
1:04:55
we saw one the other day where, like, I literally had to
1:04:57
stop my tracks. My jaw hit the floor. I had to spend,
1:04:59
like, the next hour like, process what I
1:05:01
just seen. It's one
1:05:02
of those literally like this can change everything moments. And
1:05:04
so biotech's
1:05:05
going through a lot of really kind of profound
1:05:07
change and then crypto web three we've talked about
1:05:09
it, like, before as a firm, but, like, it's it's I think, know, very profound. And then there's
1:05:11
a bunch of others. So there's really three big ones,
1:05:12
but, like, there's a whole bunch of other spaces
1:05:15
that are like that. you know,
1:05:17
actually space, you know yeah. Elon, god bless him as, like, reinvigorated, you know, the the whole
1:05:19
idea of space. He's what did he do? The other
1:05:21
the other day he tweeted is, like, he said the other
1:05:23
day on Twitter. He said, they
1:05:26
have not yet discovered an upper limit on how many launches they can do
1:05:28
with the same rocket. And and it's just
1:05:30
like like rockets went from, like,
1:05:34
disposable, which is, like, one and done. Right? All the
1:05:36
way to, like, he's not sure. You know, he just doesn't know.
1:05:38
It's, like, they got fifty times, five hundred times, five
1:05:40
thousand times. you
1:05:40
know, we'll see. And so, like, you know, really, really revolutionary things happening
1:05:42
on that front, which is exciting. You know,
1:05:44
a couple of the big things
1:05:46
I would highlight. So
1:05:48
one
1:05:48
is just like the long term kind of implications the Internet, I think, are still
1:05:51
in the early stages. And and one that that's in
1:05:53
the background, but it's like really important relative to everything
1:05:55
we've been been discussing is
1:05:58
that, like, Most people in most of history have not had
1:06:00
access to the leading edge. Right? No. No. No. No.
1:06:02
No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No. No.
1:06:04
No. No. and
1:06:06
super curious and super ambitious and willing to work hard. Most of
1:06:08
them have not ever gotten to one of these,
1:06:10
you know, cities that I was talking about. Most of
1:06:13
them never worked in an event. feel
1:06:15
like most of them were doing subsistence farming or they were
1:06:17
working at other kinds of, you know, whatever. They were doing whatever
1:06:19
was the local thing to do. They never got an
1:06:22
opportunity. You know, most even you said just like
1:06:24
music. Most of the people who could have composed great
1:06:26
music never got the chance. They never
1:06:27
got the training, they never had access to the culture, they
1:06:29
never had the ability to produce music like they just could do it.
1:06:31
And you could say that's true everything from music
1:06:33
to like, every every field of human
1:06:35
activity. And, you know, the Internet really is, like,
1:06:37
the great equalizer leveler
1:06:40
opportunity machine. right,
1:06:42
for basically anybody in the planet is curious with Internet connection to
1:06:44
be able to learn and explore and start to
1:06:46
create and start to join and meet
1:06:49
like minded people. So there's
1:06:50
a collective effect, like, like, in my in my optimistic moments, I kinda
1:06:53
think of
1:06:53
it as, like, humanity as sort of
1:06:55
a group organism, like, like, it's just like
1:06:57
the through global society. like,
1:07:00
is is actually just waking up for the first time as a consequence of
1:07:02
being, you know, all connected online. And and,
1:07:04
of course, there's there's good aspects of that and
1:07:06
maybe bad aspects of that, but it it
1:07:08
it is sort of the this this this final thing
1:07:10
is happening. And so, you know, hopefully, we will discover that there are
1:07:13
I don't know what the number is, but ten or a hundred
1:07:15
or a thousand or ten thousand or a million
1:07:17
more people around the world. who
1:07:19
could be doing a really original creative work, who just never had the chance
1:07:21
in prior generations, and all of a sudden, you know, you can imagine
1:07:23
things progressing much faster as a
1:07:25
result of that.
1:07:27
Yeah. I
1:07:27
I think even another aspect of that is
1:07:30
more people who are coming online, you
1:07:32
are having them get access to remote work and, you
1:07:34
know, another second order effect of that is
1:07:36
that remote tends to be
1:07:38
more asynchronous. So certain people
1:07:40
don't always benefit from the nine to five as
1:07:42
an example. And, you know, I saw this
1:07:44
infographic today of just like the
1:07:46
schedules of all these
1:07:48
luminaries from back in the day who had created
1:07:50
wonderful things, and they were all over the
1:07:52
place. So that's just, again, one
1:07:54
example of as more people get access to this information, as more people
1:07:56
have access to different types of schedules or different types
1:07:58
of companies or people
1:07:59
facilitated through the Internet,
1:08:00
I think we're gonna be really surprised. by
1:08:03
what comes
1:08:04
out of that. That's right. One
1:08:06
thing
1:08:06
I really want to ask you about is
1:08:09
how society values certain things.
1:08:11
And I've heard you talk about this to an extent
1:08:14
before, but society will fluctuate throughout
1:08:16
time and different people
1:08:18
within that society will value different The society as a whole to
1:08:20
find virtue in in certain things at
1:08:22
certain times. So for example, you mentioned
1:08:24
before, in history, maybe
1:08:26
entrepreneurs were valued more. So
1:08:29
as people built things, society rewarded that type of achievement. How
1:08:31
do you view that changing today? What do you think society
1:08:33
is valuing today?
1:08:36
And perhaps If you're willing
1:08:38
to share, what do you wish society valued more? So basically, there's this theory, the sky james articulated
1:08:40
this theory back
1:08:41
in the actually the nineteen forties
1:08:43
that that I think applies he
1:08:46
called at the time he called the managerial revolution. So basically, what
1:08:49
he what he said was there basically
1:08:51
two phases to capitalism. There was basically
1:08:53
what he called the the phase
1:08:55
of bourgeois capitalism. which
1:08:56
is sort of famously the sort of phase that, you know, the Robert Berens
1:08:58
and then, you know, the phase that the communist hated and, like, I'll I'll I'll, you know, the the rise to
1:09:02
kind of my culture. And and this is the, you know, the capitalism era was the
1:09:04
era of, like, Henry Ford and top of Edison
1:09:07
and then the Carnegie. JPMorgan and and all of
1:09:09
these kind of comic Henry Ford, all
1:09:11
these kind of business builders. who
1:09:13
built the companies off often named the after
1:09:16
through kinda sheer force and
1:09:18
personal kinda animal agatism at force.
1:09:23
And then he said, basically, there's an evolution that takes place, and
1:09:25
it kinda started in the thirties and forties and
1:09:27
extensive to our our period. And and
1:09:29
he called out the era
1:09:31
of managerial capitalism. or just more generally, you
1:09:33
call it managerialism. And basically, the the idea there is it's
1:09:35
it's sort of the second phase of capitalism.
1:09:37
And it's basically the phase
1:09:40
at which you
1:09:42
can no longer just have a guy. You know, you can no longer just have a Henry Ford or whatever who's just got a
1:09:45
car
1:09:46
company called Ford
1:09:48
just like orders everybody around and
1:09:51
tells everybody what to do. Like, that
1:09:53
basically modern business, modern technology. Right? Modern society. It's
1:09:55
too complicated for that. And
1:09:59
so you're not
1:09:59
basically, the the model of capitalism goes
1:10:02
from basically sole proprietorship in the bourgeois
1:10:04
era. Right? To basically what you
1:10:06
consider to be like the modern,
1:10:08
multi correctional, you know, Delaware, C Corporation with,
1:10:10
like, a Board of Directors and an executive management team and a CEO and
1:10:14
the CEO is probably the founder, in fact, the CEO was probably trained in
1:10:16
a business school, like and Richard
1:10:18
never went to business school. Right? But,
1:10:22
you know, Jim Farley was running kind of check, but I'm sure Jim
1:10:24
Farley was running forth today. I'm sure that he did. Right? So,
1:10:26
like, all of a sudden, it's basically like I said, you
1:10:29
had the rise of sort of the managerial class the
1:10:31
rise of basically these people who are sort of
1:10:33
technocratic experts who would never start
1:10:35
their own thing. Like, they would
1:10:37
never invent the car or they
1:10:39
would never start the power company, but they
1:10:41
are necessary to run the large scale industrial organization that is
1:10:43
like a modern, you know, giant power company
1:10:45
or a modern giant electrical
1:10:47
grid or a modern telephone
1:10:49
network or a modern giant. Whatever. You name it. Chip company, like, whatever it is. And it's a little bit
1:10:51
of you think about it. It's a little bit of
1:10:54
like, basically, founder led
1:10:56
companies to basically pro
1:10:58
quote unquote professional CEO led companies. You know, it's basically top down kind of dictatorial management
1:11:01
versus kind of bottoms
1:11:03
up consensus management. It's
1:11:06
basically the principal running the
1:11:08
company. Right? With his like, often with his name
1:11:10
on the line and his entire network is
1:11:12
in the company and, like, his name is
1:11:14
on the line. to what, you know, you refer to as the principal agent problem,
1:11:16
which is, okay, these companies are run by people,
1:11:18
but, like, you know, are these people really
1:11:21
gonna be with the company for
1:11:23
the next thirty years? you
1:11:24
know, get paid annually or their incentives
1:11:26
more annual as opposed to long term. And what Bernie basically said was,
1:11:28
this is an inevitable process
1:11:30
to go from bourgeois capitalism to
1:11:33
a drill capillos if it's inevitable because of scale and complexity. Right? Henry Ford today could not
1:11:35
run for a company like it's too big and
1:11:37
complicated. You need you need a different
1:11:40
skill set. But
1:11:42
he said,
1:11:43
look, it it it is a very different it is
1:11:45
a very big social cultural change. It's a
1:11:47
change basically from valuing
1:11:50
sort of individual aggression individual individual individual accomplishment,
1:11:52
individual force of will, right,
1:11:55
to a much
1:11:56
more
1:11:59
collective right, way of operating.
1:11:59
Right? Groups groups operating at sort
1:12:02
of consensus collaborative form. You know,
1:12:04
people haven't come to
1:12:06
agreement on things, committees, bureaucracy.
1:12:09
Right?
1:12:09
And he basically said it sort of, you
1:12:11
know, stage stage one, stage two. The way I view kind of what we do, like, in what we do is
1:12:13
we basically are we're
1:12:15
basically the throwback. So the
1:12:19
startups that we fund are being funded we fund startups that
1:12:21
fit that old model. Like, we're trying to find
1:12:23
the
1:12:23
next generation of Henry
1:12:24
Ford's. Right? in in your carnegies
1:12:27
and so forth and, you know, live in Stanford, you
1:12:29
know, the railroad guy who, you know, ultimately founded
1:12:31
Stanford University was a was a Robert Baron in
1:12:33
the eighteen eighties. we're trying to basically go find
1:12:35
those sort of modern bourgeois capitalists who are
1:12:37
kind of throwbacks to the old model. Why
1:12:39
are we doing that? because, like, that's the
1:12:41
only way to do something new. Right? That that was the model for
1:12:43
doing new things. All these things used to be new.
1:12:46
Like, if you wanna do anything new today, that
1:12:48
is the model. You do need to
1:12:50
bring back this model of booth like capitalism. And
1:12:52
then, basically, we work with our companies to
1:12:54
try to basically keep them from basically just turning into sort of
1:12:58
this workflow managerial capitalism kind of outcome on the other side becoming just
1:13:00
like every other big company, which, by the way, many
1:13:02
of them kind of follow that path,
1:13:04
and many of them just become like big companies just
1:13:06
like all the rest of by That's that's disgusting. So
1:13:09
I was gonna quickly ask, how do you
1:13:11
stop that? I don't know that you
1:13:13
do. Well, so Bernard would
1:13:15
say that you don't. like, Bernum would say that it's
1:13:17
inevitable process. Basically, he's like, what what Bernum would say if he was talking to me, he's
1:13:19
like, okay. Look, smart guy. You're
1:13:22
you're just gonna keep repeating history. start you're a Ford kind of clerical
1:13:24
in charge of them. At some point, they're gonna reach
1:13:26
a level of scale and complexity where,
1:13:29
like, one guy just
1:13:31
can't, like, run everything. and
1:13:32
you're gonna need to bring in
1:13:34
the the experts. Right? The experts, the technically trained experts, the managers. Right?
1:13:36
The people with white business school degrees,
1:13:38
the people
1:13:39
with white, you know, the
1:13:42
people who have grown up basically getting trained
1:13:44
to run large scale systems. And,
1:13:46
you know, these companies are naturally
1:13:49
going to evolve kind of in
1:13:51
that direction. I go close. Okay. So
1:13:51
there's there's all that. And then and then and then Bernad made the the following point is
1:13:54
he said, look, the the transition from bourgeois capitalism
1:13:56
to managerial
1:13:58
capitalism. is not just happening in business. It's
1:13:59
happening everywhere else in life. It's happening
1:14:02
everywhere else in our society. So for
1:14:04
example,
1:14:04
it's also happening in the
1:14:06
government. Right? And so the government
1:14:08
Right? Just take the federal federal government as
1:14:10
an example. Like, there's basically been three different forms of federal government in the last
1:14:13
hundred years. There was sort of
1:14:15
the pre FDR era which
1:14:17
the federal government was just basically
1:14:19
small and basically not very irrelevant. And then there was the FDR era he basically
1:14:21
appointed
1:14:24
himself king like, he
1:14:26
made himself basically essentially, the headroom for the government or something. And then he basically just, like, told everybody what to do, and that was, you know, the
1:14:28
new deal and world war two and,
1:14:30
like, all this stuff and serving four terms
1:14:32
and and
1:14:34
certain this this model of Imperial presidency. And then there's sort
1:14:37
of the model of the government we have today, which
1:14:39
is it's just basically democrats. It's
1:14:41
politicians and democrats as far as the I can
1:14:43
see. oh, that's not how these this that's not structurally how anything is supposed to be running
1:14:45
anywhere. You're not supposed to just have a guy who
1:14:47
just tells people what
1:14:49
to do. You're supposed to have process. You're supposed to have managers. You're supposed to
1:14:52
have experts. Right? You're supposed to have professors. You're
1:14:54
supposed to have think tanks. You're supposed to have
1:14:56
the press. You're everybody's
1:14:58
supposed to weigh in. You're have this big conversation. Everybody's supposed to get
1:15:00
along. Everything has to be negotiated. Right? And
1:15:02
so basically, what
1:15:03
Bernie would say is, like, the
1:15:05
entire society transition from
1:15:07
kind of individualistic to
1:15:09
basically collectivist or let's call it bureaucratic or is he would
1:15:11
call it managerial. And, like, basically, that
1:15:12
that that's happened to
1:15:15
basically all of society. Right?
1:15:17
And
1:15:17
and this is kind of the experience that you have is that if you think about this
1:15:19
as an individual, this is a kind of experience that you have because, like, everywhere you go at life now, you're,
1:15:22
like, dealing with synbureaucracy.
1:15:24
Right? It's
1:15:26
like you've got a problem with your cable Internet
1:15:28
hookup. You know? You're gonna talk to the cable company's
1:15:30
bureaucracy. Like, don't you can't call the CEO of
1:15:32
the cable company. And by the way, if you did,
1:15:34
you can't do Like, it's gonna be somebody deep at the balls of the organization that's gonna get
1:15:36
your Internet to work. You know, you need to get
1:15:38
your driver's license renewed. You can't go
1:15:42
to DMV. Right? You it's just, like, everything you you
1:15:44
know, you go out to eat and, like, the place where you
1:15:46
go out to eat is, like, one of three thousand, right,
1:15:48
restaurants that that company manages
1:15:50
that are all identical. Right? everything
1:15:52
you do, mass manufacturing, everything you buy,
1:15:54
right, has been manufactured by
1:15:57
a company's manufacturer
1:16:00
for scale. all of your entertainment, you watch a movie.
1:16:02
It's the same movie, a hundred million other people are watching. You know, it's like, you know, the actors don't come to your house anymore and,
1:16:04
like, after the play,
1:16:06
you're watching a mass produce production.
1:16:09
built by this tribe, you're obviously in Hollywood. So and then
1:16:11
you're like, wow, this movie. It seems like I've seen this movie like eighty
1:16:16
times before. know, why aren't they making, like, more creative movies?
1:16:18
Well, it's because it's movie making is a machine now. Right? It's a, you know, movies cost three hundred million dollars and there's,
1:16:22
like, a whole process and a whole bureaucracy. making And what Bernie
1:16:24
would say basically is the whole country, the
1:16:26
whole society, has evolved into this kind
1:16:31
of bureaucratic managerialism. And,
1:16:32
you know, in other words with that, it's
1:16:34
just stagnation. Right? It's just, like, the whole the whole system is on autopilot. Like, the the whole society
1:16:36
is not autopilot.
1:16:39
The government's autopilot. it's
1:16:40
all an autopilot. And and then, you know, every once in
1:16:42
a while, you get a new one must. Right? Or, you know, you get
1:16:44
the kinds of founders that we deal with that they
1:16:46
they kind of step forward. They say, well, actually,
1:16:49
I have a different idea. And then
1:16:51
they have the temerity to, you know, build a new piece of software or to start a new kind of company or to
1:16:54
propose some other creative
1:16:56
idea. So anyway,
1:16:58
you you see kind of all these kind of, you
1:17:00
know, perturbations in the force where these kind of creative individuals pop up.
1:17:02
And then, you know, like society freaks out and everybody's got
1:17:06
opinion on the whole thing. But if there is to be progress, right?
1:17:08
If there if there are to be new ideas
1:17:10
in the world, new concepts, new forms of art,
1:17:12
new forms of culture, new ideas, by
1:17:14
the way, new forms of politics, Right? By
1:17:16
the way, new ways to think about how you raise
1:17:18
kids, you know, basically anything new. It's gonna come from some unusual individual basically stepping
1:17:21
up and saying, I think that
1:17:23
the system is wrong. so
1:17:26
that's kind of the fundamental battle that we'll probably spend the rest of our, you know, probably spend the next years of our civilization to balance
1:17:31
between. Yeah.
1:17:32
I mean, I can see the example with a company.
1:17:34
Right? These companies stagnate, and then new startups come and replace them. And that's that's
1:17:37
something we've seen over
1:17:39
and over and over. also happens
1:17:41
at the individual level. Right? You see celebrities become popular and then they
1:17:43
just go and do the same movie over and over and
1:17:45
over because they're trying to
1:17:47
retain, they're following, and
1:17:49
then people move on to the next new thing. Something that we talked to Balaji about
1:17:51
because he came on to talk about the network state is the need for
1:17:53
this kind of revolution or
1:17:56
innovation at the
1:17:58
government level or the state level. I'm just interested
1:18:00
in your perspective on that because we
1:18:02
have seen it at the individual
1:18:05
level, the company level, but we are seeing
1:18:07
stagnation at that higher government level. So
1:18:09
are you also thinking that
1:18:11
we're going to see some
1:18:13
of these smaller nations or
1:18:15
completely new nations come up
1:18:16
as we see the stagnation in the
1:18:18
Western world? Yeah. So it was this great word that gets used,
1:18:20
well there's this great word against used
1:18:22
reforms reform. Right? Referring
1:18:24
one of those words or my my ears always per company when
1:18:26
I see the word because I kinda know the game is
1:18:28
being played. And so I'll give you
1:18:30
an example. I have all these are very
1:18:32
into this thing. They call it education reform. Right? And so
1:18:35
they're, you know, philanthropists,
1:18:35
you know, they've been successful. They've got money
1:18:37
and they're they've got
1:18:40
a foundation and they wanna, like, make the world better. And
1:18:42
so what do they do? They look around? They're like, what are the big problems in the world? And they in other ways, a lot
1:18:44
of times, they
1:18:44
end up looking at public education.
1:18:47
And they're just like, wow. like,
1:18:49
public education is like this huge force in our society and all these kids, you know, legally, you, like,
1:18:51
have to send your kids to basically, you know, most most people legally are
1:18:53
required to send their kids to a
1:18:56
public school. And
1:18:59
it's like, wow. Like, the outcome seem like really bad. And we
1:19:01
keep injecting more money into these schools and get
1:19:03
the results don't get better. And then,
1:19:05
by the way, there's all these problems and there's all
1:19:07
like, you know, child abuse, you know, that takes place, and there's all this, you know,
1:19:10
these, like, features sexual animals, and then
1:19:12
there's all these controversies over the curriculum
1:19:14
and, like, her kids being taught the right
1:19:16
things. And it's
1:19:18
just like, wow, this thing just seems
1:19:20
like a giant mess. And so therefore, we need education
1:19:22
reform. Right? We need to, like, go
1:19:23
into the bureaus see. And
1:19:25
we knew, like, re you know, we need to retool it. We
1:19:27
need to we need to improve it. We need to make it better. We need to reengineer you know, the
1:19:30
story
1:19:30
is kind of always
1:19:32
which is they go in, they do all this work,
1:19:34
they spend all this money, and then basically nothing changes. There have been many famous cases of this. I won't pick
1:19:39
on people I will say the Gates Foundation has done a lot of work in this area. And they actually
1:19:41
to their credit, they actually came out with a public
1:19:43
report about four years ago
1:19:45
where
1:19:46
they did this retrospective They did a
1:19:48
retrospective of the last fifty years of education reform efforts and all of
1:19:50
the different work that has been done, all the different ideas that
1:19:53
people have had to,
1:19:55
like, make schools better. and
1:19:56
they did this big report. And as the result of
1:19:58
the report was nothing has worked, like, nothing has worked for fifty years. Like, there has not been a new
1:19:59
idea in fifty years that has
1:20:01
been, like, tried in large scale education that's
1:20:03
had any impact at all. basically,
1:20:06
the
1:20:06
whole the whole the whole effort has just been,
1:20:08
like, a complete a complete zero. And
1:20:10
I'd become convinced that's basically universally
1:20:12
true, which is basically things don't get
1:20:15
reformed if you have a problem with XYZ existing institution existing system for
1:20:17
the reasons we discussed earlier. Like, it's
1:20:18
it's just it's not going to
1:20:21
get better. It's not going to get reformed. you can go
1:20:23
spend an arbitrary amount of time and money trying to reform it and improve it.
1:20:25
It's not gonna happen. Why is it not gonna
1:20:27
happen? Because it doesn't, you know,
1:20:28
it doesn't have to happen. The people running
1:20:31
it don't want it to happen. this
1:20:33
this so called principle agent problem, the people in charge aren't actually responsible for
1:20:35
it. You know, there's the incentives problem, especially when people with
1:20:37
large
1:20:37
geographies, people are much more focused
1:20:40
on that fired
1:20:42
than they are on improving anything. And so they'll
1:20:44
basically well, a lot of presidents encountered this.
1:20:46
Right? A lot of presidents come into the White
1:20:49
House, right, for the time, and they're like, wow,
1:20:51
I have all these ideas on how to
1:20:53
make the government better. They issue all these orders, and then the bureaucracy just
1:20:55
ignores them. Right? You know, because -- Right. because
1:20:58
the federal the government employees, like, what do they know? They know
1:21:00
the president's come and go. Right? Like, the president's
1:21:02
gonna be gone in four years or
1:21:04
worst case, eight years. Nobody's
1:21:06
even remember what that guy tried to do. Right? And so all they they just, like, they just wait out Right? And
1:21:09
then and then
1:21:12
nothing changes. And so, anyway, so,
1:21:14
like, by default, it's it's like stagnation as far as the I can see. I've reached the conclusion. This is kinda how I spend my
1:21:16
time. This is why I I continue to do what I
1:21:18
do, and I I think I'll basically do it forever.
1:21:23
is I really no longer believe in the concept of institutional reform. Like,
1:21:25
I I think fundamentally, it it doesn't happen
1:21:27
or it's so rare as to, like,
1:21:29
be be basically something that you
1:21:31
can't ever count on. I think, basically, progress happens by
1:21:33
starting new things. Right? And so, like, if if you wanna like reform the school system, the thing you
1:21:35
need to do is build new
1:21:38
schools. Right? From scratch, like, built the correct way. If you want
1:21:41
to have, like, a new car company, if you're Elon Musk and
1:21:43
you wanna have a car company, you don't spend your
1:21:45
time trying to go get whatever existing car company
1:21:47
to build a better car, You just,
1:21:49
like, start the car company. And then our friend biology extends it one step further,
1:21:51
which is, like, okay. You don't try to reform the country. You
1:21:54
just start into country.
1:21:56
Now, The
1:21:57
obvious problem, the challenge right, is there was an era of human history where people were starting to countries like all the countries that we have today
1:21:59
countries at some
1:22:02
point that somebody started. Unfortunately,
1:22:05
in the modern era, the real estate of the planet's kind of been divided up. And the
1:22:07
world's not really that amenable to,
1:22:11
you know, changing which
1:22:13
country controls their territory through conquest. Like, that's really kind of frowned upon these days. And so, you know, starting a
1:22:15
new country in the modern era is probably not
1:22:18
a process of life going and taking out
1:22:20
a bunch land
1:22:23
and then, like, clearing yourself a new kingdom and then, you know, it's lending a bunch of new
1:22:26
whatever your own legal system. It's probably something else. And, of
1:22:28
course, Balaji's
1:22:30
book kind of explores the something else. with with this idea of the network state.
1:22:32
So, you know, as usual, I analogy pushes it to a
1:22:34
to a level that, you know, probably beyond where I would.
1:22:36
But I think, look,
1:22:38
I think everything he says everything
1:22:40
like his situational analysis, I think, for sure, is a hundred percent correct. Overall,
1:22:42
you could say, like, this is a very depressing analysis in the state
1:22:45
of affairs, and, like, this
1:22:47
basically means the world segued on how much
1:22:49
is gonna happen. I I and there's a lot to that. There is
1:22:51
always this concept of arbitrage, which is like if
1:22:53
most of the world is not
1:22:55
doing new things, then the person
1:22:57
who can do something new has an outsized opportunity. Right? And and this is the thing that kinda gets me up
1:22:59
every morning, which is like, okay, because
1:23:02
most of the world will not
1:23:04
change, because
1:23:06
most existing companies won't change, because most
1:23:08
existing geographies won't change, systems won't
1:23:10
change, people won't change. Because of that,
1:23:12
the person who has the genuinely new idea who
1:23:14
is willing to put themselves in line to build something
1:23:16
new, because they're really big outsized opportunity. because if they succeed, right,
1:23:19
they'll they'll just they'll get all the benefit.
1:23:22
They'll get all the gains. Like, they'll all of a sudden be the person who's,
1:23:24
like, building or running everything. Right? Because it's not
1:23:26
going to be the status
1:23:27
quo that's going to
1:23:29
adapt. Right? So so therefore, like, all opportunity in the world is still
1:23:32
off, but basically available to all of these,
1:23:34
you know, kind of disruptive entrants. And so,
1:23:36
basically, the more
1:23:38
the sort of older world stagnant the bigger the opportunity on entrepreneurial side.
1:23:40
And honestly, like, I think that's what keeps
1:23:42
us in business basically in perpetuity is
1:23:44
the model of kind of
1:23:46
entrepreneurial capitalism or or more generalized
1:23:49
five different neural isn't that, you know, having new
1:23:51
ideas and putting them in the world. Like, that that model is the only source of Yeah. I love the way
1:23:53
you put it of not counting on
1:23:55
the status quo to innovate. I
1:23:59
think that by nature allows the really
1:24:01
intelligent, the really creative, the really
1:24:03
innovative people to have an opportunity.
1:24:05
Right? Because of these companies as
1:24:07
an example, if smart companies
1:24:09
that became big, stayed smart, then we wouldn't have opportunity for new smart
1:24:11
people to innovate. Right? There
1:24:14
there wouldn't be room. So
1:24:17
think that's a wonderful place to end because
1:24:19
I think this you've so opportunities people get involved and for
1:24:22
new businesses to be
1:24:24
built I think most
1:24:26
importantly, as we started off this conversation, you've really highlighted why it's still important. As the ladies sounds,
1:24:28
it's still so important
1:24:30
for us to build because
1:24:34
things do stagnate, and there is so much more
1:24:36
to be built. So thanks, Mark.
1:24:37
Thank you so much for talking to us
1:24:39
today. Thank you, stuff.
1:24:41
Thanks for
1:24:41
listening to the a sixty z podcast. If you like this
1:24:43
episode, don't forget to subscribe, leave a review,
1:24:44
or
1:24:48
tell run. We also recently launched on YouTube at
1:24:50
youtube dot com slash A6CZ underscore video, where
1:24:52
you'll find exclusive video content.
1:24:54
We'll see
1:24:55
you next time.
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More