Podchaser Logo
Home
Inside Google’s big AI shuffle — and how it plans to stay competitive, with Google DeepMind CEO Demis Hassabis

Inside Google’s big AI shuffle — and how it plans to stay competitive, with Google DeepMind CEO Demis Hassabis

Released Monday, 10th July 2023
Good episode? Give it some love!
Inside Google’s big AI shuffle — and how it plans to stay competitive, with Google DeepMind CEO Demis Hassabis

Inside Google’s big AI shuffle — and how it plans to stay competitive, with Google DeepMind CEO Demis Hassabis

Inside Google’s big AI shuffle — and how it plans to stay competitive, with Google DeepMind CEO Demis Hassabis

Inside Google’s big AI shuffle — and how it plans to stay competitive, with Google DeepMind CEO Demis Hassabis

Monday, 10th July 2023
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:02

Hello and welcome to Decoder. I'm Nilay Patel, Editor-in-Chief

0:04

of The Verge, and Decoder is my show about

0:07

big ideas and other problems. Today,

0:09

I'm talking to Demis Hassabis, the CEO

0:12

of Google DeepMind. That's the newly

0:14

created division of Google responsible for

0:16

AI efforts across the company. Google

0:19

DeepMind is the result of what you might call an internal

0:22

merger. Google acquired Demis'

0:24

DeepMind startup in 2014 and

0:27

ran it as a separate company inside

0:29

of its parent organization, Alphabet, while

0:31

Google itself had an AI team called Google

0:34

Brain. Google's been showing off AI

0:36

demos from both groups for years now, but

0:38

with the explosion of chat GPT and

0:40

a renewed threat from Microsoft in Search,

0:43

Google and Alphabet CEO Sundar Pichai

0:45

made the decision to bring DeepMind into

0:47

Google itself earlier this year, creating

0:50

Google DeepMind. What's interesting is

0:52

that Google Brain and DeepMind were not necessarily

0:55

compatible or even focused on

0:57

the same things. DeepMind was famous

0:59

for applying AI to things like games

1:02

and protein folding simulations. The

1:05

AI that beat world champions at Go, the

1:07

ancient board game? Yeah, that was DeepMind's

1:09

AlphaGo. Meanwhile, Google Brain

1:11

was more focused on what's come to be the familiar

1:13

generative AI toolset. Large

1:15

language models for chatbots, editing features

1:18

in Google Photos, and so on. So

1:20

this was a big structure decision with a goal of being

1:22

more competitive and faster to market with

1:24

AI products, but Demis had to manage

1:27

a culture clash between two very different

1:29

organizations.

1:30

And the competition isn't just OpenAI and

1:32

Microsoft.

1:33

You might have seen a memo from a Google engineer

1:36

floating around the web recently claiming that Google

1:38

has quote no moat in AI

1:40

because open source models running on commodity

1:42

hardware are rapidly evolving and

1:45

catching up to the tools run by the tech giants.

1:48

I asked Demis about that memo and he confirmed

1:50

that it was real, but he said it was part of Google's

1:52

debate culture and that he disagreed with it.

1:54

And we talked about his other ideas about where Google's

1:57

competitive advantages might come into play.

1:59

Of course, we also talked about AI risk

2:02

and especially artificial general intelligence.

2:04

Demis is not shy that his goal is building

2:06

an AGI. And we talked about what risks

2:09

and regulations there should be and on what

2:11

timeline. Demis recently signed

2:13

on to a 22-word statement

2:15

about AI risk with OpenAI, Sam

2:17

Altman, and others that simply reads, quote,

2:20

mitigating the risk of extinction from AI

2:22

should be a global priority alongside

2:24

other societal-scale risks such as pandemics

2:27

and nuclear war.

2:29

So that's pretty chill. But is that

2:31

the real risk right now or just distraction

2:33

from other, more tangible problems like

2:36

AI replacing a bunch of labor in various

2:38

creative industries?

2:39

We also talked about the new kinds of labor that AI

2:41

is creating. Armies of low-paid

2:44

taskers classifying training data

2:46

in countries like Kenya and India.

2:48

We just did a big feature on these taskers, which

2:50

we'll link to in the show notes. I wanted

2:52

to know if Demis thought these jobs were here to stay

2:55

or just a temporary side effect of

2:57

the AI boom. I got

2:59

to say, this one really hits all the decoder high

3:01

points. There's the big idea of AI. There's

3:04

all the problems that come with it, an infinite

3:06

array of complicated decisions to be made, and

3:08

of course, a gigantic org chart

3:10

change in the middle of it all.

3:12

Demis and I got pretty into the weeds and I still don't think

3:14

we covered it all, so we'll have to have him back soon.

3:16

A one-hour AI metadata

3:19

show, I promise you. We're going to make this happen. Okay,

3:22

Demis Asabas, CEO of Google DeepMind.

3:24

Here

3:25

we go.

3:37

Demis Asabas, you are the CEO

3:39

of Google DeepMind. Welcome to Decoder. Thanks

3:42

for having me. I am very excited to talk to

3:44

you. I don't think we have ever had

3:46

a more perfect decoder guest.

3:48

There's a big idea in

3:51

AI. It comes with a bunch of challenges and

3:53

problems. With

3:55

you in particular, there's a

3:58

gigantic org chart. move

4:00

and a set of high-stakes decisions to be made,

4:02

I am thrilled that you are here.

4:06

Glad to be here. Let's start with Google DeepMind

4:08

itself. Google DeepMind is a new

4:11

part of Google that is constructed of two

4:13

existing parts of Google. There

4:15

was Google Brain, which was the AI

4:17

team we were familiar with as we covered Google. That

4:20

was run by Jeff Dean. And there was DeepMind,

4:23

which was your company that you founded. You

4:25

sold it to Alphabet in 2014. You

4:27

were outside of Google. It was run

4:29

as a separate

4:29

company, instead of holding company

4:32

Alphabet structure until just now. Start

4:35

at the very beginning. Why were DeepMind and Google Brain

4:38

separate to begin with? Well, as you mentioned,

4:40

we started DeepMind actually back in 2010, a long

4:42

time ago now, especially

4:45

in the age of AI. So that's prehistory.

4:50

Myself and the co-founders, we realized

4:53

coming from academia and seeing what was going on there,

4:55

things like deep learning had just been invented. We

4:58

were big proponents of reinforcement learning. We

5:00

could see GPUs and other hardware was

5:02

coming online. A lot of great progress

5:05

could be made with a focused effort on

5:07

general learning systems and also taking

5:09

some ideas from neuroscience and how the

5:12

brain works. And so we put all those ingredients

5:14

together back in 2010. We

5:16

had this thesis, we'd make fast progress, and that's what

5:18

happened with our initial game

5:20

systems. And then we decided in 2014

5:23

to join forces with Google at the time,

5:26

because we could see that a lot more compute

5:28

was going to be needed. And obviously, Google has

5:31

the most computers and had the most computers in the world.

5:34

And so that was the obvious

5:36

home for us to be able to focus

5:38

on pushing the research

5:39

as fast as possible.

5:42

So you were acquired by Google and then somewhere

5:44

along the way, Google reoriented

5:47

itself. They turned into Alphabet and

5:49

Google became a division of Alphabet. There's other

5:52

divisions of Alphabet and DeepMind

5:54

was out of it. And that's just the part

5:56

I want to focus on

5:57

right here at the beginning. There

6:00

was what Google was doing with Google Brain,

6:03

which is a lot of LLM research. I recall six

6:05

years ago, Google was showing off LLMs at Google

6:08

IO. But DeepMind was focused

6:10

on winning game, AlphaGo and

6:13

protein folding, a very different kind of AI

6:15

research fully outside of Google. Why

6:18

was that outside of Google? Why was that an alphabet

6:20

proper?

6:20

That was part of the agreement as we were

6:23

required, was that we would pursue

6:25

pushing forward research into

6:28

general AI or sometimes called AGI.

6:30

The system that out of the box can

6:33

operate across a wide range

6:35

of cognitive tasks and

6:38

basically has all the cognitive capabilities

6:41

that humans have. And also using

6:43

AI to accelerate scientific

6:45

discovery. That's one of my personal passions.

6:48

That explains projects like AlphaFold that I'm

6:50

sure we're going to get back to. But also from

6:52

the start of DeepMind and actually prior to even DeepMind

6:55

starting, I believe that games

6:57

was a perfect sort of testing or proving ground

7:00

for developing AI algorithms

7:02

efficiently. Quickly, you can

7:04

generate a lot of data and the objective functions

7:06

are very clear, obviously winning games

7:09

or maximizing the score. So

7:11

there were a lot of reasons to use games in

7:13

the early days of AI research. And

7:15

that was a big part of why we were so successful and

7:17

why we were able to advance so quickly with things like

7:20

AlphaGo, the program that beat

7:22

the world champion at the ancient game

7:24

of Go. Those were really important

7:26

proof points for the whole field really

7:29

that these sort of general learning techniques

7:31

would work. And of course, we've done a lot of

7:33

work on deep learning and neural networks as well.

7:36

And our specialty, I suppose, was combining

7:38

that with reinforcement learning to

7:40

allow these systems to actively

7:42

solve problems

7:43

and make plans and do things

7:45

like win games. And in terms

7:47

of the differences, we always

7:50

had that sort of remit to push the research agenda

7:52

and push things advanced science. And that

7:54

was very much the focus we were given

7:56

and very much the focus that I wanted to have.

8:00

that the internal Google AI teams like Google

8:02

Brain, they had slightly different remits

8:04

and were a bit closer to product, and

8:06

obviously to the rest of Google, and infusing

8:08

Google with amazing AI technology.

8:11

And we also had an applied division that was introducing

8:14

DeepMind technology into Google products too,

8:16

but the cultures were quite different and the remits

8:19

were quite different.

8:20

So from the outside, the timeline

8:23

kind of looks like this. Everyone's

8:25

been working on this for ages. We've all been talking about

8:27

it for ages.

8:29

It is a topic of conversation for

8:31

a bunch of nerdy

8:33

journalists like me, a bunch of researchers,

8:35

we talk about it in the corner at Google events. Then

8:38

chat GPT is released, not even

8:40

as a product. I don't even think Sam would call it a great

8:42

product when it was released, but it was just released

8:44

and people could use it and everyone freaked out, and

8:47

Microsoft releases Bing based on chat GPT,

8:50

and the world goes upside down and

8:52

Google reacts by merging DeepMind

8:54

and Google Brain.

8:56

That's what it looks like from the outside. Is

8:58

that what it felt like from the inside?

9:00

That timeline is correct, but

9:03

it's not these sort of direct consequences.

9:05

It's more indirect in a sense. So Google

9:07

and Alphabet have always run like this. They

9:10

let many flowers bloom, right? And

9:12

I think that's always been the

9:14

way that, even from Larry and Sergey

9:16

from the beginning, set up Google and it's served them very

9:18

well. And it's allowed them to organically

9:21

create incredible things and become the amazing

9:23

company that it is today. On the research

9:25

side, I think it's very compatible

9:28

with doing research, which is another reason we

9:30

chose Google as our partners back in 2014. I

9:33

felt they really understood what fundamental

9:36

and blue sky research was, ambitious research

9:38

was, and they were gonna facilitate us

9:40

being, and enable us to be super

9:43

ambitious with our research. And you've seen the

9:45

results of that right by any measure, alpha

9:47

go alpha fold, but more than 20 nature

9:50

and science papers and so on. All the normal

9:52

metrics one would use for

9:54

really cutting edge, delivering amazing

9:57

cutting edge research, we were able

9:59

to do.

9:59

But in a way, what chat GPT

10:02

and the large models and the public reaction

10:04

to that confirmed is that AI

10:07

has entered a new era. And

10:09

by the way, those of you, it was a little bit surprising

10:12

for all of us at the coal face, including

10:14

I think open AI, how viral that

10:16

went, because we all had

10:18

us and some other startups

10:21

like Anthropic and OpenAI. We all had these

10:24

large language models. They're all roughly the same

10:27

capabilities. And it was surprising,

10:29

not so much what the technology was, because we all

10:31

understood that, but the public's appetite

10:33

for that, I would say. And obviously

10:35

the buzz that generated. And I think that's

10:37

indicative of something we've all been feeling for the

10:40

last, I would say, two, three years, which is these

10:42

systems are reaching a level of maturity

10:45

now and sophistication, where

10:48

it can come, really come out of the

10:50

research phase and the lab

10:53

and go into powering incredible

10:55

next generation products and experiences

10:58

and also breakthroughs, things

11:00

like AlphaFold, directly being

11:02

useful for biologists. And so to

11:05

me, this is just indicative

11:08

of a new phase that AI

11:10

is in of being practically useful to

11:12

people in their everyday lives and actually

11:14

being able to solve really hard real world

11:16

problems that really matter, not just the

11:18

curiosities or fun like games. When

11:21

you recognize that shift, I think

11:23

that necessitates a change in your approach

11:26

as to how you're approaching the research and

11:29

how much focus you're having on products and those

11:31

kinds of things. And I think that's what we

11:34

all came to the realization of, which

11:36

now was the time to streamline our

11:38

AI efforts and focus the

11:41

more the obvious conclusion of that

11:43

was to do the merger.

11:44

I want to just stop there for one

11:47

second and ask a philosophical question. It

11:50

feels like the chat GPT

11:52

moment that led to this AI explosion this

11:54

year

11:55

was really rooted in the AI being

11:57

able to do something that regular people could do. I

12:01

want you to write me an email. I want you to write me a screenplay.

12:03

And maybe it's a C, maybe the output of the LLM

12:06

is a C plus,

12:07

but it's still like something I can do, right? Like people

12:09

can see it. I want you to fill out the rest of this photo.

12:12

That's how many people can imagine doing, maybe

12:14

they don't have the skills to it, but they can imagine doing it.

12:17

All the previous AI demos that we

12:19

have gotten, even yours, Alpha Fold, you're

12:21

like, this is gonna

12:22

model all the proteins in the world. Like

12:25

I can't do that. That's like, great, a computer should do

12:27

that. Like even a microbiologist might think,

12:29

that is great, I am very excited that a computer can

12:31

do that because I'm just looking at how much time it would

12:33

take us and there's no way we could ever do it. I

12:36

wanna beat the world champion at Go. I

12:39

can't do that. It's like, fine, a computer can do that. There's this

12:42

turn where the computer is starting

12:44

to do things I can do.

12:45

And it's not even do it, and they're not even like necessarily

12:48

the most complicated tasks. Like read

12:50

this webpage and deliver a summary of it to

12:52

me. But that's

12:54

the thing that unlocked everyone's brain. And

12:56

I'm wondering why you think the industry didn't see that turn

12:59

coming. Because we've been very focused

13:01

on these very difficult things that people couldn't

13:03

do. And it seems like what got

13:05

everyone is when the computer started doing things

13:08

people do all the time.

13:10

Yeah, I think that analysis is correct. I think

13:12

that is why the large language

13:15

models have really entered the public consciousness

13:17

because it's something the average

13:19

person, the Joe public can

13:21

actually understand and interact

13:24

with. And of course, language is core

13:26

to human intelligence and our

13:28

everyday lives. So I think

13:30

that does explain why chat bots specifically

13:33

have sort of gone viral in the way they

13:35

have. Even though I would say

13:37

things like alpha fold,

13:40

I mean, of course I'd be biased in saying

13:42

this, but I think it's actually had the

13:44

most unequivocally biggest

13:47

sort of beneficial effects so far in AI

13:49

on the world, because there's a million biologists

13:52

now, researchers and medical researchers have used

13:54

alpha fold. I think that's nearly every biologist in the

13:56

world, every big pharma company is using

13:58

it to advance their drug discovery. programs. I've

14:01

had dozens of Nobel Prize winner

14:03

level biologists and chemists talk

14:06

to me about how they're using alpha-fold. So

14:08

a certain set of all the wealth scientists,

14:10

let's say, they all know alpha-fold. It's

14:13

affected and massively accelerated their

14:15

important research work. But of

14:17

course, the average person in the street doesn't know

14:20

what proteins are even and

14:22

doesn't know what the importance of those things

14:25

are for things like drug discovery. Whereas

14:27

obviously, for a chatbot, everyone can understand

14:29

this is incredible. And it's very visceral to

14:32

get it to write you a poem or something

14:35

everybody can understand and process and measure

14:38

compared to what they do

14:40

or are able to do.

14:42

It seems like that is the focus of

14:44

productized AI, these

14:46

chatbot interfaces or these

14:49

generative products that are going to make stuff

14:51

for people. That's

14:53

where the risk has been focused. Even a conversation

14:56

about risk has dramatically escalated.

14:58

I want to make sure we talk about that at length. But

15:00

even the conversation about risk has escalated

15:03

because people can now see, oh, these schools can do

15:05

stuff. Did you perceive the

15:07

same level of scrutiny when you

15:09

were working on alpha-fold? It

15:11

doesn't seem like anyone thought, oh, alpha-fold is going to

15:14

destroy humanity.

15:15

No, but there was a lot of scrutiny,

15:17

but just again, it was in a very specialized area

15:20

with renowned experts. And actually, we did talk

15:23

to over 30 experts

15:25

in the field from top

15:27

biologists to bioethicists to biosecurity

15:30

people. And actually, our partners, we

15:32

partnered with the European Bioinformatics Institute

15:34

to release the alpha-fold database of all the protein

15:37

structures. And they guided us

15:39

as well on how this could be safely put

15:41

out there. So there was a lot of scrutiny.

15:43

And the overwhelming conclusion from the people we

15:46

consulted was that the benefits far outweighed

15:48

any risks, although we did make some small

15:50

adjustments based on their feedback about which

15:53

structures to release. There was a lot of scrutiny,

15:55

but again, it's just in a very expert

15:58

sort of domain. And with the... Just

16:00

going back to your first question on the generative models,

16:02

I do think we are right at

16:05

the beginning of an incredible

16:07

new era that's going to play out over the next five,

16:09

10 years, not only in advancing science

16:12

with AI, but in terms of

16:14

the types of products we can build

16:16

to improve people's everyday lives,

16:18

billions of people in their everyday lives and help

16:21

them to be more efficient and to enrich their

16:23

lives. I think what we're seeing today

16:25

with these chatbots is literally

16:27

just scratching the surface. There's

16:29

a lot more types of AI than generative AI. Generative

16:32

AI is now the in thing, but I think

16:34

that planning and deep reinforcement

16:37

learning and problem solving and

16:39

reasoning, those kinds of capabilities

16:41

are going to come back in in the

16:43

next wave after this, along

16:45

with the current capabilities of the current systems.

16:48

I think in a year or two's time, if

16:50

we were to talk again, we're going to be

16:53

talking about

16:54

entirely new types of products

16:56

and experiences and services that

16:59

we've never seen before capabilities. I'm

17:01

very excited about building those things,

17:03

actually. That's one of the reasons I'm very excited

17:05

about leading Google DeepMind now

17:07

in this new era and focusing on

17:10

building these AI-powered

17:12

next generation products.

17:14

Let's stay in the weeds of Google

17:16

DeepMind itself for one more turn. Sunar

17:19

Pashai comes to you and says, all right, I'm the CEO of

17:21

Alphabet and the CEO of Google. I can just make this call.

17:23

I'm going to bring DeepMind into Google, merge you with Google

17:26

Brain. You're going to be the CEO. How did you

17:28

react to that prompt? It wasn't

17:30

like that. It was much more of

17:32

a conversation between the leaders of

17:35

the various different relevant groups and

17:37

Sunar about the

17:39

inflection point that we're seeing, the

17:41

maturity of the systems and

17:43

what could be possible

17:44

with those in the product

17:46

space and how to improve

17:48

experiences for our users, our billions

17:50

of users, and how exciting

17:52

that might be and what that all requires

17:55

in totality, both the change in focus,

17:57

a change in the approach to research. the

18:00

combination of resources that are required,

18:02

like compute resources. So there

18:04

was a sort of big collection of factors

18:07

to take into account that we all discussed

18:09

as a leadership group. And then

18:12

conclusions from that then result in

18:14

actions, including the merger

18:17

and also what the plans are then for

18:19

the next couple of years and what

18:21

the focus should be of that merged unit.

18:25

Do you perceive a difference being a

18:27

CEO inside of Google versus being a CEO

18:30

inside of Alphabet?

18:31

It's still early days, but I

18:33

think it's been pretty similar because

18:36

although DeepMind was an Alphabet

18:38

company, it was very unusual for

18:41

another bet, as they call it, Alphabet, which

18:44

is that we already were very

18:46

closely integrated and

18:48

collaborating with many of the

18:50

Google product area teams and

18:53

groups. We have an applied team at

18:55

DeepMind whose job it was to

18:58

translate our research work into

19:00

features and products by collaborating with

19:02

the Google product teams. And so we've

19:05

had hundreds of successful launches already actually

19:07

over the last few years, just quiet ones behind the scenes.

19:09

So in fact, many of the services

19:11

or devices or systems that you

19:14

use every day at Google will have some

19:16

kind of DeepMind technology under the hood as

19:19

a component. So we already had that

19:21

integrative structure. And then, of course,

19:24

what we were famous for was doing the scientific

19:26

advances and gaming advances. But

19:28

behind the scenes, there was a lot of bread and butter work going

19:31

on that was affecting all parts

19:33

of Google. We were different from other bets where

19:35

they have to make a business outside

19:37

of Google and become

19:40

an independent business. That was never the

19:42

goal or the remit for us, even as

19:44

an independent company. And

19:47

now within Google, we're just more

19:49

tightly integrated in terms of the product services.

19:51

And I see that as an advantage because we can

19:53

actually go deeper and do more exciting and

19:56

ambitious things in a much closer collaboration

19:58

with these other companies.

19:59

product teams than we could from outside

20:02

of Google. But we still retain some

20:05

sort of latitude to pick the processes

20:07

and the systems that optimize our

20:09

mission of producing the

20:11

most capable, in general, AI systems in

20:13

the world. There's been a bunch of reporting

20:16

that this is actually a bit of a culture clash. You're

20:18

now in charge of both. How have you structured the

20:20

group? How is Google DeepMind structured under

20:23

you as CEO? And how are you managing that

20:25

culture integration? Actually,

20:27

it turns out the culture is a lot more

20:29

similar

20:29

than perhaps has been reported externally.

20:32

And in the end, it's actually been surprisingly

20:35

smooth and pleasant because you're talking

20:37

about two world-class research groups,

20:40

two of the best AI research

20:42

organizations in the world, incredible

20:44

talent on both sides, storied

20:46

histories. As we were thinking about

20:48

the merger and planning it, we were looking

20:51

at we had some document where we listed,

20:54

I guess, the top 10 breakthroughs from each group.

20:57

And when you take that in totality, it's

21:00

like 80%, 90% over the

21:02

last decade of the breakthroughs that underpin

21:05

the modern AI industry, from deep reinforcement

21:08

learning to transformers, of course. Transformer

21:10

is a type of neural network architecture. So

21:12

it's an incredible set of people and talent.

21:15

And there's massive respect for both

21:17

groups on both sides. And there was actually a lot

21:19

of collaboration on a project-based

21:22

level ongoing over the last decade.

21:24

So of course, we all know each other very well.

21:27

I just think it's a question of focus

21:30

and a bit of coordination across both

21:32

groups, actually, and more, in

21:34

terms of what are we going to focus on,

21:38

other places that make sense to collaborate,

21:41

two separate teams to collaborate on, and

21:43

maybe deduplicate some efforts that

21:45

basically are overlapping. So

21:47

fairly obvious stuff, to be honest. But

21:50

it's important, moving into this new phase

21:52

now, where we're kind of into more of an engineering

21:55

phase of AI. And that requires

21:57

huge resources, both compute, engineering,

21:59

other things. And even as a

22:02

company the size of Google, we've got to pick

22:04

out bets carefully and be clear about

22:06

which arrows we're going to put our

22:08

wood behind and then focus on those and

22:11

then massively deliver on those things. So it's

22:13

just, I think it's part of the natural course

22:15

of evolution as to where we are in

22:18

the AI journey. So that thing

22:20

you talked about, we're going to combine these groups, we're

22:22

going to pick what we're doing, we're going to deduplate some efforts.

22:25

Those are structure questions. Have you decided on a

22:27

structure yet?

22:28

And what do you think that structure will be?

22:30

We have. I mean, the structure is still evolving.

22:32

We're only a couple of months into it. We wanted

22:34

to make sure we didn't break anything that

22:36

was working. Both teams are incredibly productive

22:39

doing super amazing research, but

22:42

also plugging in to very

22:44

important product things that are going

22:46

on. You keep saying

22:48

both teams. Do you think of it as both two

22:51

teams or are you trying to make one team? No,

22:53

no, it's for sure. It's one unified team.

22:55

I like to call it a super unit.

22:58

And I'm very excited about that. But obviously

23:00

we're still sort of combining that and forming

23:02

the new culture and forming the new grouping,

23:05

including the organizational structures as

23:07

complex things, putting two big

23:09

research groups together like this. But I

23:12

think over by the end of the summer, you

23:14

know, we'll be a single unified entity.

23:16

And I think that'd be very exciting. And we're already feeling

23:19

even a couple of months in the benefits

23:21

and the strengths of that with projects like

23:23

Gemini that you may have heard of, which is our

23:25

next generation multimodal large

23:28

models. Very, very exciting work going

23:30

on there, combining all the

23:32

best ideas from across both

23:34

world class research groups. Gemini,

23:37

by the way, is Google's next generation AI

23:39

language model. Think of it like GPT-4 compared

23:42

to GPT-3.

23:45

We have to take a quick break. We'll be back in a moment.

23:55

This

23:57

episode is brought to you by Shopify.

24:00

That's the sound of switching your business to Shopify,

24:03

the global commerce platform that supercharges

24:05

your selling. Harness the best converting checkout

24:07

and same intuitive features, trusted

24:10

apps, and powerful analytics used by

24:12

the world's leading brands. Stop

24:14

leaving sales on the table. Discover

24:16

why millions trust Shopify to build,

24:18

grow, and run their business. Sign

24:20

up today for your $1 per month trial period

24:23

at shopify.com slash tech23.

24:26

Hello, I'm Esther Perel. I'm

24:28

a psychotherapist and host of the podcast,

24:30

Where Should We Begin? Relationships

24:33

expectations are at an all-time

24:35

high, and yet the norms

24:38

are less and less clear, and we are

24:40

literally making up the new norms

24:42

as we go. Whether it's your

24:44

work relationships, friendships, or

24:47

romantic relationships, I invite

24:49

you to enter into my office

24:52

and listen in on my sessions, where

24:55

I help people explore the challenges

24:57

and choices in their relationships.

25:01

You will listen intensely to

25:03

them, but you will actually see yourselves.

25:07

And in the process, you become unstuck

25:10

and empowered in your own relationships. Join

25:14

me in my office every Monday

25:16

morning for a new episode. Listen

25:19

and follow Where Should We Begin on your

25:22

favorite podcast app.

25:28

We're back with Dennis DeSantis. You

25:33

have a lot of decisions to make, right? Well, you're describing

25:35

as much of complicated decisions than out in the world. How should we regulate this,

25:37

which is another set of very complicated decisions? And

25:41

you're a chess champion. You're

25:43

a person who's made the games. So what's

25:45

your framework for making decisions? I

25:48

suspect it's more rigorous than other ones I might

25:50

hear about. I think it probably is.

25:51

And I think

25:54

if you play a game like chess that seriously, effectively,

25:56

professionally, since, you know, in all

25:59

my childhood, since the age of 10, I think

26:01

it's very formative for your brain. So

26:04

chess is a sort of problem solving

26:06

and strategizing. I find it a very

26:08

useful framework for many things

26:11

and decision making. Chess is basically

26:13

decision making under pressure with

26:15

an opponent. It's very complex and I think it's a

26:17

great thing. I advocate it being taught

26:19

at school, actually, part of the school curriculum, because I think

26:21

it's a really fantastic training ground

26:24

for problem solving and decision making. But

26:27

then I think actually the overarching approach

26:29

is more of like the scientific method. So

26:31

I think all my training is doing my PhDs

26:34

and postdocs and so on. Obviously,

26:36

I did it in neuroscience, so I was learning about the brain,

26:39

but it also taught me how to do

26:41

rigorous sort of hypothesis testing and hypothesis

26:44

generation and then update

26:46

based on empirical evidence. So

26:49

I think the whole scientific method, as

26:51

well as the chess planning, both can

26:54

be translated into the business domain.

26:57

You have to be smart about how to translate that, so

26:59

you can't be academic about these things. And

27:01

often in the real world, in business, there's

27:03

a lot of uncertainty and hidden

27:05

information that you don't know. So

27:08

in chess, obviously, all the information is there

27:10

for you on the board. You can't just sort of directly

27:12

translate those skills, but I think

27:14

in the background, they can be very helpful

27:17

if applied in the right way. Give me an example. Make that real

27:19

for people. How do you combine those two in some decision you've made? There's

27:22

so many decisions I make every day. It's hard

27:24

to come up with one now, but I tend

27:26

to try and plan out and scenario

27:28

plan many, many years in advance. I

27:31

tell you, the way I try to approach things is

27:33

I have an end goal. I'm quite good at imagining

27:36

things, so that's a different skill. Visualizing

27:38

or imagining what an end state,

27:41

a perfect end state would look like, whether that's

27:43

organizational or it's product-based or it's

27:45

research-based. And then

27:47

I kind of work back from the end point

27:50

and then figure out what all the steps would

27:52

be required and in what order

27:54

to make that outcome as likely

27:57

as possible. So that's a little

27:59

bit chess-like. right? In the sense of like you

28:01

have some plan that you would like to get

28:03

to checkmate, you know, your opponent, but

28:06

your many moves away from that. So what

28:08

are the incremental things one must do to

28:10

improve your position in order to increase

28:13

the likelihood of that final outcome?

28:16

And I found that extremely useful

28:18

to do that search process from, you

28:21

know, the end goal back to the state

28:23

that the current state that you find yourself

28:25

in.

28:26

Let's put that next to some products, right? You said there's a lot

28:28

of deep mind technology in a lot of Google products. I

28:31

think obviously the ones that we can all look

28:33

at are barred in the new search

28:35

generative experience. Obviously there's AI

28:37

and Google photos and all the stuff, but focused

28:40

on sort of the LLM moment, it's barred in the search generative

28:42

experience. Those can't be the end

28:44

state, right? Like that, that's not,

28:46

they're not finished and Gemini is coming and we'll probably

28:48

improve both of those and all that will happen. When

28:51

you think about the end state of those products, what do you

28:53

see?

28:53

The AI systems around Google are

28:56

also not just in the, in the consumer facing

28:58

things, but also under the hood, right? That you may not

29:00

realize. So even for example, one of

29:02

the things we applied our AI systems to

29:05

very initially was the cooling systems

29:07

in Google's data centers, enormous data

29:09

centers, and actually reducing the energy

29:11

they use by nearly 30% that the

29:14

cooling systems use, which is obviously huge if you

29:16

multiply that by all of the data centers

29:18

and computers they have there. So there's

29:21

actually a lot of things under the hood where AI is being used

29:23

to improve the efficiency of those systems

29:25

all the time. But you're right, the current

29:27

products are not the end state. They're just, they're

29:30

actually just way points, I would say in the

29:32

case of chat bots and,

29:34

and those kinds of systems, I think ultimately

29:36

they will become these incredible universal

29:39

personal assistants that you kind of use

29:42

multiple times during the day for

29:44

many, many things across really useful

29:46

and helpful things across your daily lives, from

29:50

what books to read to, you know, recommendations

29:52

on, on maybe live events and things like that,

29:54

to booking your travel, to

29:57

planning, you know, trips for you to a system.

30:00

existing year in your everyday work.

30:02

I think we're still far away from that

30:05

with the current chatbots. I

30:07

think we know what's missing,

30:09

things like planning and reasoning and memory.

30:12

We're working really hard on those things.

30:15

I think what you'll see in maybe

30:17

a couple of years' time, today's chatbots

30:20

will look trivial by comparison to

30:23

what's coming in the next few years.

30:25

My background is as a person who's reported on

30:27

computers, I think of computers as somewhat modular systems.

30:29

You look at a phone, it's got a screen, it's got a chip, it's got

30:31

a cell antenna, whatever.

30:33

Should I look at AI systems that way?

30:35

There's an LLM, which is a very convincing

30:38

human language interface, and behind

30:40

it might be alpha fold that's actually doing

30:43

the protein folding. Is that how you think about stitching

30:45

these things together, or is it a different

30:47

evolutionary pathway?

30:49

Actually, there's a whole branch

30:51

of research going into what's called tool

30:53

use. This is the idea that

30:56

these large language models or large multi-modal

30:58

models, they're expert at language,

31:00

of course, and maybe a few other capabilities

31:03

like math and possibly coding. But

31:05

when you ask them to do something specialized,

31:08

like fold a protein or play

31:10

a game of chess or something like this,

31:13

actually what they end up doing is calling a tool,

31:15

which could be another AI system, that

31:18

then provides the

31:20

solution or the answer to that particular problem.

31:23

Then that's transmitted back to the user via

31:25

language or pictorially

31:28

through the central large language model

31:30

system. It may be

31:33

invisible to the user because to the

31:35

user it just looks like one big AI system that

31:37

has many capabilities. But under the hood,

31:40

it could be that actually the AI

31:42

system is broken down into smaller ones that

31:44

have specializations. I

31:47

actually think that probably is going to be the next

31:49

generation of systems will use those

31:53

kinds of capabilities. Then you can think

31:55

of the central system as almost

31:58

a switch statement that you... you

32:00

effectively prompt with language, and

32:02

it roots your query or your question

32:05

or whatever it is you're asking it to

32:07

the right tool

32:09

to solve that question for you or

32:12

provide the solution for you. And

32:14

then transmit that back in a very understandable

32:16

way, again, using through the interface,

32:19

the best interface really of natural

32:21

language. Does that get you closer to

32:23

an AGI? I know that's like in your Twitter

32:25

bio, right? This is where you are headed is AGI.

32:28

Does that

32:29

process get you closer to an AGI? Or does that

32:31

get you to some sort of maximum

32:33

state and you got to do something else? I think

32:36

that is on the critical path to

32:38

AGI.

32:39

And that's another reason, by the way,

32:41

I'm very excited about this new role and

32:44

actually doing more products and things. Because

32:46

I actually think the product roadmap

32:48

from here and the research roadmap

32:50

from here towards something like AGI

32:53

or human level AI is very

32:55

complimentary, right? So I think the kinds

32:57

of capabilities one would need to push in

33:00

order to build those kinds

33:02

of products that are useful in your everyday

33:04

life, like a universal assistant, requires

33:07

pushing on some of these capabilities like planning

33:09

and memory and reasoning that I think are

33:11

vital for us to get to AGI. So

33:14

I actually think there's a really neat feedback

33:16

loop now between products and research

33:19

where they can effectively help

33:22

each other.

33:22

I feel like I had a lot of car CEOs in the show

33:25

at the beginning of it. I asked all of them,

33:26

when do you think we're going to get self-driving cars? And they all said

33:29

five years. And they've been saying five years for five years, right?

33:31

Yes. I'm going to ask you the same version

33:34

of that question about AGI, but I feel like that the number

33:36

has gotten smaller recently with people I've talked to. How

33:39

many years until you think we have AGI?

33:40

I think there's a lot of uncertainty over

33:43

how many more breakthroughs are required

33:46

to get to AGI, big, big breakthroughs,

33:48

innovative breakthroughs, versus just scaling

33:51

up existing solutions. And I think

33:54

it very much depends on that in terms

33:56

of timeframe, obviously if there are a lot

33:58

of breakthroughs still required.

33:59

those are a lot harder to do and take

34:02

a lot longer. But right now, I would

34:05

not be surprised if we approach

34:07

something like AGI or AGI like in

34:10

the next decade. In the next decade. I'm

34:12

gonna come back to you in 10 years when I see if that happens.

34:14

Sure.

34:17

That's not a straight line though. You call it the critical path. That's

34:19

not a straight line, right? There's breakthroughs along the

34:21

way that might upset

34:23

the train and send you along a different

34:25

path, you think?

34:26

Yeah, research is never a straight line, right? If

34:28

it is, then it's not real research. If

34:31

you knew the answer before you started it, then that's

34:33

not research. So research and blue sky research

34:35

always, sort of at the frontier,

34:38

always has uncertainty around it. And

34:40

that's why you can't really predict timelines with

34:42

any certainty. But what you can look

34:44

at is trends and we can sort of look at

34:46

the quality of ideas and projects that are being

34:49

worked on today, look at how they're progressing.

34:52

That could go either way over the next five to 10 years

34:55

where we might asymptote, we might hit a

34:57

brick wall with current techniques and

35:00

scaling we may find, I

35:02

wouldn't be surprised if that happened either that we may find

35:04

that just scaling the existing systems resulted

35:07

in diminishing returns in terms

35:09

of the performance of the system. And actually

35:11

that would then signal some

35:13

new innovations were really required to

35:15

make further progress. And so at the moment,

35:17

I think nobody knows which regime

35:19

we're in. So the answer to that is you have

35:22

to push on both as hard as possible.

35:24

So both the scaling and the engineering

35:27

of existing systems and existing ideas,

35:30

as well as investing heavily

35:32

into exploratory research

35:35

directions that you think might

35:37

deliver innovations that might solve

35:40

some of the weaknesses in

35:42

the current systems. And that's one advantage

35:44

of being a large research organization with a lot

35:46

of resources is we can bet on both

35:49

of those things maximally, both

35:51

of those directions. So I think in

35:53

a way, I'm kind of agnostic to that question

35:56

of like, do we need more breakthroughs or will existing

35:58

systems just scale all the way? the way, my

36:00

view is it's an empirical question. One

36:03

should push both as hard as possible

36:05

and then the results

36:08

will speak for themselves.

36:10

This is a real tension. When

36:12

you were deep-minded and alphabet and you were very

36:14

research-focused, and you had the luxury

36:17

of that, and then the research was moved back

36:19

into Google, and Google's engineers would turn

36:21

it into products. You can see how that relationship worked.

36:23

Now you're inside of Google. Google is under a lot of

36:25

pressure as a company to win this battle.

36:28

Those are product concerns. Those are make it real

36:30

for people and go win in the market. There's

36:33

a leaked memo that went around. It was purportedly from inside Google.

36:35

The company had no moat in it. Open

36:38

source AI models or leaked models

36:40

would run on people's laptops and they would outpace

36:42

the company because the history of open computing

36:45

would outpace a closed source competitor. Was

36:47

that memo real?

36:48

I think that memo was real. Engineers

36:52

at Google often write various documents

36:54

and sometimes they get leaked and go viral.

36:56

I think that's just a sort of kind

36:59

of thing that happens. I

37:02

wouldn't take it too seriously. These

37:04

are just opinions. I think it's

37:07

interesting to listen to them and then you've

37:09

got to chart your own course. I haven't read

37:11

that specific memo in detail, but I

37:13

disagree with the conclusions from that. There's

37:16

obviously open source and publishing.

37:20

We've done tons of that in the history of DeepMind.

37:22

We gave, our fold was open sourced. We

37:25

obviously believe in open source

37:27

and supporting research and open research.

37:30

That's a key thing of the scientific discourse,

37:33

which we've been a huge part of. As Google,

37:35

of course, publishing transformers and other things

37:38

and tensorflow. You look at all the things we've

37:40

done, we do a huge amount in that

37:42

space. I also think there are

37:44

other considerations that need to be had as well. Obviously

37:47

commercial ones, but also safety questions

37:50

as well about access to these very powerful

37:52

systems. If bad actors can access

37:54

it who maybe are that technical, so they couldn't

37:56

have built it themselves, but they can certainly

37:58

reconfigure a system. that is out there,

38:01

what do you do about those things? And

38:03

I think that's been quite theoretical till

38:05

now, but I think that that is really

38:08

important from here all the

38:10

way to AGI as these systems

38:13

become more general, more sophisticated,

38:15

more powerful. That question is

38:17

going to be very important about how does

38:19

one stop bad actors just using

38:22

these systems for things they weren't intended

38:24

for, but for malicious purposes. That's

38:26

something we need to increasingly come

38:28

up with. But just to back to your question,

38:31

is look at the history of what Google

38:33

and DeepMind have done in terms of coming

38:35

up with new innovations and breakthroughs and multiple,

38:38

multiple breakthroughs over the last decade or

38:40

more. I would bet on us

38:42

and I'm certainly very confident that that will

38:44

continue and actually be even

38:47

more true over the next decade in

38:49

terms of us producing the next

38:51

key breakthroughs, just like we did

38:54

in the past. Do you think that's the moat? We invented

38:57

most of this stuff, so we're going to invent the most of the next

38:59

stuff. Yeah, I don't really think about those moats,

39:01

but I'm incredibly competitive

39:03

person. That's maybe another thing I got from

39:05

Jess and many researchers

39:07

are. Of course, they're doing it to discover

39:10

knowledge and ultimately that's what we're here

39:12

for, is to improve the human

39:14

condition. But also we want to be first to do

39:16

these things and do them responsibly

39:18

and boldly. We have some of the world's best

39:20

researchers. I think we have the biggest collection of

39:23

great researchers in the world, anywhere in the world. And

39:26

incredible track record and there's no reason

39:28

why that shouldn't continue in the future.

39:31

And in fact, I think with our new organization

39:33

and environment might be conducive to

39:36

even more and faster paced

39:38

breakthroughs than we've done in the past.

39:40

You're leading me towards risk and regulation.

39:42

I want to talk about that, but I want to start with just sort

39:44

of a different spin on it. You're

39:46

talking about all the work that has to be done. You talk about deep mind,

39:49

reinforcement learning, how that works. We just

39:51

ran a gigantic story in collaboration with

39:53

New York Magazine. It's on the cover of New York Magazine

39:55

about the taskers who are actually

39:57

doing the training, who are actually labeling.

39:59

the data, there's a lot of labor

40:02

conversation with AI along the way. Hollywood

40:05

writers are on strike right now because they don't want chat GPT

40:07

to write a bunch of scripts. I think that's appropriate.

40:10

But then there's a new class of labor that's being developed

40:12

where a bunch of people around the world

40:15

are sitting in front of computers and saying, yeah, that's a stop sign.

40:18

No, that's not a stop sign. Yep, that's clothes you can wear. No,

40:20

that's not clothes you can wear. Is that

40:22

a forever state? Is that just a new class of

40:24

work that needs to be done for these systems to operate? Or

40:26

does that come to an end? It's hard

40:28

to say. I think it's definitely

40:31

a moment in time and

40:33

the current systems and what they're requiring

40:35

at the moment. We've been very careful

40:37

just to say from our part, and I think you quoted

40:40

some of our researchers in that article to

40:42

be very careful to pay living wages

40:45

and be very responsible about

40:47

how we

40:48

do that kind of work and which partners we use.

40:51

And we also use internal teams as well. So

40:53

I think we've been actually, I'm very proud of how

40:56

responsible we've been on that type of work.

40:59

But going forwards, I

41:01

think there may be ways that these systems,

41:04

especially once you have millions and millions

41:06

of users, effectively can

41:08

bootstrap themselves. Or

41:11

one could imagine AI systems

41:13

that are capable of actually sort

41:16

of conversing with themselves or critiquing themselves.

41:18

This would be a bit like turning language systems

41:21

into a game-like setting, which of course we're

41:23

very expert in. And we've been thinking about where these

41:25

reinforcement learning systems, different versions

41:28

of them can actually sort of rate each

41:30

other in some way. And it may not be

41:32

as good as a human rater, but it's actually

41:35

a useful way to sort of do some

41:37

of the bread and butter rating and then maybe just calibrate

41:39

it by checking those ratings

41:42

with a human rater at the end, rather

41:44

than getting human raters to rate everything. So

41:46

I think there are lots of innovations

41:48

I can see coming down the line that

41:51

will help with this and potentially

41:54

mean that there's less requirement for

41:56

this all to be done by human raters. But

41:58

you think there's always human raters.

41:59

the mix. Even as you get closer to AGI,

42:02

it seems like you need someone to

42:04

tell the computer if it's doing a good job or not.

42:06

Let's take Alpha Zero as an example. Our

42:09

general games playing system that ended up learning

42:11

itself how to play any two-player game,

42:14

including chess and go. It's interesting.

42:16

What happened there is we set up the system

42:18

so that it could play against itself

42:21

tens of millions of times. In fact,

42:24

it built up its own knowledge base. It started

42:26

from random, played itself, bootstrapped

42:28

itself, trained better versions of itself, and

42:30

played those off each other in mini tournaments.

42:33

But at the end, you still want to test it

42:35

against the human world champion or something like

42:37

this, or an external computer

42:39

program that was built in a conventional

42:41

way so that you can just calibrate your

42:44

own metrics, which are telling you these systems

42:47

are improving according to these objectives

42:49

or these metrics. But you don't know for sure

42:52

until you calibrate it with an external benchmark

42:54

or measure. Depending on what that is,

42:57

a human rater or human benchmark, human expert

42:59

is often the best thing to calibrate

43:02

your internal testing against. You

43:05

make sure that your internal tests are

43:07

actually mapping reality. Again,

43:09

that's something quite exciting about products

43:12

for researchers because when you put your

43:14

research into products and millions of people are using it every

43:16

day, that's when you get real

43:18

world feedback. There's no way around that.

43:21

That's the reality. That's the best test of

43:25

any theories or any system that

43:26

you've built. Do you think that work is rewarding

43:29

or appropriate, the labeling of data

43:31

for AI system? There's just something about

43:34

that, which is I'm going to tell a computer how to understand

43:36

the world so that it might go off in the future and displace

43:39

other people. There's a loop in there

43:41

that seems like it's worth more just moral

43:43

or philosophical consideration. Have you spent

43:45

time thinking about that?

43:47

Yeah, I do think about that. I don't really see

43:49

it like that. I think that what raters

43:51

are doing is they're part of the development

43:54

cycle of making these systems safer,

43:57

more useful for everybody, and

43:59

more

43:59

helpful and more reliable. So

44:02

I think it's a critical component. In

44:04

many industries, we have safety testing, you

44:07

know, technologies and products. And today,

44:10

that's the best we can do for

44:12

AI systems, right is to is to have

44:14

human raters. I think in the future,

44:17

next few years, I think we need more,

44:19

you know, we need a lot more research. And I've been calling

44:22

for this. And we're doing this ourselves, but it needs

44:24

more than just one organization to do this is

44:26

great, robust evaluation benchmarks

44:29

for capabilities, right, so that

44:31

we know, if a system passes

44:34

these benchmarks, then it has

44:36

certain properties and it's safe, and it's reliable

44:38

in these particular ways, right. And right now,

44:41

I think we're in the space of many researchers

44:43

in academia and civil society and elsewhere, we have

44:46

a lot of good suggestions for what those tests

44:48

could be. But I don't think they are robust

44:51

or practical yet. I think they're basically

44:53

theoretical and philosophical in nature. And

44:55

I think they need to be made practical so

44:58

that we can measure our systems, you

45:00

know, empirically against those tests.

45:02

And then that gives us some assurances about about

45:05

how the system will perform. And I

45:07

think it once we have those, then

45:09

the need for this sort of human

45:12

rating testing feedback

45:14

will be reduced. I just think that's

45:17

required in the volumes is required now,

45:19

because we don't have these kinds of independent

45:22

benchmarks yet, partly because we

45:24

haven't rigorously defined what

45:26

those properties are. I mean, it's almost neuroscience

45:30

and psychology and philosophy area

45:32

as well, right. A lot of these terms have not been

45:34

defined properly, even for you know, the

45:36

human brain.

45:40

All right, one more sharp break, we'll be right back.

45:54

We're back with Dennis Asabas, the CEO of

45:56

Google DeepMind. You've

45:58

signed a letter from in the center for AI

46:01

safety, open AI, Sam Altman,

46:03

others who have also signed this letter, that warns

46:06

against the risk from AI.

46:09

And yet you're pushing on, right? Like

46:12

Google's in the market, you've got to win, you've described yourself

46:14

as competitive. There's a tension there,

46:16

right? Needing to win in the market with products and

46:19

oh boy, please regulate us because

46:21

raw

46:22

capitalism will drive

46:24

us off the cliff with AI if we don't stop it in some

46:26

way. How do you balance that risk? It

46:28

is a tension, it's a creative tension. What we

46:31

like to say at Google is we

46:33

wanna be bold and responsible.

46:35

And that's exactly what we're trying

46:37

to do and live out and role model. So

46:39

the bold part is being brave and optimistic

46:42

about what AI, the benefits,

46:44

the amazing benefits, incredible benefits

46:46

AI can bring to the world and to help

46:49

humanity with our biggest challenges, whether that's

46:51

disease or climate or sustainability.

46:54

I think AI has

46:56

huge part to play in helping our

46:58

scientists and medical experts

47:01

solve those problems. And we're working hard

47:03

on that and on all those areas and alpha

47:05

fold again, I point to as a poster child

47:07

for that, what we want to do there. So

47:10

that that's the bold part. And then the responsible

47:12

bit is to make sure we do that

47:14

as thoughtfully as possible with

47:17

as much foresight as possible ahead

47:19

of time, you know, try and anticipate what

47:22

the issues might be if one was

47:24

successful ahead of time, not

47:26

in hindsight, like perhaps has happened with

47:28

social media, for example, where, you

47:30

know, it is this incredible growth

47:33

story, obviously it's done a lot of good in the world,

47:35

but then it turns out 15 years later,

47:37

we realize there are some consequences,

47:40

unintended consequences as well to those

47:42

types of systems. And I would like to chart

47:44

a different path with

47:45

AI. And I think it's such a profound

47:47

and important and powerful technology. I

47:49

think we have to do that with something as potentially

47:52

as transformative as AI.

47:54

And it doesn't mean no mistakes will be made.

47:56

It's, you know, it's very new, anything new, some

47:59

things you...

47:59

You can't predict everything ahead

48:02

of time, but I think we

48:04

can try and do the best job we can. And

48:06

that's what signing that letter was for, is just to point

48:09

out that there are, we

48:11

don't know the, I don't think it's likely, I

48:13

don't think it's, I don't know on the time

48:15

scales, but it's something that we

48:17

should consider too, in the limit, is

48:19

what these systems can do and might be

48:22

able to do as we get closer to AGI. We're

48:24

nowhere near that now. So this is not a question

48:26

of today's technologies or even

48:29

the next few years, but at some point,

48:31

and given the technology is accelerating very fast,

48:34

we will need to think about those questions. And we don't

48:36

want to be thinking about them on the eve

48:39

of them happening. We need to use the time

48:41

now, the next five, 10, whatever

48:43

it is years, to do the research and

48:46

to do the analysis and to engage

48:48

with, various stakeholders,

48:50

civil society, academia, government,

48:53

to figure out as this stuff is

48:55

developing very rapidly, what the

48:57

best way is of making

48:59

sure we maximize the benefits and minimize

49:02

any risks. And that includes mostly at this stage,

49:05

doing more research into these areas, like

49:08

coming up with better evaluations and benchmarks

49:10

to rigorously test the capabilities

49:13

of these frontier systems.

49:15

You talked about tool usage for AI models,

49:18

right? You'd ask an LOM, do something that goes

49:20

off and ask off of all the full protein for you. Combining

49:23

systems like that, integrating systems like that, historically,

49:26

that's where emergent behaviors appear, things

49:28

you couldn't have predicted happen start happening. Are

49:31

you worried about that? There's not a rigorous way

49:33

to test that.

49:34

You assemble a computer, the computer starts doing

49:36

stuff you never expected.

49:38

Right, exactly. I think that's exactly the

49:40

sort of thing we should be researching and thinking

49:42

about ahead of time, is as

49:45

tool use becomes more sophisticated and you can

49:47

combine different AI systems together in different

49:49

ways, there is scope for

49:51

emergent behavior. Of course, that

49:54

emergent behavior may be very desirable and

49:56

be extremely useful, but it could

49:58

also potentially be harder.

49:59

in the wrong hands and

50:02

in the hands of bad actors, whether that's individuals

50:04

or even nation-states.

50:06

Let's say the United States and the EU and China

50:08

all agree on some framework to regulate AI.

50:10

And then North Korea or Iran

50:12

says, fuck it, no rules. And that becomes

50:14

a center of bad actor research.

50:17

How does that play out? Do you see a world

50:19

in which that's possible?

50:20

Yeah, I think that is a possible world.

50:23

This is why I've been talking to governments,

50:25

UK, US mostly, but also EU on,

50:28

I think whatever

50:30

regulations or guardrails or whatever that

50:33

is that transpires over the next few years

50:35

and tests, they ideally would

50:37

be international. And there would be international

50:40

cooperation around those safeguards

50:43

and international agreement around

50:46

deployment of these systems and other things.

50:48

Now, I don't know how

50:50

likely that is given the geopolitical

50:53

tensions around the world, but that is

50:55

by far the best state. And I think what

50:57

we should be aiming for if

50:59

we can.

51:00

If the government here passes a rule, it says, here's the here's

51:02

what Google is allowed to do. Here's what Microsoft is allowed

51:05

to do. You are in charge. You

51:07

are accountable. And you can go say, all

51:09

right, we're just not running this code in our data center.

51:11

We are not going to have these capabilities. It's not legal.

51:13

If I'm just a person with a MacBook,

51:16

would you accept some limitation

51:18

on what a MacBook could do because

51:20

the threat from AI is so scary? That's

51:22

the thing I worry about. Practically, if you have open

51:24

source models and people are going to use them for

51:26

weird things, are we going to tell

51:28

Intel to restrict what its chips can do?

51:30

How would we implement that such that

51:33

it actually affects everyone and not just

51:35

we're going to throw Dennis in jail if Google does stuff

51:37

we don't like?

51:38

Those are the big questions that are being debated

51:40

right now. And I do worry about that.

51:44

On the one hand, there are a lot of benefits of open sourcing

51:46

and accelerating scientific discourse.

51:48

And lots of advances happen there. And it gives

51:50

access to many developers. On

51:53

the other hand, there could be some negative consequences

51:55

with that. If they're about individual actors

51:58

that do bad things with that,

51:59

access

52:00

and that proliferates. And I think that's

52:03

a question for the next few years that will need to

52:05

be resolved because right now I think it's okay

52:07

because the systems are not that sophisticated or that

52:09

powerful and therefore not that risky. But

52:12

I think as the systems increase

52:14

in their power and

52:17

generality, the access question will need

52:19

to be thought about from government and how they

52:21

want to restrict that or control that or

52:24

monitor that is going to be an important question. I

52:26

don't have any answers for you because I think this is a societal

52:28

question actually that requires stakeholders

52:31

from right across society to come together

52:34

and weigh up the benefits with the

52:36

risks there.

52:37

You said we're not there yet, but Google's own work

52:39

in AI certainly had

52:41

some controversy associated with this around

52:44

responsibility, around what the models

52:46

can do or can't do. There's a famous

52:48

stochastic parrots paper from Emily

52:50

Bender and Tim Nicki-Brew and Margaret Mitchell that led

52:52

to a lot of controversy inside of Google, led to

52:55

them leaving. Did you read that paper and

52:57

think, okay, this is a correct. LMS

53:00

are going to lie to people and Google

53:02

will be responsible for that. How

53:05

do you think about that now with all of the scrutiny?

53:08

I think

53:09

the large language models and

53:12

I think this is one reason that

53:14

Google has been very responsible with this is that we know

53:16

that they hallucinate and they

53:18

can be inaccurate. That's one

53:20

of the key areas I think that has to be

53:22

improved over the next few years is factuality

53:26

and grounding and making sure

53:28

that they don't spread disinformation,

53:30

these kinds of things. That's very much top

53:32

of mind for us. We

53:35

have many ideas of how to improve

53:37

that and our old

53:39

DeepMinds Sparrow language

53:41

model was an experiment into,

53:44

which we published a couple of years ago, was an experiment

53:46

into just how good can we get

53:48

factuality and rules adherence

53:51

in these systems. Turns out we can

53:53

maybe make it an order of magnitude better, but

53:55

it sometimes comes at the expense of lucidness

53:58

or creativity on the part. of the language

54:00

model and therefore usefulness. So it's a bit of

54:02

like a Pareto frontier where if you improve one

54:04

dimension, you sort of reduce the

54:07

capability in another dimension. Ideally,

54:09

what we want to do in the next phases and the next

54:12

generations of systems is combine the best of

54:14

both worlds, keep the creativity

54:16

and lucidness and fondness of

54:18

the current systems, but improve

54:21

their factuality and reliability.

54:23

We've got a long way to go on that, but

54:25

I can see things improving and I don't

54:27

see any theoretical reason why

54:30

these systems can't get to

54:32

extremely high levels of accuracy

54:34

and reliability in the next few years.

54:37

When you're using the Google search generative

54:39

experience, do you believe what it says?

54:42

I do. I sometimes double check

54:44

things, especially in the scientific domain where

54:47

I've had very funny situations where actually all of

54:49

these models do this, where you ask them to

54:51

summarize an area of research, which I think would

54:53

be super useful if they could do that, and

54:55

then say, what are the key papers I should read?

54:58

They come up with very plausible signing

55:00

papers with very plausible author

55:02

lists, but then when you go and look into

55:05

it, it turns out they're just like the most

55:07

famous people in that field or

55:10

the titles from two different papers

55:12

combined together, but of course they're extremely plausible

55:15

as a collection of words. I think

55:18

there what needs to happen is these systems need to

55:20

understand that citations

55:22

and papers and author lists

55:25

are a unitary block rather than

55:27

a word-by-word prediction. There's

55:31

interesting cases like that where we need to improve

55:33

and that's something which is, of course, us

55:36

as wanting to advance the frontiers of science. That's

55:38

a particularly interesting use case that we would like

55:40

to improve and fix for our own

55:43

needs as well. I'd love these systems to better

55:45

summarize for me. Here are the top five

55:47

papers to read about

55:48

a particular disease or something

55:50

like that to just quickly onboard you

55:53

in that particular area. I think would be incredibly useful.

55:56

I will tell you, I clicked a link that was basically

55:58

a link to Google, my friend. and John Gruber

56:00

and SGE confidently told me that

56:02

he pioneered the use of a Mac in newspapers

56:05

and invented WebKit. And I

56:07

don't know where that came from. And is

56:10

there a level, is there a quality level, a truthfulness

56:13

level that you need to hit before you roll that out to

56:15

the mass

56:16

audience?

56:17

Yeah, we think about this all the

56:19

time, especially at Google because

56:22

of the incredibly high standards

56:24

Google holds itself to on things

56:26

like search and that we all rely

56:29

on every day in every moment of every day,

56:31

really. We want to kind of get towards

56:33

that level of reliability. Obviously we're a long, long,

56:35

long way away from that at the moment with not

56:38

just us, but anybody with their generative systems.

56:40

But that's the gold standard. And

56:43

actually things like tool use can

56:45

come in very handy here where you could in

56:48

effect build these systems so that

56:50

they fact check themselves, perhaps

56:52

even using search or other

56:55

reliable sources, cross

56:57

reference, just like a good researcher

56:59

would, cross reference your facts. Also

57:02

having a better understanding of the world, what

57:05

are research papers, what entities

57:07

are they, these kinds of things. So

57:09

they can,

57:10

these systems need to sort of get a, have a better understanding

57:13

of the media they're dealing

57:15

with. And maybe also give

57:18

the systems the ability to reason and

57:20

plan because then they could potentially turn

57:22

that on their own outputs and sort

57:24

of critique themselves. And again,

57:26

this is something we have a lot of experience in, in

57:29

games programs that are, you know,

57:31

they don't just output the first move that

57:33

you think of in chess or go, right? You

57:35

actually plan and do some, do

57:38

some search around that and

57:40

then back up. And sometimes they change

57:42

their minds and

57:43

switch to a better move. And that you could

57:45

imagine some kind of process like that with

57:47

words and language as well.

57:49

There's the concept of model collapse,

57:52

right? That we're going to train LLMs

57:54

on a bunch of LLM generated data and that's going to go

57:56

into a circle. When you talk about,

57:59

cross-referencing facts and I

58:02

think about Google, Google going out in the

58:04

web and trying to cross-reference a bunch of stuff

58:06

But maybe all that stuff has been generated by LMS

58:09

that were hallucinating in 2023. How

58:11

do you guard against that?

58:12

We're working on some pretty cool solutions

58:14

to that I think the answer is and

58:17

this is an answer to deepfakes as well is to do

58:20

some sort of encrypted watermarking Sophisticated

58:22

walk marking that can't be removed Easily

58:25

or at all and it's probably built into

58:27

the genitive models themselves So

58:29

it's part of the genitive process We hope

58:31

to release that and maybe provide

58:33

it to third parties as well as a as a generic

58:36

solution But I think that the industry

58:38

in the field needs those types of solutions where we

58:41

can mark Generated media

58:43

be that images audio perhaps even

58:45

text with you know Some kind of kite mark

58:48

that says to the user and

58:51

future AI systems that with

58:53

ease with AI generated And

58:55

I think that's a very very pressing need

58:57

right now for near-term issues

58:59

With AI like deepfakes and disinformation

59:02

and so on but I actually think

59:04

a solution is On the horizon

59:06

now.

59:07

I had Microsoft CTO and EVP of AI

59:09

Kevin Scott on the show a few weeks ago He said something

59:11

very similar. I promised him that we would do a one-hour

59:14

episode on metadata. So you're coming

59:16

for that one That

59:19

will be our most popular episode if I know this

59:21

audience a full hour on metadata

59:23

ideas Will be our most popular episode sounds perfect.

59:26

Well Dennis. Thank you so much for coming on decoder. We

59:28

have to come back soon. Thanks so much

59:32

Thanks again to Dennis

59:34

Sabas for taking the time to chat and decoder today and

59:36

thank you for listening I hope you enjoyed it as

59:39

always. I'd love to hear what you think of decoding You

59:41

can email us at decoder at the verge comm I

59:43

read all the emails We can hit us up

59:45

directly on tick tock check it out.

59:47

It's at decoder pod It's a lot of fun if

59:50

you like decoder, please share with your friends subscribe

59:52

wherever your podcast I really like the show hit

59:54

us with that five-star review Decoder is a

59:56

production of urge and part of the Vox media podcast network

59:59

today's episode is for produced by Raghum Manavallam and Jackie

1:00:01

McDermott. It was edited by Callie Wright. The

1:00:03

decoder of music is by Breakmaster Cylinder. Our editorial

1:00:06

director is Brooke Mentors and our executive director

1:00:08

of video and audio is Eleanor Donovan. We'll

1:00:10

see you next time.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features