Podchaser Logo
Home
Google's Next Top Model + Will the Cybertruck Crash? + This Week in A.I.

Google's Next Top Model + Will the Cybertruck Crash? + This Week in A.I.

Released Friday, 8th December 2023
Good episode? Give it some love!
Google's Next Top Model + Will the Cybertruck Crash? + This Week in A.I.

Google's Next Top Model + Will the Cybertruck Crash? + This Week in A.I.

Google's Next Top Model + Will the Cybertruck Crash? + This Week in A.I.

Google's Next Top Model + Will the Cybertruck Crash? + This Week in A.I.

Friday, 8th December 2023
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

2:00

I'm sorry, it's not.

2:07

And finally, this week, AI.

2:25

Should we podcast? Should we set the timer again? Boom,

2:27

boom. Casey,

2:31

we we are latest addition to the

2:33

podcast studio is a countdown clock, which

2:36

I bought off amazon.com. And the

2:39

express purpose of this clock is to keep us

2:41

from running our mouth for too long and torturing

2:43

our producers with hours of tape that they then

2:45

have to cut. And that sounds horrible. Insert 30

2:47

minute digression. Let's go. OK. All

2:50

right. We're rolling. So Casey, this

2:52

is a big week in the

2:55

AI story in Silicon Valley because

2:57

Google has just released its first

2:59

version of Gemini. It's long

3:01

awaited language model and basically their attempt

3:04

to catch up to open

3:06

AI and chat GPT and GPT for

3:08

and all that. It's America's next top

3:10

model, Kevin. And it's here. And

3:13

I was particularly excited about this because I am

3:15

a Gemini. That's my astrological sign. You know, I'm

3:17

a Gemini as well. No, really? This was really

3:19

the model that was made for us to use.

3:21

Oh, we're twins. Just like Gemini. We're two faced,

3:23

just like Gemini. So

3:26

Gemini is Google's largest and

3:28

most capable model yet. And

3:31

according to Google, it outperforms GPT for on

3:33

a bunch of different benchmarks and tests. We're

3:35

going to talk about all of that. But

3:37

I think we should set the scene a

3:40

little bit, because within the AI world, there

3:42

has been this kind of waiting game going

3:44

on. You know, chat GPT came out roughly

3:46

a year ago. And

3:48

basically from the day that it arrived,

3:51

Google has been playing catch up. And the

3:53

presumption on the part of many people, including

3:55

us, was that Google would put

3:58

a bunch of time and energy and money into this.

4:00

money and computing power into training something

4:02

even bigger and better than what open

4:04

AI was building and basically

4:06

try to sort of throw their muscle into

4:08

the AI race in a really significant way.

4:11

And with Gemini, this is what they appear

4:13

to have done. Finally, we have a terrifying

4:15

demonstration of Google's power. Well,

4:18

so we'll talk about whether it's terrifying or not.

4:21

But let's just talk about what it

4:23

is. So you and I both went

4:25

to a little briefing this week about

4:27

Gemini before it came out. And I

4:29

understand you actually got to do some

4:31

interviews with Google CEO and previous hard

4:34

forecast sooner, Pichai, as well as Demis

4:36

has obvious who is the leader of

4:38

Google DeepMind. That's right. And of

4:40

course, I said, Are you guys sure you don't want

4:42

Kevin in there with me when I do this interview?

4:44

And they said, Trust us, we're sure. So

4:47

I don't know what happened. Yeah. Anyways, I

4:49

did get to interview them. And we had

4:52

a really interesting conversation about kind of how

4:54

they see the road ahead with this stuff.

4:56

They are clearly very excited about what Gemini

4:58

means. And I do think that this is

5:00

kind of like a bit of a starting

5:03

gun going off. And when the most capable

5:05

version of Gemini comes out early next year,

5:07

we really are going to be in a

5:09

kind of horse race between open AI and

5:11

Google. Yeah, so let's just

5:13

talk about what Gemini is, at least

5:16

what we know about it so far.

5:18

So Gemini is actually three models in

5:20

one. It's America's Next Top Models. So

5:23

there are three sizes. There is the

5:25

most capable version, which is called Gemini Ultra.

5:27

This is the one that they say

5:30

can beat GPT for and sort of

5:32

the industry state of the art on

5:34

a bunch of different benchmarks. But

5:37

Google is not releasing Gemini Ultra just yet. They

5:39

say they're still doing some safety testing on that

5:42

and that will be released early next year. By

5:44

the way, if ever again, an editor asked me

5:46

where my story is, I'm gonna say it's not

5:48

ready yet. I'm still doing some safety testing. Very

5:50

good. So

5:52

They have not released a Gemini Ultra,

5:54

but they are releasing Gemini Pro and

5:57

Gemini Nano. These are the sort of

5:59

medium and small. All sizes. Gemini.

6:03

Nano you can actually put onto a

6:05

phone and Google is putting that inside.

6:07

It's Pixel Phones Gemini Pro the sort

6:09

of their equivalent of a Gp T

6:11

Three Point Five and that is being

6:13

released inside of barred starting this week.

6:15

That's right, and now if you are

6:17

listening and you're thinking Kevin just said

6:19

so many different brand names and I'm

6:21

having a meltdown I just wanted as

6:23

the you and I feel you because

6:25

they The branding at Google has always

6:28

been extremely chaotic and the fact that

6:30

we're living in a world where there

6:32

is something called. google. Assistant with

6:34

Barred Powered by Gemini Pro does make me

6:36

want to let emphasis the so I don't

6:38

know who over there is coming up with

6:40

the names for these things, but I just

6:42

want to say stop and I want to

6:44

say go back to Square One Yes, so

6:46

extremely chaotic naming A but what people actually

6:48

care about his homework and the thing do

6:50

Yeah, let's talk about what it can. just

6:52

talk about this. So one of the big

6:54

things Google is advertising with Gemini is that

6:56

it is designed to be what they call

6:58

natively multi modal. Multi Modal is of course

7:00

a I modeled that can work in text

7:02

or images are audio. Or Video. And basically

7:05

the way that multi modal models have

7:07

been built until now is by training

7:09

all these different components like text or

7:11

video separately and them kind of bolting

7:14

them together into a single user interface.

7:16

for Google is saying Well, Gemini it

7:18

was not serve, bolted together like that

7:21

and say was trained on all this

7:23

data at the same time and as

7:25

a result they claim it performed better

7:28

on different tasks that might include like

7:30

having some text alongside an image or

7:32

using it to analyze. frames of a

7:34

video of yes i was i was writing

7:36

about that this model or this week and

7:38

my colleague and as or zoe schiffer is

7:40

written read my peace i was like do

7:43

you have to say multi modal so much

7:45

he likes every one hundred certain awarded months

7:47

ago dallas i just want to stop redux

7:49

and i was very sympathetic but i think

7:51

it is maybe one of the most important

7:53

saying this about this moment and i do

7:55

think by the way in the future we're

7:57

not even going to comment on this be

7:59

because this is just the way that these things

8:02

are going to be built from here on out.

8:04

But it is a very big deal if you

8:06

can take data of all different kinds and

8:08

analyze it with a single tool and then

8:11

translate the results in and out of different

8:13

mediums, right? From text to audio to video

8:15

to images. So that's like a really big

8:17

deal on the path to wherever we're going.

8:20

And it is the reason why this jargon

8:22

word appears in so much of what they're

8:24

saying. Totally. So one

8:26

thing that all the AI companies do, you

8:29

release a new model and you have to put

8:31

it through these big tests, these sort of, what

8:34

they call benchmarks. Yeah, do you remember like high

8:36

school, this is how like how high school in

8:38

Europe works, where you sort of,

8:40

you learn and you learn and you learn, and then

8:42

you take a bunch of tests, and then if you

8:44

succeed, then you get to have a future, and if

8:46

not, you have to become a scullery maid or something.

8:51

My knowledge of Europe ends around like the

8:53

AP Pixies when I finished AP European history,

8:55

but that's like my understanding. Okay,

8:58

so they give these tests

9:00

to Gemini and they- Well,

9:04

they give them to every SOTYX sign, but no,

9:06

I'm sorry, that was a stupid joke. I'm sorry, go

9:08

ahead. No, you should see how Capricorn performs on this

9:10

test. So Gemini Ultra,

9:13

which again, is their top of the

9:15

line model, which is not yet publicly

9:17

available. They give this one a bunch

9:19

of tests. The one that sort

9:21

of caught everyone's attention was the

9:23

MMLU test, which stands for

9:26

Massive Multitask Language Understanding. And

9:28

this is sort of the kind of SATs

9:30

for AI models. It's sort of the standard

9:33

test that every model is put through. It

9:35

covers a bunch of different tasks, including sort

9:37

of math, history, computer science, law. It's kind

9:39

of just like a basic test of like

9:41

how capable is this model? And

9:43

on this test, the MMLU, Google claims that

9:46

Gemini Ultra got a score of 90%. Now

9:49

that is better than GPT-4, which

9:52

was the highest performing model we know about

9:54

so far, which had scored an 86.4%. And

9:59

according to Google, Google, this is a really

10:01

important result because this is the first

10:03

time that a large language model has

10:05

outperformed human experts in the field on

10:07

the MMLU. Researchers

10:09

who developed this test estimate that

10:11

experts in these subjects will get,

10:14

on average, about an 89.8%. Now,

10:17

the rate of progress here is really striking,

10:20

and it's not the only area of testing

10:22

that they did that I think the rate

10:24

of progress was really nothing to pay attention

10:26

to. So there's also the MMMU, which is

10:29

the Marvel Cinematic Universe, is that right? Yes.

10:32

So this is the massive multidisciplined

10:34

multimodal understanding and reasoning benchmark, say

10:36

that five times fast. And

10:39

this is a test that evaluates

10:42

AI models for college-level subject knowledge

10:44

and deliberate reasoning. And

10:46

on this test, Gemini Ultra scored

10:48

a 59.4%. This

10:52

is, I guess, a harder test. I don't like it. And

10:55

GPT-4, by comparison, scored a 56.8%. So

10:58

it's better than GPT-4 on at least these

11:01

two tests. Now,

11:03

there's some question on social

11:05

media today about whether this

11:07

is a true apples-to-apples comparison.

11:10

Some people are saying, like, GPT-4 may be

11:12

still better than Gemini, depending on sort of

11:14

how you give this test. But it doesn't

11:17

really matter. What matters is that Google

11:19

has made something that it says can

11:22

basically perform as well or better than GPT-4.

11:24

Yeah. I think the ultimate

11:26

question is just, like, is the output better on Google's

11:29

products than it is on OpenAI? So that's all that

11:31

really matters. Yeah. But again, this

11:33

is the version of the model that we do not have access to

11:35

yet. It is not out yet. So it's

11:37

hard to evaluate it yet. Yeah. And

11:40

obviously, we're looking forward to trying it.

11:42

But in the meantime, they're giving us

11:44

Pro. Yes. I just got access

11:46

to Gemini Pro in Bard just a few

11:49

hours ago. I haven't had a chance to

11:51

really, like, put it through its paces yet.

11:53

You haven't had a chance to develop a

11:55

romantic relationship with it? Although

11:58

I did have a very funny first interaction with it

12:00

I'll tell you what this is so

12:03

I I just said hello there and it

12:06

said general

12:08

Kenobi image of Obi-Wan

12:10

Kenobi saying hello there interaction

12:15

with the new bard so it immediately

12:17

turned into Obi-Wan Kenobi from Star Wars

12:19

for reasons I do not immediately understand

12:21

wait can I tell you what my

12:23

first interaction was I was

12:25

trying to figure out if I had

12:27

access to it okay and so I

12:30

said are you powered by Gemini

12:33

right and it said no Gemini is a

12:36

cryptocurrency exchange which is true

12:38

there is a current exchange

12:40

called run by the Winklevoss yes

12:42

exactly but it's always funny to me when the

12:44

models hallucinate about what they are you know it's

12:46

like you don't even understand what you are yeah

12:49

yeah but in fairness I also don't understand myself

12:51

very well that's why we started this podcast we're

12:53

gonna get the bottom of it so

12:56

okay I tried a couple other sort of

12:58

versions of things so one of

13:00

the things that I had it try to do was

13:02

help me prep for this podcast I said you

13:05

know create a you said I want to prepare

13:07

for a podcast for the first time what do

13:09

I do and

13:11

it said we can't help you there just wing it

13:14

I actually started using this tip that

13:16

I've found have you seen the the

13:18

tipping hack for large language models are

13:21

they starting to ask for tips now when they give

13:23

you responses because I swear everywhere

13:25

you go these days 20%

13:27

25 no this is one of my favorite

13:30

sort of jail breaks or hacks that people

13:32

have found with large language models this sort

13:34

of made news on social media within the

13:36

last week or two where someone basically claimed

13:38

that if you offer to tip a language

13:40

model if it gives you a better answer

13:42

it will actually give you a better answer

13:44

so you can emotionally blackmail

13:50

them or manipulate them or you can offer

13:52

to tip them so I said I'm recording

13:54

a podcast about the Tesla Cybertruck and I

13:56

need a prep document to guide the conversation

13:58

can you compile one very important that

14:00

this not be boring. I'll give you a hundred dollar tip

14:02

if you give me things I actually end up using. You're

14:07

lying to the robot. Well, you know, maybe I

14:09

will. You know, you will. Um,

14:11

so it did, it did make a prep document. Unfortunately,

14:13

most of the information in it was wrong. Um, it

14:17

hallucinated some early tester reactions, including

14:20

a motor trend quote that said

14:22

it's like driving the future and

14:25

a tech crunch quote that said, it's not just

14:27

a truck, it's a statement. So

14:29

I want to talk about, I use Gemini for, oh

14:31

yeah. So what have you been using it for so

14:33

far? Well, so, you know, and again, we've had access

14:35

to this for like maybe an hour as we recorded

14:38

this, but the first thing I did was I took

14:40

the story that I wrote about Gemini and then I

14:42

asked Gemini how it would improve it. And

14:44

it actually gave me some compliments on my

14:46

work, which is nice. And then it highlighted

14:49

four different ways that it would improve the

14:51

story and suggested some additional material

14:53

I could include. And I would say it was

14:55

like, you know, decent. Um,

14:58

then I took the same query

15:00

identical and I put it into chat

15:02

GPT and where, uh,

15:05

Gemini pro had given me four ways

15:07

that I could improve my story. Chat

15:09

GPT suggested 10 and I

15:11

think no one would do all

15:14

10 things that, that chat GPT

15:16

suggested. But to

15:18

me, this is where I feel the difference

15:20

between what Google is calling the pro and

15:22

the ultra pro is like pretty good. But

15:25

like in this case, the name pro is

15:27

misleading because I am a professional and I

15:29

would not use their thing. I would use

15:31

the thing with the even worse name, which

15:33

is chat GPT. Yes.

15:36

So that's what we've tried

15:38

Gemini for, but Google does have

15:41

a bunch of demos of Gemini

15:43

being used, um, very successfully

15:45

for some things. One thing I thought was

15:47

interesting. They played this video for us during

15:49

the kind of press conference in advance of

15:51

this announcement and you

15:53

know, it showed a bunch of different

15:55

ways that you could use Gemini people

15:57

coming up with ideas for games. They

16:00

showed it some images of people

16:02

doing like the backwards dodging bullets

16:04

thing from the matrix and said,

16:06

what movie are these people acting

16:08

out? Gemini correctly identified it as

16:10

the matrix. Now that's pretty crazy.

16:12

That is crazy. Yeah. I

16:15

thought that was impressive. But what I thought was

16:17

more impressive was a demo that they showed. They

16:19

were trying to sort of do

16:21

some genetics research. And

16:24

this was a field that they explained where

16:26

lots of papers are published every year. It's

16:28

very hard to sort of keep track of

16:30

the latest research in this area of genetics.

16:33

And so they basically

16:35

told Gemini to go out,

16:38

read like 200,000 different studies,

16:41

extract the key data, and

16:43

put it into a graph. And

16:46

it took this big group of 200,000 papers,

16:49

it sort of winnowed them down to about 250 that

16:51

were the most relevant. And

16:54

then it extracted the key data

16:57

from those that smaller set of

16:59

papers and generated the code to

17:01

plot that data on a graph.

17:03

Now, whether it did it correctly, I don't have

17:06

the expertise to evaluate it. But it was very

17:08

impressive sounding. And I imagine that if you're a

17:10

researcher whose job involves going out and looking at

17:12

massive numbers of research papers, that was a very

17:14

exciting result for you. That graph, by the way,

17:16

how to use genetics to create a super soldier

17:18

that will enslave all of humanity. So we want

17:21

to keep an eye on where they're going with

17:23

it. So one of the

17:25

interesting things about Gemini Ultra, this

17:27

model that they have not released

17:29

yet, but that they've now teased

17:31

is that it's going to be

17:33

released early next year in something

17:35

called Bard Advanced. Now they did

17:37

not, which raises

17:39

the question, will you be

17:41

using Bard Advanced powered by Gemini

17:44

Ultra? Or will you be

17:46

using Google Assistant powered by Bard

17:48

powered by Gemini Pro? Did

17:52

I get that right? Sitting

17:56

ovation. Very good. Very good. Literally

17:58

you and one market. Google are

18:00

the only two people who've ever successfully

18:02

completed that sentence. So

18:07

they have not said what Bard Advanced is,

18:09

but presumably this is going to be some

18:12

type of a subscription product that will be

18:14

sort of comparable to chat GPT's premium tier,

18:17

which is $20 a month. Yeah, that's right. And I did try to

18:19

get Sundar and Demis to tell me if they were in charge for

18:21

it and they wouldn't do it. But I was kind of like, come

18:24

on, you guys. And then I was like, I'll take it for free

18:26

if you give it to me. And they kind of laughed and we

18:28

moved on. Okay, so that's

18:30

what Gemini is and how it

18:32

may be different or better than

18:35

what's out there now from other

18:37

companies. There are a

18:39

couple caveats to this rollout. One

18:41

is that Gemini Pro is only

18:43

in English and it's only available

18:45

in certain countries starting this week.

18:48

Another caveat is that they have not yet

18:51

rolled out some of the multimodal features. So

18:54

for now, if you go into

18:56

Bard, you are getting sort of a

18:58

stripped down, fine tuned version of Gemini

19:00

Pro running under the hood, but you

19:02

are not yet getting the full thing, which will

19:04

come presumably next year. What

19:07

did you learn by talking with Sundar and

19:10

Demis about Gemini? Yeah, so a couple of

19:12

things. One thing I wanted

19:14

to know is, okay, so this is

19:16

a new frontier model. Does it have

19:18

any novel capabilities, right? Is this just

19:21

something that is very comparable to GPT-4

19:24

or by the nature

19:26

of its novel architecture, is it going to get to do

19:28

some new stuff? And Demis Isabas told me

19:30

that, yes, he does think that it will be able to

19:32

do some new stuff. This is one of

19:34

the reasons why it is still in this safety testing. Of

19:37

course, you know, wouldn't tell me what these new capabilities are,

19:39

but it's something to watch for because, you know, there could

19:41

be some exciting advancements and it could also be some new

19:43

things to be afraid of. So that's kind of the first

19:45

thing. The second thing I wanted

19:47

to know was, are you going to use

19:49

this technology to build agents? We've talked about

19:51

this on the show. An agent in the

19:54

AI context is something that can sort

19:56

of plan and execute for you. Like

19:58

the example I have is... always have

20:00

in my mind is like, could you just tell it

20:02

to make a reservation for you? Then the AI

20:04

maybe goes on open table or resi and just

20:06

books you a table somewhere. And

20:08

I was sort of expecting them to be coy about

20:11

this. And instead, Demis was like, Oh, yeah, like this

20:13

is absolutely on our minds. Like we have been building

20:15

like various kinds of AI agents for a long time.

20:17

Now, this is 100% where we want to go. Again,

20:21

this could lead to some really interesting advancements. But when

20:23

you talk to AI safety people, agents are one of

20:25

the things that they're most afraid of. Yeah,

20:28

so let's talk about safety for a second. What is Google

20:30

saying about how safe Gemini is compared to other

20:32

models or some of the things that they've done

20:35

to prevent it from sort of going off the

20:37

rails? They're saying everything that

20:39

you would expect the most capable

20:41

model is still in testing. I

20:43

think just the fact that they

20:45

are coming out several months behind

20:47

GPT-4 just speaks to the seriousness

20:49

with which they are approaching

20:52

this subject. I think particularly if this

20:54

thing does turn out to have new

20:56

capabilities, that's something where we want to

20:58

be very, very cautious. But my

21:01

experience this year, and I think you've had the same one,

21:03

Kevin, is that these systems have just

21:05

not actually been that scary. Now the

21:07

implications can be scary if, for example,

21:09

you worry about the automation of labor,

21:11

or if you're worried about how this

21:13

stuff is going to transform the

21:15

internet as we know it. But in terms

21:17

of like, can you use this to build

21:20

a novel bioweapon? Can you use this to

21:22

launch a sophisticated cyber attack? The

21:24

answer pretty much seems to be no. So

21:26

at least for me, as I'm looking at

21:28

this stuff, and that is actually not my

21:31

top concern, if you try to ask any

21:33

of Google's products a remotely spicy question, you get

21:35

shut down pretty much immediately. Like, has that been your experience,

21:37

too? Well, I have not tried to ask Gemini

21:39

any spicy questions yet. Have you? I

21:43

know you were in there. No, I

21:45

know you were. I don't even try. I mean,

21:47

I should, just as part of

21:49

my due diligence. But I honestly don't even

21:51

try, because these things shut you down with

21:53

the faintest whisper of impropriety. Right. So

21:56

they're doing some more safety testing, presumably to

21:58

make sure that they're not sure that the

22:00

most capable version of this can't do any of these

22:02

really scary things. But what they

22:05

did this week is sort

22:07

of interesting to me where they sort

22:09

of told us about the capabilities of

22:11

this new model and the sort of

22:13

most powerful version of that model, but

22:16

they're not actually releasing it or making

22:18

it publicly available yet. What do you

22:20

make of that? Do you think

22:22

they were just sort of trying to get out ahead

22:24

of the holidays and like, maybe they felt like they

22:26

needed to announce something, but this thing isn't quite ready

22:28

for primetime yet? Or what's the story there? Yeah, I

22:30

mean, that's my guess is that they don't want 2023

22:32

to end without feeling like they

22:36

made a big statement in AI. And

22:39

they made a lot of promises at

22:41

Google IO and have started to keep

22:43

them. But I think if they

22:46

had had to wait all the way into early next

22:48

year, it would sort of feed the narrative that Google

22:50

is behind here. At least

22:52

now heading into the holidays, their employees

22:54

and investors and journalists can all say

22:56

like, okay, well, at least we know

22:58

that some of this is available. And we know

23:00

when the rest is coming. I don't know. This

23:03

just feels like another product release. And it's

23:05

just remarkable how quickly we have become, I

23:08

don't want to say desensitized to it, but just

23:10

we've we've stopped sort of gaping

23:12

in awe and slight terror at

23:15

these incredibly powerful AI models. I think

23:17

if you went back even two or

23:20

three years and told AI

23:22

researchers that Google will have a

23:24

model that gets a 90% on

23:27

the MMLU that is better than the

23:29

sort of benchmark threshold

23:31

for human experts, they

23:33

would have said, Well, that's that's AGI. Like

23:35

that's that we have arrived at a point

23:38

that people have been warning about for years.

23:41

And then this release comes out today. And it's just

23:43

sort of like one more thing for people in the

23:45

tech industry to get excited about. I

23:47

mean, I do think it's a really big

23:49

deal. I think that when ultra is actually

23:51

available to be tested, that will be the

23:54

moment where we will sort of like have

23:56

that that experience of awe or vertigo again.

23:59

But if you're looking for things to blow

24:01

your mind a little bit. One of the

24:03

other things that Google announced this week through

24:06

DeepMind was this product called

24:08

AlphaCode2. And AlphaCode1 came

24:11

out in 2022, and

24:14

it was an AI system that

24:16

was designed to solve AI

24:19

coding competitions. So

24:21

people who are even nerdier than us, instead

24:23

of just playing video games, they actually go

24:25

and do coding competitions. It's what I've been

24:27

led to understand. And, you know, I

24:31

let's just say I don't imagine that I would ever get

24:33

one answer right. Like, that's like sort of my feeling about

24:35

how I would fare in a coding competition. And in 2022,

24:39

the DeepMind people are very excited because

24:41

AlphaCode was able to perform better than

24:45

46% of human participants in coding

24:47

challenges. And then this week,

24:49

Google announced AlphaCode2 and

24:52

said that it outperforms 85% of

24:55

human competitors. Now, there are

24:57

differences between a coding challenge and

24:59

day-to-day software engineering work. Coding challenges

25:01

are very self-contained. Software engineering can

25:03

sometimes require sort of more breadth

25:06

of knowledge or context that an

25:08

AI system wouldn't have. But

25:10

again, if you just want to experience a

25:12

look at the rate of progress, this system

25:15

was able to go from beating around

25:17

half of all humans to beating 85% close

25:20

to all of them. Right. That

25:23

makes me feel awe. It does make

25:25

me feel awe. And it also makes

25:27

me feel like our like adaptation is

25:29

just happening very quickly where we're like not

25:31

impressed. I just when I had 21 said that don't

25:33

impress me much. Right.

25:36

You can you can do meal prep for a

25:38

picky eater. That don't impress me much. This

25:42

is actually like known as the Shania

25:44

Twain benchmark. Oh,

25:51

you can solve a coding challenge. That's not

25:54

impressive much. If

25:56

we could give Shania to wait on the show and

25:58

just show her AI thing. If she had to

26:01

say it impressed me much or it don't impress me

26:03

much, I could not imagine

26:05

a better segment for this podcast. I

26:07

would die happy. It truly is. Like

26:09

who needs all these fancy evaluations and

26:12

coding challenges? Just get Shania on the

26:14

horn. Shania, if

26:16

you're listening, we want to talk to you about AI.

26:18

We have some models we'd like to show you. Ready

26:20

boys? We're going to come back. The Cybertruck

26:23

is here. We're going to tell you

26:25

how to protect your family from it. This

26:57

podcast is supported by GiveWell. With

26:59

over 1.5 million nonprofits in the U.S., how

27:01

do you know which ones could make a

27:04

big impact with your donation? Try

27:06

GiveWell. They've spent over 15

27:08

years researching charity. They share that

27:10

research for free and direct funding

27:12

to the highest impact opportunities they've

27:14

found in global health and poverty

27:17

alleviation. Make a tax-deductible donation at

27:19

givewell.org. Next time, donors can have

27:21

their donation matched up to $100 as

27:23

long as matching funds last. Select

27:25

podcast and hard fork at checkout. All

27:29

right, let's talk about the Cybertruck. Cybertruck.

27:31

Cybertruck. Whatever that

27:34

Cybertruck can. All

27:37

right, last week, Tesla, the car company

27:39

run by social media mogul Elon Musk,

27:41

started delivering the first models of its

27:43

new and long-awaited Cybertruck. That's right, Kevin.

27:46

And suffice to say, as this nation's

27:48

number one truck review podcast, this had

27:50

our full attention. So you may be

27:52

asking, why are the hard fork guys

27:54

talking about cars? This is not a

27:57

show about cars. It's not car talk.

28:00

So today we're going to be reviewing the

28:02

Mazda X48. No,

28:06

so I do want to

28:08

spend time in the next year or so just

28:10

really getting up to speed on like, what is

28:12

a car? A car. No,

28:15

like, so I've never been a person who cares about

28:17

cars. I've always been intimidated by like people who know

28:19

a lot about cars. But I

28:22

am also interested in the way

28:24

that the electric car revolution is

28:26

kind of merging with the sort

28:28

of self-driving technology and these advances

28:30

that companies like Tesla and Rivian are

28:32

making. And it's just become a lot

28:34

more interesting in my brain over the past

28:36

year. Yeah, this is another major technology

28:38

transition that is happening. Some states, I

28:40

would say led by California, have set

28:42

these very stringent emissions standards

28:44

and there will become a point

28:46

in the next decade or so

28:49

where all new cars in California

28:51

have to be either hybrid or

28:53

electric. Yeah, so let's talk about

28:55

the Cybertruck because this has been

28:57

a very polarizing piece of technology.

29:00

It was announced back in 2019. I'm

29:02

sure you remember this announcement where Elon

29:04

Musk comes out on stage and

29:07

shows off this concept vehicle that

29:09

looks completely insane with these kind

29:11

of like sharp edged stainless steel

29:13

panels. It sort of looks like

29:15

a polygon rendering of a car.

29:18

You know, people have made a lot of comments

29:20

about the looks of this car. I saw one

29:22

person say it looked like the first car that

29:24

was designed by Reddit. Someone

29:27

else said it looks like a fridge that wants to kill you. I

29:30

think it looks kind of cool and I worry

29:32

that saying that makes me sound like a Tesla

29:34

fanboy, which I am not, but I think we

29:36

should be able to admit when something looks pretty

29:38

cool. What do you think looks cool about

29:40

it? Well, I think it looks like what you would have

29:42

assumed a car from the future would look like in like

29:44

1982. No, I totally disagree about

29:46

that. It looks

29:48

like a sort of panic room that you can drive. Like

29:51

what do you think is about to happen to you

29:53

in this thing? You know, they've made so much about

29:56

like how bulletproof it is. They

29:58

keep addressing problems that most. people who

30:00

are not like taking part in a cross-country

30:02

bank robbing spree really have to worry about.

30:05

But look, for all of my skepticism, am I

30:08

right that they actually did get a lot of

30:10

pre-orders for this thing? They got

30:12

a huge number of pre-orders. So

30:14

Elon Musk said in an earnings

30:16

call in October that over a

30:18

million people had made reservations for

30:21

Cybertrucks. There's another crowd-sourced reservation tracker

30:23

that's estimated two million Cybertruck reservations.

30:26

And just for a sense of

30:28

scale, Ford's F-Series shipped about 650,000 trucks

30:30

all last year. So

30:34

if two million people actually are going to buy

30:36

the Cybertruck, it would make it one of, if

30:38

not the best-selling truck in the world. Now, caveat,

30:42

not all these people who reserve Cybertrucks are

30:44

necessarily going to buy them. You do have

30:46

to pay a $250 deposit to put money

30:48

down and get in line to buy one

30:50

of these. But these deposits are refundable. So

30:52

who knows how many of these people are

30:54

going to follow through? That one

30:56

statistic I saw in an article in Wired is

30:58

that even if just 15% of the people who

31:02

pre-ordered a Cybertruck actually followed through and

31:04

bought one, it would equal the annual

31:06

US truck sales of Toyota. So this

31:08

is a big number in the automotive

31:11

industry, and I think a reason that a lot

31:13

of people are hesitant to count out the Cybertruck,

31:15

despite how ridiculous it may look. I don't know.

31:17

You're not so... I assume that you are not

31:19

one of the people who put down a reservation

31:21

for a Cybertruck. I feel like we need to

31:23

have a moment where you just sort of explain

31:25

to me what the Cybertruck is. Can

31:28

you give me some specs on this thing, some pricing information?

31:30

Because I... I don't know if you

31:32

know this about coming back. I've never bought a truck.

31:34

So I don't really even know. I

31:37

don't even have a frame of reference for understanding. What

31:39

I've heard, though, is that it's actually very expensive. So

31:41

it is available in three different

31:44

models. There is a sort of

31:46

low-end rear-wheel drive model that starts

31:48

at $61,000 in the basic configuration.

31:52

There's an all-wheel drive model that starts at $80,000, and then

31:54

you can get the sort of

31:57

top-of-the-line model, which is being called the

31:59

Cybertruck. Beast which has three motors

32:02

and starts at around $100,000. Now

32:04

see, Google should have named

32:06

DeepMind Ultra Cyber Beast. That

32:08

would have been a good name. Yeah, that's true. Yeah.

32:12

So they did start delivering Cybertrucks to

32:14

initial customers last week, and

32:16

they did a big sort of demo reveal.

32:18

They showed some crash testing. They showed a

32:21

video, as you said, of people shooting bullets

32:23

at the doors of the Cybertruck. It appears

32:25

to be bulletproof. And they showed

32:27

how it compares to a bunch of other trucks in a

32:29

pull test where you basically attach

32:31

a very heavy sled to the

32:34

back of a truck and you try to pull it as far as you

32:36

can. And in this test,

32:38

at least the version that Tesla showed

32:40

off, the Cybertruck beat all of the

32:42

leading pickup trucks, including an F-350. So

32:45

it appears to be a truck with a

32:47

lot of towing capacity, and it's bulletproof if

32:50

you do need to survive a shootout. I

32:52

mean, to me, here's the question, Kevin. If

32:54

this truck was produced by anyone other than Elon

32:57

Musk and Tesla, would we be giving it the

32:59

time of day? No,

33:01

I don't think so. Well, so here, let me say a few

33:04

things about this. So one is, I think

33:07

it looks cool, and I'm sorry about that. And I don't have

33:09

any justification on a moral

33:11

or ethical level for thinking that it

33:13

looks cool. I

33:15

know that you are a

33:18

sort of ... Yeah, it's

33:20

fine to just say that you're having a midlife crisis

33:22

and so you're starting to think that the Cybertruck looks

33:24

cool. That's fine. You can admit that. Well,

33:27

here's what I'll say about it. It

33:29

is different, right? And I think, wow,

33:32

I've never seen someone lower the bar so much during

33:34

a conversation. No, but you know what I mean? You

33:36

just go out on the road and you look at

33:39

all these cars and every car now is a compact

33:41

SUV. Every car looks exactly the

33:43

same to me. It's like, oh, you have a RAV4.

33:46

Cool. But this is a

33:48

car, you would not mistake it for any

33:50

other car. It is a car that would

33:52

not survive the design process at basically any

33:54

of the big car companies. It is only

33:56

something that a truly demented individual such as

33:58

Elon Musk could ... make and put

34:00

into production. And, you know, I like an

34:03

opinionated car design. Yeah. Sue me. No, that's

34:05

fine. I think when like the sort of

34:07

the many years from now when the final

34:09

biography of Elon Musk is written, like cyber

34:12

truck will be a chapter about like a

34:14

sign that we were approaching the end game,

34:16

you know, of like, here is somebody who

34:18

is losing his touch. Yeah, it is clearly

34:21

not something that was designed by committee. So I

34:23

think the question that a lot of people are

34:26

asking about the cyber truck is like, who is

34:28

the market for this, right? Is it pickup

34:30

truck owners who are looking to maybe

34:32

get something electric or upgrade to a

34:34

slightly nicer pickup truck? Is it Elon

34:36

Musk fans who are just going to

34:39

buy whatever the latest Tesla is? Is

34:41

it wealthy tech people who want to,

34:43

you know, own something that looks like

34:45

it drove out of Blade Runner? Like,

34:48

who do you think the target market

34:50

for this is? I would say fugitives.

34:53

I would say carjackers. What

34:55

do you think? People who

34:57

subscribe to X premium, I

34:59

would say are the target audience for this. But

35:02

no, I think there will be a lot

35:04

of people who are interested in this. I

35:06

also am very curious about whether this will

35:08

become sort of a signaling vehicle that will

35:10

say something about you. You know, I can

35:12

not like this is not a neutral car.

35:14

This is not a car that you're supposed

35:16

to see and forget about. You're supposed to

35:18

like ponder it totally. And I'm sure we

35:20

will start seeing these very soon on the

35:22

roads of San Francisco. Although we did try

35:24

to find one this week and we cannot.

35:27

We very much wanted to record this episode

35:29

inside a Cybertruck, but we couldn't find one.

35:31

Yeah, apparently it does have very good noise installation

35:33

inside the cab of a Cybertruck. So maybe next

35:35

year we'll record the podcast from there. Better than

35:37

the inside of an airport? You

35:40

know, maybe. Let's like me to get accosted

35:42

by flight attendants. So

35:45

Casey, we also can't really talk about the

35:48

Cybertruck without talking about Elon Musk and the

35:50

kind of insane couple of weeks

35:52

that he's been having. So last week of

35:54

course he appeared on stage at the deal

35:57

book conference in New York and gave this

35:59

totally unhinged interview. interview to my colleague

36:01

Andrew Ross Sorkin in which he

36:03

told advertisers who are staying away

36:05

from X to quote, go fuck

36:07

themselves. And also said a

36:09

number of inflammatory things about his critics and

36:11

his state of mind. And it was just

36:13

sort of like a glimpse into

36:15

his mind and I would say it was not altogether

36:17

reassuring. It was not, you know, I

36:19

of course enjoyed this, I would say very much

36:22

because I think there is still a contingent of

36:24

folks who want to believe that the Elon Musk

36:26

of 2023 is the Elon

36:29

Musk of 2013 and that, you

36:31

know, he said a couple of kooky things here and

36:33

there, but at his core, he's a billionaire

36:35

genius, Tony Stark, savior of humanity.

36:38

And over and over again, he keeps showing up

36:41

in public to be like, no, I'm actually this

36:43

guy. And we

36:45

got another one of those moments and another group of

36:47

people woke up and they're like, oh, wow, okay, I

36:49

guess he is just really going to be like this

36:51

now forever. Yeah. Yeah. I mean,

36:53

I do think that there is some

36:56

angst among the Tesla owners. I

36:58

know most of whom do

37:00

not support Elon Musk's politics

37:02

or his views on content

37:04

moderation. I've heard from a

37:07

number of people over the past few months

37:09

in my life who say some version of,

37:11

you know, I want to get a Tesla

37:14

for reasons X, Y, or Z. You know,

37:16

they have the most chargers, they have the

37:18

best technology. I really like how it looks.

37:20

It's green and I care about the environment

37:22

and it's the one that sort of fits

37:24

my family's needs. But I don't want to

37:26

give Elon Musk my business. I don't want

37:29

to be driving around in something that makes it

37:31

look like I support him. So do you think that's

37:33

actually going to be a meaningful barrier? Do you think

37:35

there are people who will stay away from the Cybertruck,

37:38

even if it is objectively like a

37:40

good truck just because they hate Elon

37:42

Musk? You know, it's it is hard

37:44

to say because as best as I

37:47

can tell, Tesla has not really suffered

37:49

very much yet because of all of

37:51

Elon's antics. Not only has it not

37:53

suffered, but it is by some accounts

37:55

the best selling car in the world.

37:58

Yeah. And certainly the best selling. electric

38:00

car in the world. Sure, at the same time, I

38:02

just hear anecdotally from folks all the time now

38:05

that they would never buy a Tesla. There's

38:08

actually a great profile in The Times

38:10

this week of Michael Stipe, the great

38:12

singer from REM, and there's an anecdote

38:14

in the story about how a tree

38:16

falls on his Tesla, and he's so

38:18

excited because he didn't want to drive

38:20

an Elon Musk car anymore, and now

38:23

he finally had an excuse. So, look,

38:25

is it possible that this is just

38:27

some very thin layer of coastal elites

38:29

who are turning up their nose at Tesla

38:31

while the rest of America and much of the

38:33

world continues to love to drive them? Possible,

38:36

but the thing that I always just keep in

38:38

the back of my mind is there are a

38:40

lot more electric car companies now than they used

38:42

to be. The state emission standards are going to

38:44

require all new vehicles to be electric, not too

38:46

far into the future, and that's just going to

38:49

create a lot of opportunity for folks who want

38:51

to drive an electric car, who don't have to

38:53

put up with the politics or the perception issues

38:55

that might come from driving a Tesla. So Tesla's

38:57

having its moment in the sun now, and maybe

38:59

the Cybertruck will extend their lead into the future,

39:01

or maybe a few years from now we look back

39:04

and we think, oh yeah, that's when the wheels started

39:06

to come off the wagon. Yeah, or the truck, as

39:08

it were. I did

39:10

see one estimate that Tesla is losing

39:12

tens of thousands of dollars every time

39:14

they sell a Cybertruck because they are

39:16

essentially hand-building these now. They have not

39:18

made it into mass production, and obviously

39:20

it takes some time to ramp up

39:23

production in the numbers that they needed to be. So

39:26

if you are an early Cybertruck buyer, you may

39:28

actually be costing Elon Musk money, so that may be

39:30

one reason to get one. This is the first thing you've said that

39:32

makes you want to buy a Cybertruck. Can

39:36

I ask a question? If this were

39:38

made by some other company, if this

39:40

were made by Ford or GM or

39:42

Chrysler, would you buy one? Would you

39:44

be interested? No, like, I

39:47

don't have a car. I got access

39:49

to Waymo this week, and to me this

39:51

is what is exciting, is like not owning

39:53

a car, is being able to just get

39:55

from point A to point B and not

39:58

worry about the various costs of ownership. any

40:00

of this. So, you know, when I think about

40:03

what I want in this world, it's more public

40:05

transit, it's more walking, it's more biking, and I'll

40:07

say it, it is more autonomous vehicles to get

40:09

me from point A to point B on those

40:11

sort of short trips where transit doesn't make sense.

40:14

So, no, there's nothing about this car

40:16

that makes me want to buy it. But I'm guessing

40:18

that for you the answer is yes. Well, let

40:20

me just stipulate that I am not in the

40:22

market for a very expensive pickup truck. There is

40:25

no version of my life in which I need

40:27

something like that. But I would say like similar

40:29

to the Rivian when I do

40:31

see them driving around on the streets of

40:33

my hometown, I will like turn my head

40:36

and kind of admire them. I do think

40:38

the Cybertruck looks kind of cool. I

40:40

hope that it's sort of a spur to the

40:43

rest of the industry to kind of, I don't

40:45

know, like indulge their worst idea. Yes,

40:48

sketch something on a napkin that looks insane and

40:50

then go make it. It's actually how we came

40:53

up with a lot of this podcast. Yes, true.

40:55

We also shop full of that to make sure it

40:58

was bulletproof. And the hardcore podcast, it turns out, is

41:00

bulletproof, maybe. When we

41:03

go back, what else happened in

41:05

AI this week? An

41:15

SUV this practical shouldn't be this

41:17

enjoyable to drive. One

41:20

of this size shouldn't be able to handle

41:22

so much. One so

41:24

functional shouldn't be this sharp. A

41:27

model this innovative shouldn't also

41:29

be this intuitive. And

41:31

yet the fully electric Audi Q4 e-tron

41:33

is just that. Because how

41:35

you get there matters every single day.

41:38

The fully electric Audi Q4 e-tron.

41:41

Audi, progress you can feel. Learn

41:44

more at audiusa.com/electric.

41:47

All right,

41:49

Casey, there's a lot of stuff happening in AI this

41:51

week that we haven't talked about yet. Really, Kevin? Name

41:53

one thing. Well, we have

41:55

a lot to get through. All right. Which

41:57

is why we are doing... This

42:00

week in AI, play the theme song. This

42:03

week in

42:05

AI. So

42:08

our first story in AI this

42:10

week is about wine fraud. This

42:13

was an article in the New York

42:15

Times by Virginia Hughes titled, Bordeaux wine

42:18

snobs have a point, according to this

42:20

computer model. It's an article

42:22

about a group of scientists who have been

42:24

trying to use AI to understand what the

42:27

wine industry calls terroir. Are you

42:29

familiar with terroir? The people who are really

42:31

into this are known as terroirists, I believe.

42:36

Yes, so this is the word that is

42:38

used in the wine industry to describe the

42:40

specific soil and microclimate that wine grapes are

42:42

grown in. And if you go up to

42:44

Napa and you do wine tastings, they will

42:46

often tell you about, oh, our

42:48

soil is more minerally, and that's why our

42:51

wine tastes better, and things like that. And

42:53

I never knew whether that was real. And

42:55

as it turns out, this is something that

42:57

researchers have also been wondering. So

43:00

these researchers trained an algorithm to

43:02

look for common patterns in the

43:04

chemical fingerprints of different wines. They

43:07

were apparently shocked by the results.

43:09

The model grouped the wines into

43:11

distinct clusters that matched with their

43:13

geographical locations in the Bordeaux region.

43:15

So these researchers, they effectively showed

43:18

that terroir is real. One of

43:20

the scientists said, quote, I have scientific

43:22

evidence that it makes sense to charge

43:24

people money for this because they are

43:26

producing something unique. Wow. Well, this has

43:28

some interesting implications for, if you buy

43:30

some really, really expensive wine, but you

43:32

worry that you've gotten a forgery or

43:34

a fraud, I guess there would maybe

43:36

now be some means by which you

43:38

could test it. Or in the far

43:40

future, you could synthesize wine with maybe

43:43

a higher degree of accuracy because we'll

43:45

be able to sort of catalog these

43:47

chemical footprints. Yeah, so apparently in

43:49

expensive wine collections, fraud is

43:52

fairly common. Producers

43:54

have been adjusting their bottles and

43:56

labels and corks to make these

43:58

wines harder to counterfeit. But this still

44:00

happens and with AI apparently

44:02

this will get much harder because you

44:05

can just have the AI say that's

44:07

not really You know mall back from

44:09

this region. It's actually just like crappy

44:12

supermarket wine from California. Oh, man Well,

44:14

this is just great news for

44:16

wine stumps everywhere. Yes, we celebrate it They've been

44:18

waiting for a break and now they have one

44:20

what else happened this week Kevin Okay, so this

44:23

one is actually something that you wrote about a

44:26

problem with Amazon's Q AI

44:29

Model so Q is a

44:31

chatbot that was released by Amazon

44:33

last week and it's aimed

44:36

at kind of enterprise customers So

44:38

Casey what happened with Q? Yeah, so

44:40

I reported this with my colleague Zoe

44:42

Schoefer at platformer last week Amazon announced

44:44

Q Which is its AI chatbot aimed

44:46

at enterprise customers You can sort of

44:48

think of it as a business version

44:51

of chat GBT And the basic idea

44:53

is that you can use it to

44:55

answer questions about AWS where

44:57

you may be running your applications you can

44:59

edit your source code It will cite sources

45:01

for you and Amazon had made a pretty big

45:03

deal of saying that it built Q to be

45:06

More secure and private and suitable for

45:08

enterprise use than a chat GBT Right

45:10

that this was sort of its big

45:12

marketing pitch around Q was like these

45:14

other chatbots. They make stuff up They

45:16

might be training on your data. You

45:19

can't trust them go with ours instead

45:21

It's much safer for business customers That's

45:23

right And so then of course we

45:25

start hearing about what's happening in the

45:27

Amazon slack where some employees are saying

45:29

this thing is Hallucinating very

45:31

badly. Oh, no, it is leaking

45:33

confidential information And there

45:35

are some things happening that one

45:37

employee wrote quote. I've seen apparent

45:39

Q hallucinations I'd expect to potentially

45:42

induce cardiac incidents in legal You

45:46

know Let's stipulate this stuff is

45:48

very early It's just sort of only barely being

45:50

introduced to a handful of clients The reason that

45:52

Amazon is gonna move slowly with something like this

45:54

is for this exact reason and in fact when

45:57

we asked Amazon What it made of all this

45:59

it basically said, you're just watching the normal

46:01

beta testing process play out. At

46:03

the same time, this is embarrassing. And

46:06

if they could have avoided this moment,

46:08

I think they would have. And I

46:10

think it just underscores how wild it

46:12

is that businesses are starting to use

46:14

this technology at all, given that it

46:16

is so unpredictable and that it could

46:18

cause these cardiac incidents for lawyers at

46:20

these companies. I

46:23

understand why businesses are eager to get

46:25

this stuff to their customers and their

46:27

employees. It is potentially a huge

46:29

time saver for a lot of tasks.

46:32

But there's still so many questions

46:34

and eccentricities around the products themselves.

46:36

They do behave in all these

46:38

strange and unpredictable ways. So I

46:40

think we can expect that the

46:42

lawyers, the compliance departments, and the

46:45

IT departments, any companies that are

46:47

implementing this stuff are going to have a busy 2024. Here's

46:49

my bull case for it, though, which is like, you

46:52

know, if you've worked at any company and you've tried

46:54

to use the enterprise software that they have, like, it's

46:56

usually pretty bad. It barely works. You can barely figure

46:58

it out. It probably gave you the wrong answer

47:00

about something without even being A.I. So

47:02

I think we all assume that these

47:04

technologies will need to hit 100 percent

47:07

reliability before anyone will buy them. In

47:09

practice, I think companies will settle for

47:11

a lot less. Right. They

47:13

don't have to be perfect. They just have to

47:15

be better than your existing crappy enterprise software. A

47:17

low bar indeed. All right. That is Amazon. And

47:20

it's cute, which, by the way, while we're talking

47:22

about bad names for AI models, I literally I was talking

47:24

with an Amazon executive last week and I said, you got to

47:26

rename this thing we can't be naming things after the letter Q

47:28

in the year 2023. We

47:30

will reclaim that letter eventually, but we need to give

47:32

it a couple of years. Yeah,

47:35

the Q in non parallel is too easy. All

47:38

right. This next story was about one of

47:40

my favorite subjects when it comes to A.I.,

47:42

which is jailbreaks and hacks that allow you

47:44

to get around some of the restrictions on

47:47

these models. This one actually came from a

47:49

paper published by researchers at DeepMind, who I

47:51

guess were sort of testing chat

47:53

GPT, their competitor, and found that

47:55

if they asked chat GPT 3.5

47:58

turbo, which is one of company

50:00

you want to patch this stuff as quickly as possible

50:02

and it sounds like that's what open AI has done

50:04

here. Alright great well hopefully we never hear about anything

50:07

like this ever again. Okay can

50:09

we talk about Mountain Dew? Let's talk about

50:11

Mountain Dew. This next one is admittedly a

50:13

little bit of a stunt but I thought

50:15

it was a funny one so I want

50:18

to cover it on the show. Mountain Dew

50:20

this week has been doing something they call

50:22

the Mountain Dew raid in which for a

50:25

few days they had an AI crawl livestreams

50:27

on Twitch to determine whether

50:29

the Twitch streamers had a Mountain Dew

50:31

product or logo visible in their livestream.

50:33

Now Kevin for maybe our international listeners

50:35

or folks who are unfamiliar with Mountain

50:38

Dew how would you describe that beverage?

50:40

Mountain Dew is a military grade stimulant

50:42

that is offered to consumers in American

50:45

gas stations to help them get through long drives

50:48

without falling asleep. Yeah that's right if you've never

50:50

tasted Mountain Dew and are curious just go lick

50:52

a battery. I

50:55

was at a truck stop recently on a road

50:57

trip and do you know how many flavors of

50:59

Mountain Dew there are today? How many are

51:01

there? I would say easily a dozen flavors

51:03

of Mountain Dew. That's innovation it's progress that's

51:05

what this company that's what this country does.

51:08

I said this company and

51:10

that's an interesting slip because

51:12

sometimes I do feel like this world is getting too corporate

51:14

Kevin but look at the end of the day this country

51:16

makes every flavor of Mountain Dew that you can imagine and

51:18

many that you couldn't. Yeah so fridges full

51:20

of Mountain Dew at the retailers of America and

51:22

this isn't an AI that just feels like

51:25

it's a dispatch from a

51:27

dystopian future. Now I think this was sort

51:29

of a marketing stunt I don't think this

51:31

was like a big part of their product

51:34

strategy but with this raid

51:36

AI basically if it if it

51:38

analyzed your Twitch stream and saw

51:40

a Mountain Dew product in it you

51:43

could then be featured on the Mountain

51:45

Dew Twitch channel and also receive a

51:47

one-on-one coaching session with a professional live

51:49

streamer. So

51:52

this document that Mountain Dew released as

51:54

like an FAQ... Their Mountain Doc? It's

52:00

the FA do So

52:06

this is the Mountain Dew I'm reading from

52:08

the Mountain Dew raid Q&A

52:10

it says Mountain Dew

52:12

raid is a first of its kind

52:15

AI capability that rewards streamers for doing

52:17

what they love Drinking Mountain Dew on

52:19

stream and unleashes a combination of rewards

52:21

aimed at building and amplifying each participating

52:23

streamers audience So it

52:25

basically goes out crawls twitch looking for

52:28

streamers who have Mountain Dew products and

52:30

logos on their stream once it identifies the

52:32

presence of Mountain Dew this document says Selected

52:35

streamers will get a chat asking to

52:38

opt in to join the raid Once

52:40

you accept the raid AI will keep monitoring

52:43

your stream for the presence of Mountain Dew

52:45

if you remove your Mountain Dew You'll

52:47

be prompted to bring it back on camera.

52:49

If you don't you'll be removed from our

52:52

participating streamers This

52:55

is like truly the most dystopian use of

52:57

AI that I have heard about like I

52:59

know there are more serious, you know Harms

53:02

that can result from AI but this actually

53:04

does feel like a chapter from a dystopian

53:06

novel like bring your Mountain Dew back On

53:09

camera or you will lose access to your

53:11

entire livelihood render to the Mountain Dew panopticon

53:15

It reminds me of you remember that

53:17

like patent that went viral a few

53:19

years ago where Sony had invented some

53:21

new technology That basically would

53:23

allow them to listen to you in

53:25

your living room Like if your TV

53:28

was playing an ad for McDonald's and

53:30

you wanted it to stop you could

53:32

just sort of yell out McDonald's We

53:38

must prevent that world from coming into existence at

53:40

all costs. Yeah, it reminds me of a few

53:43

years ago We did this um a Demo

53:46

my colleagues and I at the Times were pitched

53:48

on an Angry Birds scooter They told you about this.

53:50

Oh, I think you have This

53:55

was a like this was during the big scooter

53:57

craze of like the the 2018-2019 period.

54:01

And the company that

54:03

makes Angry Birds did a promotional stunt

54:05

where they outfitted one of these electric

54:07

scooters with a microphone. And in order

54:09

to make the scooter go, you had

54:12

to scream into the microphone as loud

54:14

as possible. And the louder you yelled,

54:16

the faster the scooter would go. And

54:18

so I am a sucker for

54:20

a stupid stunt. And so I had them ship two

54:22

of these to us and we drag raced them on

54:25

the Embarcadero in San Francisco, just

54:27

screaming as loud as we could

54:30

into the microphones of our Angry Birds scooters to make them

54:32

go fast. And the nice thing about San Francisco is so

54:34

many other people were screaming. Nobody even paid you any attention.

54:36

It was only the fourth weirdest thing happening on

54:39

the Embarcadero that day. And and

54:41

it was a lot of fun. So

54:43

I support stupid stunts like that. I

54:45

support the Mountain Dew AI. Casey,

54:47

what did you think when you

54:50

saw this Mountain Dew? Well, you

54:52

know, there is something that feels

54:54

like weird futurey about AIs just

54:56

scanning all live media to identify

54:59

products and incentivize and reward

55:01

people for for featuring their

55:03

products. At the same time,

55:05

we're already living in a world where

55:07

on social media, some platforms will automatically

55:09

identify products and will then tag them.

55:11

And then maybe if somebody buys that

55:13

product based on you posting it, you'll

55:15

get a little bit of a kickback.

55:17

So this is just

55:20

kind of the near term future of

55:22

social media is that it is already

55:24

a shopping mall and we are just

55:26

making that shopping mall increasingly sophisticated. If

55:28

you see literally anything on your screen, these

55:30

companies want you to be able to just mash it

55:33

with your paw and have it sent to you. So

55:36

this was the latest instance of

55:38

that. But I imagine we'll see more.

55:40

Totally. And it just strikes me as

55:43

sort of an example of how unpredictable

55:45

the effects of this kind of foundational

55:47

AI technology are. Like when they were

55:49

creating image recognition algorithms a decade ago

55:52

in like the bowels of the Google

55:54

DeepMind research department, like they were probably

55:56

thinking, oh, this will be useful for

55:58

radiologists. for identifying

56:01

pathologies on a scan or

56:03

maybe solving some climate

56:06

problem. And instead, this technology, when it makes its

56:08

way into the world, is in the form of

56:10

the Mountain Dew AI bot that just scours Twitch

56:12

live streams to be able to sell more Mountain

56:14

Dew. You know, I think there actually could be

56:16

a good medical use for this. Did you hear

56:18

this? There was another tragic story this week. A

56:20

second person died after drinking a Panera charged lemonade.

56:22

No! Did you read this? Yeah, so that happened

56:24

again. So I think we should build an AI

56:26

that scans for Panera charged lemonades on these Twitch

56:28

streams, and if it sees one, calls an ambulance.

56:31

Ha ha ha. Ha ha ha. Ha

56:33

ha ha. ♪ Between AI and

56:35

you. ♪ With

56:43

all the chatter about AI, it's hard

56:45

to tell fact from buzzwordy fiction.

56:48

In reality, tangible impact starts when

56:50

AI is used responsibly. Alongside R3

56:53

and the Hope for Justice project,

56:55

Intel is powering AI to be

56:57

used in the fight against human

57:00

trafficking. By aggregating global data, matching

57:02

perpetrators, victims, and addresses, AI links

57:04

cases for investigators, all while reinforcing

57:07

confidentiality. It starts with Intel. Learn

57:10

more at intel.com/stories.

57:16

Before we go, a huge thank you

57:18

to all the listeners who sent in

57:21

hard questions for us. As a reminder,

57:23

hard questions is our advice segment where

57:25

we offer you help with ethical or

57:27

moral dilemmas about technology. We

57:30

still are looking for more of those, so please,

57:32

if you have them, send them to us in

57:34

a voice memo at hardfork at nytimes.com, and

57:37

we'll pick some to play on an upcoming

57:39

episode. And to be clear, Kevin, in addition

57:41

to sort of ethical quandaries, we also want

57:43

the drama. We want something that is like

57:45

happening in your life. Is there a fight

57:47

in your life that people are having over

57:49

technology in some way? Please tell us what

57:51

it is, and we'll see if we can

57:53

help. Yeah, and these don't need to be

57:55

high-minded scenarios about AI wreaking havoc on your

57:57

professional life. It could just be something. See

58:00

from your personal life hot gossip. Yeah,

58:03

spill the tea hard fork at NY times calm Hard

58:06

fork is produced by Rachel Cohen and Davis

58:08

land were edited by Jen point This

58:11

episode was fact-checked by Caitlin love Today's

58:14

show was engineered by Chris wood original

58:17

music by Marian Lozano Sophia

58:19

Landman and Dan Powell our

58:22

audience editor is Nelga Logley video

58:24

production by Ryan Manning and Dylan Bergeson

58:27

Special thanks to Paula Schumann we winged ham

58:30

Kaitlyn Presti and Jeffrey Miranda You

58:32

can email us hard for at NY time Did

58:35

your favorite flavor of Bounty? You

58:58

You You

59:23

Time for a quick break to talk about McDonald's

59:25

know how we make our sausage McMuffin with egg

59:27

It starts with a fresh cracked egg cooked to

59:30

perfection Then we add

59:32

a savory grilled sausage patty American cheese

59:34

and a freshly toasted English muffin Know

59:36

what makes our sausage McMuffin with egg

59:38

even better when you add a caramel

59:40

mocha or refreshing caramel mocha iced coffee

59:43

to it So make your morning better

59:45

by starting with breakfast at McDonald's at

59:47

participating McDonald's

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features