Podchaser Logo
Home
Daniel Simons Reveals Why We’ll All Be Fooled Again

Daniel Simons Reveals Why We’ll All Be Fooled Again

Released Tuesday, 26th December 2023
Good episode? Give it some love!
Daniel Simons Reveals Why We’ll All Be Fooled Again

Daniel Simons Reveals Why We’ll All Be Fooled Again

Daniel Simons Reveals Why We’ll All Be Fooled Again

Daniel Simons Reveals Why We’ll All Be Fooled Again

Tuesday, 26th December 2023
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

Want to be more active this winter? Sierra

0:02

helps you save on everything from skis to

0:04

snow boots, sleds to scarves. And if you

0:06

still don't want to go outside, you can

0:08

always move somewhere warmer. Sierra,

0:10

let's get moving. To your local

0:12

store, like now. Go! You

0:16

can crush your fingers and all your toes during a

0:18

data center migration. You can

0:20

knock on wood, pluck a dozen four-leaf clovers, or

0:22

look to your lucky stars for a successful office

0:25

expansion. You could hold your breath,

0:27

shut your eyes, and say all the world wishes to

0:29

help avoid cyber attacks. But none

0:31

of that truly helps you. Because

0:33

next-level moments need the next-level network,

0:35

with the security, reliability, and expertise

0:38

to take your business further. AT&T

0:41

Business, the network more businesses

0:43

are choosing. Today

0:46

on What the Hack, join us as we

0:48

unravel the mysteries of deception and the subconscious

0:51

tendencies that make us all vulnerable. We

0:53

talk to Dr. Daniel Simons, co-author

0:55

of a new book called Nobody's

0:57

Fool, to learn why we get

0:59

scammed. And learn how to stay safe by

1:01

thinking the way a scammer thinks. Welcome

1:03

to What the Hack, a show about hackers, scammers,

1:06

and the people they go after. I'm

1:08

Adam Levin. I'm Beau Friedlander. And

1:11

I'm Travis Taylor. Today

1:23

we have one of the co-authors of the book,

1:25

Nobody's Fool, Why We Get Taken In, and What

1:27

We Can Do About It. Please welcome Daniel Simons.

1:29

Thanks for having me on. Oh, thank you so

1:31

much for coming. We're delighted to

1:33

have you here, and we're particularly excited to

1:35

talk to you about your new book, The

1:38

Psychology of Scams, and How People Can Better

1:40

Insulate Themselves from Falling for Scams. Dan, I

1:42

am interested in the nature of deception as

1:44

you discuss in your book. I mean, I

1:47

guess it can range from everything. There's a

1:49

wide, there's a

1:51

kaleidoscopic selection of

1:53

things that can happen, from

1:55

Ponzi schemes to political misinformation

1:58

to disinformation. What exactly? is

2:00

a scam? And is it different

2:02

from a con or, you know, other negative

2:04

things that can happen in the realm of

2:06

information? There's a wide

2:08

range, as you say, and what we're

2:10

more interested in broadly is the nature

2:12

of deception and how we get deceived.

2:15

And that can take the form of cons,

2:17

which we think of as sort of the

2:19

more grand sort of impersonation or pretending to

2:21

be somebody that you aren't or pretending to

2:23

do something you aren't. Scams

2:25

which might be much more

2:27

simplistic, trying to take people in in lots

2:29

of different ways, as well

2:32

as just deception, which might be completely inadvertent

2:34

where somebody might mislead somebody without even realizing

2:36

they've done it in the first place. So

2:39

I view cons as kind of a combination of scams

2:42

that last over a long time that are much

2:44

more complex. Okay, so what's the most common thing

2:46

that you come across in Nobody's Fool?

2:48

What do you really... What am

2:50

I going to walk away and say, okay, I

2:52

really understand blank. Okay,

2:55

so I think there's a key principle that we are trying

2:57

to get at in this book that we

2:59

think is missed by a lot of the

3:02

other coverage of scams, cons, deception, which

3:04

is that the vast majority of the

3:07

movies and podcasts and stories and articles

3:09

about scams focus on

3:11

the dynamics of the scam itself.

3:14

Right? So if you watch a con movie, you

3:17

are following the action almost as

3:19

if you're watching a narrative and a storyline and you know

3:21

from the outside what it is. What

3:23

we're trying to do is gather

3:25

information about deception in a wide,

3:27

wide range of contexts to

3:30

see if there's any commonalities about the people

3:32

who fall for it, not just in their

3:34

personality, but in terms of their thinking and

3:36

how they think about scams. So I

3:38

think the one... If there's one

3:40

big point here, it's that it's

3:43

really easy to look at a scam

3:45

from the outside and say, yeah, the

3:47

person who fell for it, they were gullible, they didn't

3:49

spot the red flags they should have, but

3:51

in reality, we all can be targeted

3:54

because all forms of deception take

3:56

advantage of how we think and reason by

3:58

default in naturally good ways. So

4:00

I think one thing that

4:03

we try and focus on is what is it

4:05

that leads us to be deceived by looking at

4:07

commonalities across many forms of deception as opposed to

4:10

going into depth and looking at why

4:12

somebody fell for a particular scam. So

4:15

is there a particular kind of person who falls

4:18

for things like this? Well,

4:20

it's a good question. There's certainly going

4:22

to be a range of skepticism versus

4:25

sort of openness or willingness to believe things,

4:27

right? And some people are going to be

4:30

more willing to fall for things than others

4:32

or more willing to accept as true something

4:34

that's not. But I

4:36

think one of the key insights here is that we

4:38

all can be targeted by scams and cons,

4:42

more often by scams and just deception. We all

4:44

can be targeted if they happen to tap into

4:46

the sorts of things we're interested in or that

4:48

we want. So any of us

4:51

can recognize something is too good to be true

4:53

if it's kind of implausible to us. But all

4:56

of us can also find something to be just good

4:58

enough to be plausible if it's framed

5:00

correctly. So what I'm hearing is

5:02

there is no like this. There are a panoply

5:06

of situations, and depending on where you are

5:08

in your day or your life, you may

5:10

or may not be susceptible. Exactly.

5:12

I mean, most people, for

5:14

example, are not going to respond to an email from

5:17

a Nigerian prince promising a share of their inheritance. Most

5:19

people won't. Some people will. I

5:21

always do, no. Yeah. Some

5:24

people will only because they want to kind of scam the

5:26

scammer. But a lot

5:29

of times, those sorts of scams, it

5:31

might appeal to somebody and they might

5:33

respond to it. But that same person, under

5:36

different conditions, hours later or a day later,

5:38

might not. And it really depends

5:40

on whether it catches people at the right moment

5:42

in the right way that they respond to it. Some

5:44

people will never respond to that sort of email because

5:46

they recognize it immediately as too good to be true.

5:50

But those same people could

5:52

be victimized by something that seemed completely

5:54

plausible to them and implausible to somebody

5:56

who would respond to a Nigerian prince

5:58

scam. In

6:02

terms of the research for the book, I'm wondering what

6:04

surprised you the most. For me

6:06

it was how consistent, across

6:08

all of these forms of deception, over thousands

6:10

of years, right, everything from the Trojan horse

6:13

on up, that they

6:15

all rely on a small set of principles

6:17

that are pretty consistent. So

6:20

I think that's a

6:22

positive, right? It means that if we can understand how

6:24

these sorts of deceptions take advantage of the

6:27

way we think and the way we reason

6:29

and the kinds of information we find appealing,

6:31

then as new variants of these things emerge,

6:34

we'll be able to better spot those consistencies. So

6:37

I think that's something that for me

6:39

was a bit surprising how well it holds

6:41

together when you look across everything from art

6:44

forgeries to chess cheaters, that

6:47

they all kind of rely on a similar

6:49

set of tactics. And the big ones, like

6:51

the long-running sort of scams and cons, rely

6:53

on a lot of them, whereas other

6:56

sorts of deception might just rely on one or two

6:58

at a time. Now, are we talking about hope, despair,

7:02

love, all the Shakespearean and

7:04

or Greek tragedy type stuff,

7:06

or what are we talking

7:08

about? I think hopes

7:10

and wishes and fears have to underlie most

7:12

of these sorts of things. They put us

7:15

in the right mindset to

7:17

be targeted. But what

7:19

we're talking about more is what leads us

7:21

to say, yeah, that's appealing.

7:23

Yes, I'm going to believe that and

7:26

not ask another question or check it out. So

7:28

we're focusing more on the thinking aspect of

7:31

it than on the motivation. So

7:33

we're not so much interested in the personality

7:35

of a con artist or of the victim.

7:37

We're interested in the cognitive mechanisms, how we

7:39

think, what our habits of thought are, and

7:43

what sorts of information we inherently find appealing,

7:45

often for really good reason. So can you walk

7:47

us through that small set of principles that make

7:49

scams work? So we divide

7:52

these sort of cognitive principles into what we

7:54

call habits, which are cognitive

7:56

tendencies, things we do really well

7:58

and they're really useful most of the time. the time,

8:01

but that can be turned against us, and

8:03

hooks, which are kinds of information that we find

8:05

inherently appealing, and that most of the time that's

8:08

a good thing. They draw our attention to the

8:10

things we care about, but that

8:12

again, if somebody's looking to deceive us, they

8:14

can present information that we find appealing when

8:17

it shouldn't actually be trustworthy. And

8:19

all of these sorts of cognitive

8:22

tendencies, habits, and hooks are

8:24

premised on the idea that we mostly accept what

8:26

other people are telling us as true.

8:29

We by default assume that we're interacting with people who are

8:32

being honest with us, that the things

8:34

we encounter in the world are truthful, and

8:36

it takes effort to stop

8:38

and think to yourself, well,

8:40

maybe I shouldn't believe that. Maybe that's

8:42

not true. Okay, so the first principle

8:44

is that the person we're

8:46

talking to or communicating with is for

8:49

real, and what they're saying is true.

8:51

Exactly. And we tend to do that by

8:53

default very quickly. And it takes effort

8:56

to kind of say, oh, wait a second. They're

8:58

potentially lying to me. And this

9:00

is true not just of other people, but of

9:02

everything we encounter in our daily lives. Dan,

9:05

it's true that this morning there was a

9:07

coyote in my upper field, and

9:10

he was limping with his back leg. And

9:13

when I got out of the car, he took off not limping. He

9:16

was trying to con me out of the car.

9:18

A little jerk. Yeah. So

9:21

we see it in wildlife all the time.

9:23

Scams have bruises in order to

9:26

get something to happen generally in

9:28

the terms of hunting. And I guess

9:30

scams are a form of hunting, aren't they? In

9:33

a lot of ways, yeah. And the key

9:35

element of that is that it makes sense to believe

9:37

it's true most of the time because cons

9:40

and deceptions and scams are relatively rare.

9:42

Most of us are never going to be a victim of a

9:45

large con. Most of us are

9:47

never going to fall for a Bernie Madoff's Ponzi scheme.

9:50

Most of us can't afford to fall for them. Sure.

9:53

A lot of us don't have that opportunity.

9:55

I don't have the ante. Yeah, exactly. I

9:57

mean, most of us are never going to try and purchase.

10:00

you know, really expensive fine art and buy a

10:02

forgery by mistake. And most of us aren't going

10:04

to invest in Theranos. That's right. Yeah. And most

10:06

of us don't have the opportunity to be scammed

10:08

in those big ways. We

10:11

have plenty of opportunities to be scammed in small ways, but

10:13

this idea that we accept as true, what we see

10:16

and what we hear makes

10:18

a lot of sense because we

10:20

can't go around our daily lives

10:22

constantly double-checking everything. We can't go

10:25

into the grocery store and disbelieve everything on

10:27

the packaging of a product. I mean,

10:30

we could, but we'd never get out of the grocery store. If

10:33

you go and buy some fruit and it says that it's

10:35

organic, you're probably not going to

10:37

go out to the farm and monitor the farmer

10:39

24 hours to

10:41

make sure that there's no pesticides being used. We

10:44

have to just accept that some of these things are true

10:46

and they might not be. And most of

10:48

the time it doesn't matter. So it's a

10:50

really good default assumption to have because it allows us

10:53

to have a conversation and it allows us to interact

10:55

with other people and be a part of a

10:57

community. It does mean that somebody

10:59

who's looking to deceive us, whether

11:01

it's a con artist or a magician,

11:04

will say things that simply aren't

11:06

true in order to get you to believe what

11:08

they want you to believe, to focus your attention where they want

11:10

to focus it. So that's a precondition

11:12

for any act of deception, is this willingness

11:15

to believe. Okay. So I'm hearing one principle.

11:17

There has to be more to it though.

11:20

Yeah. I think that's the underlying principle that

11:22

affects everything. You can't have deception without

11:25

a willingness to take what

11:27

people are telling you as true. And

11:29

when people are deceiving you, they're presenting something as

11:32

true that's not. That's kind of

11:34

the fundamental aspect of it. But

11:36

the cognitive tendencies are a bit different. There are things

11:38

that we do all the time that

11:41

most of the time work great. So for

11:43

example, we have a cognitive tendency to focus,

11:47

which is to pay attention to the information we have right in

11:49

front of us and often to

11:51

not think about the information we don't have in front

11:54

of us. And most of the time

11:56

that works great because what we actually want to do is

11:58

focus on what we're dealing with. and not

12:00

be distracted by irrelevance. But

12:03

that means that if somebody wants you to believe

12:05

something in particular, all they have to

12:07

do is put that information in front of you and keep

12:09

you busy with it and have you

12:12

not think about what you're missing. Can

12:14

you give us an example? Here's one

12:16

example. Imagine you're at a psychic performance.

12:18

You've gone to see the psychic who

12:20

helps people commune with the dead,

12:22

right? Well, what do they do in that

12:24

performance? They rattle off a whole

12:27

bunch of things at once until somebody in their

12:29

audience says, oh yeah, that was me. That fits

12:31

me. And then they zero in on

12:33

that person and talk about nothing but that person and

12:35

how they were right about them. What

12:37

does that do? It draws the audience's attention

12:39

just to that one case and not all

12:41

of the guesses that they'd missed up to

12:43

that point. So if they say, is there

12:45

a Rob, a John, a Frank, a George?

12:48

Oh yeah, there's a George, right? That's

12:50

all they focus on at that point. And

12:52

it's a great tactic. And

12:54

magicians essentially do this all the time,

12:57

right? They're directing your attention. So a lot

12:59

of magic works by misdirection. Well, that's really

13:01

just focus. They're focusing your attention

13:03

on one thing and you're not being

13:05

asked about anything else, right? And you don't think about those

13:07

other things. It comes out in

13:09

other contexts where people aren't trying to deceive, but they might

13:12

inadvertently do that. So a lot

13:15

of business books are about successful leaders,

13:18

right? Well, what are those books

13:20

gonna talk about? They're gonna talk about all of

13:23

the things that preceded that person's success, right?

13:26

And all of the things that they do as

13:28

a successful person. And we

13:30

focus in on those success stories and

13:33

only think about those. We don't think

13:35

about all of the people who

13:37

might've done exactly the same thing and didn't become

13:40

successful leaders, right? We don't think

13:42

about all of the other things that that same person could have

13:44

done that would have led to success or

13:47

all the other things they could have done that would have led to failure. We

13:49

don't think about the information we don't have. We

13:52

only think about that set of successful things and

13:55

we don't stop to ask, hey,

13:57

do those things have anything to do with why they were successful?

13:59

Or were they just... lucky. They could be lucky

14:01

or they could have done some homework and

14:03

picked something or represented themselves in a way

14:05

that makes someone more likely to be deceived.

14:08

One of the hooks that we talk about

14:10

is the hook of familiarity and having

14:13

a familiar name, having something we've encountered before

14:15

makes us somewhat more trusting of it, right?

14:18

And most of the time that's

14:20

a great thing because most of the time the

14:22

people we're really familiar with are people we have

14:25

known for a while, we've grown to trust and rely

14:27

on them because they haven't mistreated

14:29

us in the past. But

14:31

these days what counts as familiar, what

14:33

counts as a close friend, might

14:36

be somebody you've never met before, right? It

14:38

might be somebody you just know on social media and

14:40

have interacted with virtually and you've never actually met them

14:42

or interacted with them in person. And

14:44

that can lead to a lot of problems. I'm

14:47

sure you've encountered a lot of romance scams. The

14:50

victims of those romance scams have never actually

14:52

met the person that they have been in

14:54

a long relationship with. It's only been virtual,

14:56

right? And the person might not look

14:58

or be anything like what

15:00

they think they are because they're they

15:02

gradually build up this sense of familiarity which

15:04

leads to a sense of trust. Lauren.

15:13

Mike. So we host a podcast for

15:15

Wired called Gadget Lab. We do. We

15:18

do. Yes, that is correct. Tell

15:21

the good people some more about it. Well

15:23

I think the good people should definitely

15:25

tune in every week because they get

15:27

to hear me roasting you. I know.

15:29

Alright. No, really what Gadget Lab is

15:31

is Mike and I tackling the biggest

15:33

questions in the world of technology. I

15:36

like to think of it as the

15:38

best of Wired's journalism but in audio

15:40

form. We cover the big news of

15:42

the week in Techland but we also

15:44

offer our expert analyses and opinions on

15:46

all things consumer tech. Whether that's mobile

15:49

apps, hardware, startups, cryptocurrency. Mike, what's been

15:51

a recent highlight episode for you? We

15:53

did a deep dive on the group behind the

15:55

massive Okta hack. We also had a great conversation

15:57

about web 3 and the Metaverse. What

16:00

stands out for you? Never met a verse you didn't like.

16:03

I really enjoyed our recent podcast about

16:06

Peloton. And recently, the legendary

16:08

tech journalist Kara Swisher joined us to talk

16:10

all about Elon Musk and the future of

16:12

Twitter. So I guess we should tell people

16:14

how they can listen to our pod. We release a

16:16

new episode of Gadget Lab every week. And

16:18

you can listen and follow us on Apple

16:21

Podcasts, Spotify, or wherever you pod. So

16:27

what is HelloFresh? Farm fresh, pre-portioned ingredients and

16:29

seasonal recipes that get delivered right to your

16:31

doorstep. So they send you a box of

16:33

stuff. How do you know it's the right

16:35

stuff? It comes with everything you need in

16:37

the box. It's the season for giving and

16:39

gathering. And with HelloFresh, it can also be

16:41

the season of saving with fresh recipes that

16:44

are delivered cheaper than takeout. I love to

16:46

cook, but at the same time, after a

16:48

full day of work, there's still always something

16:50

else to do. With HelloFresh, it helps turn

16:52

busy weeknights into just great, memorable meal times,

16:54

practical options. Whether you're the home chef or

16:56

you know a chef, HelloFresh would make

16:59

a fantastic holiday gift. And it's

17:01

America's number one meal kit. If

17:03

people want to go ahead and

17:05

get this HelloFresh, what do they

17:07

do? They go to hellofresh.com/hackfree and

17:10

use code hackfree for free

17:13

breakfast for life. One

17:15

breakfast item per box while

17:17

subscription is active. That's free

17:19

breakfast for life. hellofresh.com/hackfree

17:24

with code hackfree. So

17:31

in your book, you talk about what's

17:33

called the possibility grid. And

17:35

I thought that was fascinating. Can you tell our listeners

17:37

a little bit about that? Yeah,

17:40

so this is a tool that's designed

17:42

to just help us realize when we're

17:44

focusing and not thinking about missing information,

17:46

right? So the possibility grid's a

17:48

simple idea. It just takes a little effort

17:50

to implement it for yourself. So

17:54

imagine a two by two grid, right? So

17:56

we can use psychic predictions since we've already

17:58

talked a little bit about psychic. The

18:00

top row would be predictions that the psychic

18:02

made, and then the top

18:05

left column would be predictions that were

18:07

true. So the

18:09

psychic predicted that there would

18:12

be a hurricane that would hit Miami in this

18:16

year. Right.

18:20

And it did. So that's

18:22

the top left cell of that

18:24

grid, the top left box. That's the one

18:27

we tend to focus all of our attention on. That's

18:29

the one with the success

18:31

stories of a successful leader, the activities

18:33

of a successful leader. But you

18:36

can also look at what are the

18:38

predictions that psychic made that didn't come

18:40

true, that a hurricane would

18:42

hit Galveston. We

18:45

need to know how many of those also

18:48

failed, those failed stories, those cases

18:50

where they predicted something that didn't come true. We

18:52

need to know about those in order to know whether they're any good at

18:54

making predictions. We also need to

18:56

know about the things that they didn't predict that

18:59

did come true. And there's

19:01

a great project out of Australia, Richard Saunders

19:04

helped lead this, where they cataloged

19:06

a whole bunch of world events. Everything

19:08

from the Indonesian tsunami to 9-11 to

19:11

a whole bunch of international world events that

19:13

you'd think if psychics were able to predict

19:15

the future, they should predict. They then looked

19:17

at all of the prominent psychics in

19:20

the Australia area, in the media, in

19:22

television, YouTube, and looked to see

19:24

if any of them actually predicted those things. And it turns out,

19:27

no, they hadn't. So those are

19:29

the ones that would go in that

19:31

bottom row of predictions that weren't made,

19:33

but things that actually happened. And

19:36

then there's all the predictions that weren't made and that

19:38

didn't happen, which we just don't know about.

19:42

So if you want to understand whether

19:44

or not the information you have is complete, you

19:46

have to look at all four cells. And

19:48

the idea of the possibility grid is not to

19:51

necessarily make you kind of calculate all the numbers

19:53

and figure out all those details. We can't do

19:55

that for anything most of the time. But

19:57

what you can do is say, okay, all I've got are things

19:59

that are saying, success stories? What

20:01

about the people who did

20:03

the same thing and failed? What about

20:05

people who did something different and succeeded?

20:07

And that will tell you that those,

20:09

say, executive leadership stories might

20:12

just be stories. It might turn out that all

20:14

of those things that were in that top left

20:16

part of the grid were

20:19

there but arbitrary, and that if you took the same

20:21

people and had them do those things, had them do

20:23

those things again, they might fail. And

20:25

maybe those things actually hurt your performance, make

20:27

you less likely to succeed. It's hard to

20:30

know. But in order to know there's

20:32

any pattern at all, you have to have all of that

20:34

information. No, we do live in a world where you have

20:36

people that have made a fortune teaching

20:38

other people how folks succeeded.

20:41

But they never talk about, well, these

20:43

are the folks who didn't succeed, and maybe

20:46

we should also look at what they might

20:48

have done wrong as to why they didn't

20:50

succeed. Yeah, absolutely. There's

20:52

a real danger in case studies of

20:54

success, in that you can fool

20:56

yourself into thinking that the things that happened in

20:59

that one case are representative

21:01

of all people or that

21:04

they're even associated. It may turn out that those

21:06

things are antipredictive of success. It may

21:08

turn out that they actually are worse for you to

21:10

do. And it might be that that person just got lucky.

21:13

Are these case studies universal or are these

21:15

tendencies limited to the US or Western culture?

21:18

I think we're trying to focus on

21:20

cognitive tendencies that are fairly universal, that

21:22

are just a process that we need.

21:26

None of these things like focus are inherently a bad

21:28

thing. So there

21:31

are things that we need in order to be able to

21:33

function effectively. Some of the other hooks that we talk about

21:35

are things like prediction, kind of

21:37

anticipating future events. That's something

21:39

we have to do all the time. We

21:41

can't function effectively without being able to anticipate

21:43

what might happen next. And we

21:47

have to form expectations and that governs

21:50

how we act. So if you expect

21:52

that it's going to rain tomorrow, you

21:54

will take an umbrella with you. And that's

21:57

going to be true anywhere. You'll adapt to

21:59

the environment appropriately. based on your expectations.

22:02

The same principles will apply in our

22:05

interactions in social media, right? So if

22:08

you expect or

22:11

believe some pattern to be true, and

22:13

you predict an outcome, you're going to

22:16

be more receptive to things that are consistent with

22:18

it. You're not going to question them as much as

22:21

you would if something were completely opposite

22:23

of predictions. So we tend to challenge

22:25

things that counter what

22:27

we believe pretty effectively. We're all pretty good at

22:30

this, right? We're all pretty good at shooting down

22:32

that claim that we just don't agree with. And

22:35

the problem is we're not as good at shooting down claims

22:37

that we agree with. We don't tend to do that as

22:39

much. I think that's probably

22:41

something that's cross-culturally universal, right?

22:44

We all are going to have to make predictions. We

22:46

have to do this. Same thing with commitments, which are

22:48

just basically strongly held assumptions that

22:50

we really can't discount, right? Once

22:53

we've made this strong assumption, we're not as

22:55

likely to challenge things that are consistent with

22:57

it. That's going to be

22:59

true for anybody who makes an assumption anywhere. It

23:02

doesn't have to do with individual differences

23:04

in that way. Some people are going to be more

23:06

committed to their beliefs than others. Of

23:08

course, yeah. And there might be some cultural variation in

23:10

that. Yeah,

23:12

absolutely. But is it something

23:14

that's going to be true, at least in some cases,

23:16

for everyone? Yeah, it kind of has to be.

23:19

Because we have to be able to do these sorts of things.

23:21

We have to be able to focus. We have to be able to

23:23

make predictions. We have to be able to form

23:25

commitments and stick to them. And efficiency

23:29

is the last of the hooks. We have to kind

23:31

of try and act efficiently because we have limited time,

23:33

limited capacity. So we have to act

23:35

in ways that allow us to get things done

23:37

effectively. And scammers really

23:40

prey on people, in particular,

23:42

where they have an opportunity to make

23:44

you make a snap decision. Yeah,

23:47

absolutely. Because they create this urgency.

23:50

And that almost kind of fits in with the

23:52

whole concept of the efficiency of the thought process.

23:54

You want to be efficient. And

23:56

something happens. You want to move immediately because

23:58

you think the world... was about to end if

24:01

you don't. Yeah, exactly. And that

24:03

works in both a positive way and a negative

24:05

way, right? So offers

24:07

that are expiring in the next hour, or there are

24:09

only two left, right, are attempting

24:11

to get you to act with efficiency to

24:13

kind of pick up on that deal, which

24:16

is almost never necessary, right? It's almost never

24:18

that urgent. But in the same

24:20

way, it can work as a negative, right? So there

24:22

are scams right now that are fairly

24:25

pervasive that involve calling up parents

24:27

or grandparents or cousins and saying,

24:29

Hey, this kid has been arrested,

24:31

we need you to send over money to get them out

24:33

of jail, right? There are lots of variants of these, some

24:36

of them more horrific than others. But

24:38

all of them are preying on that

24:40

immediacy, we need you to respond right

24:42

away. It's an urgent situation. Even

24:45

the call center scams where you people will call up

24:47

and say, you made a mistake on

24:49

your visa application, the police are coming to arrest you

24:51

and deport you right now, right? That

24:55

same sort of scam is putting people under

24:58

immediate pressure to say, if you pay it off, you

25:00

can make this go away. And we

25:02

really like that sort of efficiency, making it go

25:04

away quickly when you're under pressure to deal with

25:06

something that takes you takes your

25:09

opportunity to say, wait a second, this doesn't make

25:11

any sense, and takes away your

25:13

ability to check on it really quickly. As

25:15

far as that's concerned, what advice would you have

25:17

just for a layperson in terms of how to

25:20

avoid that tendency to just jump to that action?

25:23

Yeah, it's really hard because that's that is our

25:25

natural tendency to try and resolve a problem. And

25:29

one thing that I think helps is seeing a whole

25:31

bunch of examples of how that sort of time pressure

25:33

works in settings ranging

25:35

from terrible to completely innocuous. So

25:39

magicians use that all the time, right? They put you under

25:42

pressure to make a decision right then. And you don't think

25:44

about it when you're doing it. And that forces you to not

25:46

think about the other things they're trying to hide from you. Anything

25:49

where it's kind of a rapid decision, you have

25:51

to decide on this right now, almost

25:54

never the case that you need to decide on something right now.

25:57

For things as specific as the sort of call center

25:59

scams or the or the kidnapping or

26:01

your kids in jail for bail

26:03

scams. There are a couple of things you

26:05

can do, and the best way I think

26:07

to handle those is to make sure that you put

26:09

yourself in a situation where you

26:11

can counteract that time

26:14

pressure quickly. And one

26:16

of the things that we do is kind

26:18

of like what growing up I had with

26:21

my family, which was when there was a big

26:23

scare about stranger kidnappings. We all had a family

26:25

passphrase, so that if somebody came up to you

26:27

in a car and said, hey, your mom asked

26:29

me to take you home, you

26:32

could give them the passphrase, and if they didn't ask

26:34

my passphrase, and if they didn't know it, you run,

26:36

right? Right. Same principle applies, right?

26:38

Scams really haven't changed all that much. If you

26:41

have a family passphrase and somebody calls you up

26:43

and says, hey, your cousin's been arrested, or

26:46

your kid's been arrested, what's the

26:49

passphrase? And they'll try and dodge that

26:51

and get away with it, but it's a way of getting

26:53

out of that scam and getting rid of the time

26:55

pressure preemptively, right? Being aware of

26:58

what you could have happened to you. But

27:00

that's a really specific solution to a specific

27:02

kind of problem. I think the

27:04

bigger thing is just to recognize that time

27:06

pressure almost always is a bad thing. Right,

27:09

when somebody's putting you in time pressure, it's

27:11

almost always a sign that you don't

27:13

want to rush. So being able

27:15

to stop and say, okay, wait a second, I need to think about this

27:17

a little longer, is a good mode, a

27:19

good move to make. Asking a couple

27:21

of additional questions can

27:24

really make a difference there. So how

27:26

do you strike a balance between trust

27:28

and skepticism? Because if

27:31

you trust too much, you're in trouble.

27:33

If you're too skeptical, you have no

27:35

life. Yeah, exactly.

27:37

And I think

27:39

that's one danger that I think you

27:42

commonly see among people who are kind of contrarian

27:44

is that, well, I'm a skeptic, I'm a contrarian,

27:46

I'll never fall for any of this. And it's

27:48

like, well, no, that's not true either. If somebody

27:50

targets your skepticism, you could absolutely fall for it.

27:53

There are probably good, fake skeptic organizations

27:55

out there catering to people who believe

27:57

they're skeptics. Right, but... We

28:00

can't be skeptical of everything all the time or we'd never

28:02

get anywhere. We have to be able to trust other people.

28:05

And I think, again, it's important

28:07

to keep in mind how rare big

28:10

cons are, how common probably

28:12

small forms of deception. And

28:15

those small forms of deception most of the time don't

28:17

matter enough to most of us to worry about

28:19

them. So for example, if

28:22

a grocery store happens to scan prices

28:24

slightly higher than they were marked on

28:27

the shelves. For

28:29

some people that really matters and they have to

28:31

check every penny that they're spending. But if you

28:33

can afford to lose a couple of dollars here

28:35

and there when you go grocery shopping, it's probably

28:37

not worth your effort to cross check every receipt

28:39

against every price in the store because that form

28:41

of deception just isn't a big deal. But

28:44

if you are investing your retirement, it

28:47

probably does make sense to double check,

28:50

especially if you're planning on investing your retirement with a

28:52

money manager who reaches out to you, it probably

28:55

makes sense to check them out really thoroughly. I

28:58

think there's some ways you can kind of think about

29:00

this thinking about the big risks are going to be

29:03

for things that are relatively unregulated.

29:07

So if you're making decisions about things that

29:09

are relatively unregulated and involve the potential

29:11

for identity theft or large amounts of money. And

29:15

those are cases where you can be extra cautious.

29:18

If you're investing your retirement savings with

29:20

a giant bank or Vanguard or something like

29:22

that, those are fairly

29:25

well regulated industries. They've been around a

29:27

long time. The odds that

29:29

Vanguard is pulling a massive scam on

29:31

the whole world is probably pretty low.

29:34

You probably don't need to spend a lot of time checking

29:36

out the company itself in this

29:39

context. But if it's just a money manager who a friend

29:41

of a friend recommended, you probably do want to

29:44

check them out. And one

29:46

thing that we found reading through scams in

29:48

all of these fields is it's remarkable how

29:50

often people who are pulling off

29:52

scams and cons have done it before and been

29:54

caught for it. They've

29:57

actually often been convicted or

29:59

charged. with scams and

30:01

fraud, and then they pull it off again.

30:04

There's a great example of this, a relatively

30:06

recent example of this, of an Orlando

30:08

art museum that had an

30:10

exhibit of Basquiat paintings that

30:13

had been provided to them. And the FBI relatively

30:15

recently raided the museum and took all of the

30:17

paintings because they suspected almost all of them, if

30:19

not all of them, were forgeries. The

30:23

Orlando Museum of Art fired its

30:25

director after the FBI seized 25

30:27

possibly fake paintings on display. The

30:29

paintings were said to be works

30:32

of the late artist Jean-Michel Basquiat.

30:34

But there are authenticity. It's now in

30:37

question. The New York Times reports a

30:39

man who initially said he bought the

30:41

paintings from Basquiat, told federal agents he

30:43

never met the painter. It's not the

30:46

OMA's job to authenticate art. They

30:48

came to us authenticated by the top

30:50

specialist on Basquiat. FBI

30:54

officials now plan to forensically examine

30:56

the paintings. And the

30:59

people who provided them to the museum claimed

31:02

to have found them in a storage locker. And

31:05

so they were newly disfigured and

31:09

worked with the museum to have an exhibition, and they were planning

31:11

on selling it for millions of dollars, selling these

31:13

paintings for millions of dollars. Possibly

31:16

if the museum director had checked out the backgrounds

31:18

of the people who were selling them the Basquiat's,

31:20

they might have been a little less likely to

31:22

work with them because among the several of them,

31:24

they had about seven convictions and a couple of

31:26

them for fraud. So this is

31:29

not an unusual pattern. But that brings me to

31:33

another wishy-washy part of this world, which

31:35

is in the art world,

31:37

for example. Yeah. It's a very thin

31:39

line between a fraud and business

31:42

as usual. And so,

31:44

you know, when you look up somebody,

31:46

let's say you look up a contractor.

31:48

Let's make it a little more run-of-the-mill.

31:51

You go online and you need some work done

31:54

around your house. Now, chances are that

31:56

at the moment you need that work done, you

31:59

really need it done. So you're in a

32:01

bit of a rush, you got some stress about

32:03

it now. No one's available, no

32:05

one's available, no one's available, that person's

32:07

available. Oh my gosh, that person's available.

32:09

You look them up and you see

32:11

there's a couple of court dates in

32:13

there that have been made the

32:16

public record and you think, hmm, you

32:18

know, and you ask them about them and

32:21

they say, oh, you know, that was just,

32:23

you know, people can be really unreasonable. That

32:25

was one of those, that was an unreasonable

32:27

person. You still

32:30

might go for

32:32

it. Yeah, you might if you're out

32:34

of options. That's also a good

32:36

example because what are the, what

32:38

is the information you have in front of you? Right.

32:42

You have the information of what they've done in the past and they

32:44

might tell you about all of their successful contracting

32:46

projects, right? And

32:48

you have the information about a couple of

32:50

cases where people filed suit against them for

32:53

not coming through. And

32:56

what you should think about then is what

32:58

information do I not have? So

33:01

one question you can ask is how

33:04

likely is it that somebody would

33:06

actually follow through and sue

33:08

somebody as opposed to just giving up on them and

33:10

moving on in a court

33:13

case? So the odds are good that

33:15

if there are a couple of court cases, there are a lot

33:17

more cases that didn't make it to court. So

33:22

that's something to keep in mind. There might be a lot more cases where they

33:24

did a great job too, but the ones you

33:26

care about are the ones where people were

33:28

upset enough to follow through. And

33:30

the thing is, it's just like

33:33

three-card Monte or any other trick

33:36

where I might say, Dan, Dan, Dan,

33:38

you know, yes, that woman sued me,

33:40

but I will give you the number

33:42

of six people who didn't sue me.

33:46

Yeah. And

33:48

you call them up and they're like, oh,

33:50

Bo, he's great. You know, and they're all

33:52

people who are on the take. Yeah,

33:55

exactly. And I was just

33:57

out in LA a week or so ago and

33:59

watching. somebody doing three-card monty on

34:03

Venice Beach. And if you

34:05

watch long enough, I was just kind of, I had nothing else

34:07

to do for an hour or two, so I was just wandering

34:09

back and forth, and the three-card monty setup would just move down

34:11

Venice Beach. And every single

34:14

time it was this same woman who didn't

34:16

look anything like the person who was doing

34:18

the moving the caps around, same

34:21

person would be the first one to pick

34:23

and would win like a hundred dollars. Every

34:25

single time, it's the same woman walking

34:28

down the block. I was in

34:30

Paris, same situation, same game, and

34:33

I was watching, and I figured out who

34:35

the players in the scam were. So I

34:37

went over and I said, I

34:40

totally know what's going on here,

34:42

and without missing a beat, he

34:45

showed me a knife. And

34:49

that was the end of the conversation. Yeah,

34:51

I often feel that way too, but you

34:53

know. No, but I mean, and that is

34:55

part of it. It's like if you do,

34:57

if the veil trembles and you look on

34:59

the other side of it, be

35:02

careful. Yeah. Sure. These

35:04

are criminals who choose to make

35:07

money through trickery, which means they're not the ones

35:09

who probably, probably not the ones who are throwing

35:11

you up against a brick wall, but

35:13

it's unreasonable to believe they don't know

35:15

someone who will do that. Yeah,

35:18

exactly. And the other thing to watch when you're watching

35:20

three-card Monty is you should have your hand on your

35:22

wallet at all times. Yeah. Because that's a classic

35:26

place where pickpockets work, and they work in

35:28

tandem with the people doing three-card Monty. Yeah.

35:30

So I have a friend, Paulo

35:32

Robbins, who's a professional thief, right? He's not

35:34

in the negative way. He's a gentleman's thief.

35:36

He's a Vegas performer. He'll take your watch

35:38

off your wrist. He'll take your glasses off

35:40

your face. He won't notice. Oh, really impressive.

35:42

But he carries a business card around. And

35:44

sometimes if he sees people pulling this three-card

35:46

Monty, he has a fake wallet, and he'll

35:48

position himself so that they'll try and take

35:51

it. And they'll take it and he'll have

35:53

him, he'll hand them his business card. He

35:55

says professional thief. That's

35:57

funny. Yeah. Then there's the the

36:00

folks who bury you in facts. Because

36:04

they figure that the more professional

36:06

they sound, the more successful they

36:09

sound, the more they can be

36:11

very specific about what they're doing, the more you'll

36:13

buy into it. I guess

36:15

an analogy of that might be Theranos,

36:17

right? I mean, this was some impressive

36:20

operation. And when you talk about high

36:22

profile people who can be scammed, uh,

36:25

that was the list of who's who of the

36:27

Western world who got scammed by that. I

36:30

think it's a fascinating case because it highlights,

36:32

first of all, that highly successful people who

36:34

might view themselves as very good, critical thinkers

36:36

were taken by that one, right? That,

36:39

that board of directors had retired generals

36:41

and admirals that had a, you know,

36:43

prominent secretary of state who were major

36:45

leaders in world world figures. Right.

36:47

And they were taken by it. They weren't

36:49

necessarily biotech experts and a lot of biotech

36:52

experts didn't want anything to do with Theranos.

36:54

It was a lot of California tech money

36:56

and not as much sort of biotech money.

36:59

For, I think for good reason, that people

37:01

wanted to kind of have a feel for

37:03

this. They had, it seemed really appealing. The

37:05

idea of it was a really

37:07

compelling narrative. And that, that

37:10

was one of those sort of big cases

37:12

that used all of these sorts of principles,

37:14

um, to good effect. The idea

37:16

that you were talking about a very rapid banter with a

37:18

lot of facts thrown at you, sometimes

37:20

that plays into what we call the hook

37:22

of precision, where people will give a concrete

37:25

specific number associated with something that doesn't actually

37:27

have any bearing in reality. So we can

37:29

do over a thousand tests with

37:31

this single drop of blood. Never could.

37:34

Right. But that concrete number instead of saying,

37:36

Oh yeah, we have a lot notes, like

37:38

we can do 1000 and whatever. That

37:41

precision is something that usually is great.

37:44

If it's not true, that's a problem.

37:46

So precision, if somebody can actually genuinely

37:48

give you a number to a great

37:50

degree of precision, and it's based on

37:52

a good understanding, that's really cool. Right.

37:54

So physicists can measure a constant down

37:56

to, you know, however many decimal places

37:59

and the. Precision they're fighting with there is maybe

38:01

the eighth decimal place or something. For

38:04

most human behavior, we don't have that level

38:06

of precision, but if you give people precision,

38:08

they tend to assume that there's something underlying

38:10

it. And this can be for something as

38:12

dumb as the price of a house

38:15

listing. If you give it

38:17

to a concrete number like it's 378,500 as opposed to 380,000, the

38:19

more precise number

38:26

will lead people to negotiate less and end up

38:29

paying a higher price because they

38:32

kind of intuitively assume that if

38:34

there's that much precision, there must have been a

38:36

reason for it. There must have been something behind

38:38

it. Maybe the comparables were more precise and more

38:41

detailed. And we don't tend to think

38:43

about the way that that influences us, that we just jump

38:45

right in and say, oh, that's precise. And it just keeps

38:47

us from thinking about it more critically. I'm

38:54

CBS News correspondent Major Garrett, host

38:56

of the podcast, Agent of Betrayal,

38:58

the double life of Robert Hanson.

39:00

During the Cold War FBI agent

39:02

Robert Hanson traded classified secrets to

39:04

the Kremlin in exchange for cash

39:07

and jewels. In the podcast, you'll

39:09

hear from Hanson's closest friends, family

39:11

members, victims and colleagues for the

39:13

most comprehensive telling of who Robert

39:15

Hanson really was. Binge the entire

39:17

series now, Agent of Betrayal, the

39:19

double life of Robert Hanson is

39:21

available wherever you get your podcasts.

39:24

Hey, Brad, you know nationwide is

39:26

more than an insurance company. Yeah,

39:28

they're one of America's largest financial

39:30

services companies. We get that in

39:32

a song like business life retirement

39:34

or nationwide there to protect. I'm

39:36

kind of the jingle guy. Not

39:38

sure I agree with that. I'm

39:40

not sure I like your hat.

39:43

Well, it would never fit on you. Products

39:45

issued by Nationwide Life Insurance Company or

39:47

Nationwide Life and Annuity Insurance Company, the

39:49

general distributor for variable products is Nationwide

39:51

Investment Services Corporation member FINRA Columbus, Ohio.

39:57

So just to zoom out here a little bit for our

39:59

listeners. if they are considering playing 300 mod

40:02

D, which they shouldn't. Don't. Or if

40:05

they're looking for a contractor looking to buy

40:07

a house or invest in some big new

40:09

stock or something like

40:11

that. Crypto. Yeah, don't. Yeah,

40:13

don't. Outside of just

40:15

don't, is there one piece of advice that you would

40:18

give to avoid falling for a scam? You

40:21

know, I'm not sure that there's any single piece

40:23

of advice that works well for everything, but

40:26

there are a lot of questions you can ask yourself

40:28

that if you take a step and take a step,

40:31

pause and ask yourself a question. One

40:34

way to think about whether you might be in a

40:36

situation where you're at risk of being scammed is, if

40:39

I were a scammer, what would I do? Right?

40:42

If I were a scammer trying to fool me,

40:45

what would I do? How would I do it? And

40:48

often if you ask that question, you'll realize I

40:51

couldn't tell. Right? The things

40:53

you'd come up with doing would be very similar to what

40:55

was happening right then, and you wouldn't be

40:57

able to tell if it was a scam or not because

40:59

you're not thinking in the way that they would. So as soon

41:01

as you start thinking the way a scammer would, then

41:04

you realize what they might be up to. Right?

41:07

So if somebody were hiring a contractor and say,

41:09

okay, if I were a scamming contractor, if I

41:11

were somebody who was trying to scam somebody, take

41:13

their money and then not complete the work, what

41:16

would I say and do? Well, I'd tell

41:18

people all of my success stories. I would tell

41:20

people to ignore the cases that were, you know,

41:24

negative cases and they were just disgruntled people. I'd

41:28

say all of those sorts of things. And

41:31

if they're saying all those sorts of things, it's like, okay, maybe these

41:33

are red flags. Maybe they're not,

41:35

but maybe they are. So that's

41:37

one of my favorites. They're more subtle

41:39

ones you can use that are simple in

41:41

other contexts. One of my favorites is, is

41:44

that really true? Just

41:46

asking yourself, is that really true? And

41:49

I think this is one of those

41:51

things that can help us from accidentally

41:53

sharing information that is not true. It

41:55

can help us avoid spreading disinformation. Most

41:58

of us don't want to spread disinformation. We don't

42:00

want to mislead people, right? But

42:02

we have a tendency to sort of share

42:05

or forward things that align

42:07

with our beliefs. So you get information

42:10

that kind of aligns with your beliefs. It sounds kind of cool. It's

42:12

like, oh yeah, I want to share that with all my friends. You

42:14

repost it. Just pausing for a second,

42:16

saying, is that really true? And if

42:18

it weren't, how would I know? Can

42:21

cause you to take a step before

42:23

you share it with other people and

42:25

potentially spread misinformation. Well, that's great advice.

42:28

My daughter, who is at

42:30

University of Michigan and studying neuroscience, so

42:32

she's sort of in the bucket you

42:34

would appreciate, her favorite thing to

42:36

say in the face of things

42:38

that are patently false or perhaps a little

42:40

squishy, she will say, but

42:43

that's not true. And then if

42:45

you continue, she'll say, yes, but that never

42:47

happened. And

42:52

it works for the high functioning

42:54

Rain Man guy in Utah making

42:56

the 7.6.1, whatever, return

43:01

on your dough to a lot of things. I

43:03

think that it comes down to, for

43:06

me, hearing you talk today,

43:08

what it comes down to is listening for

43:11

the many different ways in which people

43:14

say, how does that sound to you?

43:18

And you have to answer the question

43:20

thoughtfully and say, that sounds like a scam to

43:23

me. I

43:25

know there's no contractor available for three

43:27

months, but I can do it next

43:29

week. The only thing is, I need

43:32

to know that you're actually down, so you're going to give

43:34

me a $500 deposit. How

43:38

does that sound? It sounds terrible. Get out

43:40

of here. Yeah. So in

43:42

other words, what you're saying is one of the

43:44

greatest contributions that any of us can make to

43:47

society is not to

43:49

pass something along unless we

43:51

feel comfortable that it's the

43:53

real deal. And

43:56

the real question is, can I tell that it's the real deal?

43:58

Right? I think

44:00

people are too quick to kind of share things because

44:02

it's fun to share things and you get approval for

44:05

sharing things and people like it, right, when you share

44:07

something cool. And

44:09

if it's just something cool, if it's just, you know, a cat video

44:11

or something, it's like, okay,

44:13

fine. But if it's something

44:15

political or something about science or

44:18

some radical claim, taking a second

44:20

and saying, okay, is that really true? And

44:23

how would I know? What would I

44:25

need to know in order to know that that's true? Can

44:27

stop a lot of stuff, right? So you

44:30

get a crypto offer like, well, is that true?

44:33

No, I don't know. How would I know? What would I

44:35

need to know in order to verify this? And

44:37

I think we often don't think about

44:39

how much information we really should have

44:41

before forming an opinion and before actually

44:43

having an idea about whether something is

44:46

worth doing or not. Crypto

44:48

is one of those classic cases that it taps

44:50

into all of these sorts of hooks. It

44:53

uses celebrity endorsements. It promises outlandish returns

44:55

on short investments. It promises these underlying

44:57

formulas or models that are supposed to

45:00

be amazing and defeat all of

45:02

these sort of standard approaches. It

45:04

tells you you're going to be able to get these

45:06

returns really quickly, right? It has all of the

45:08

hallmarks of classic Ponzi schemes, you know,

45:10

guaranteeing big returns in short windows. But

45:13

how many of us actually understand how

45:15

blockchain works, the math of it behind it?

45:18

How many of us could actually explain one of

45:20

those models to anybody? Now I'm going to step

45:22

in like the big bad scammer and

45:24

say, you know, you can listen to

45:26

Dan sounds reasonable

45:29

or you can

45:31

make $100 in the next 10 minutes. So

45:35

it's up to you. You listen to Dan sounds

45:37

reasonable or you make a hundred bucks in the

45:39

next 10 minutes. Totally up to you.

45:42

Now listen, the 10 bucks that you need to give

45:44

me in order to make that $100 in the next

45:47

10 minutes, isn't a very high

45:49

price. And if I'm a scammer, you only lost

45:51

10 bucks. I'm going to get that

45:53

10 bucks, Dan. A

45:55

lot of people will pay that 10 bucks. A lot of people also

45:57

will play the lottery even though they're never going to win it. Any

46:02

parting words for us, Dan? I

46:04

think one of the things that we should do

46:06

is be more kind in thinking about people who

46:08

have fallen for these sorts of scams. It's very

46:10

easy to look at somebody who made a mistake,

46:13

fell for a scam, and as

46:15

a result of that, you know, within

46:17

dire straits, it's easy to look at them and

46:20

say, oh, yeah, they were just clueless. They were

46:22

just gullible. I would never have fallen for that.

46:25

And I think that's an uncharitable way to look at

46:27

things. I mean, yes, sure, some people are going to

46:29

be more gullible than others, but everybody

46:31

can be fooled if they're not careful and

46:34

if they're not, if they're targeted in the

46:36

right ways. We all can be

46:38

taken in, even if we view ourselves as

46:40

very skeptical. So I think

46:43

treating other people the same way that we would want to

46:46

be treated if we happened to fall for something like that

46:48

is something worth considering. I think the

46:50

optimist in me would argue that

46:53

over time, we're going to be aware that it's

46:56

so much easier to create fake information

46:58

and every student going through high school

47:00

and college right now is, I mean,

47:02

so many of them are using things

47:04

like chat GPT already. That

47:07

it becomes apparent that, you know,

47:09

you can't tell whether something was

47:11

AI generated or human generated, which

47:14

means that maybe people will become a little

47:16

more skeptical and look for validation of

47:18

what they're reading before they believe it. And

47:21

the optimistic view is that maybe people

47:23

will be more willing to look when they realize how easy it

47:26

is to fake things. And we're

47:28

only now kind of entering that territory where

47:30

fake images and, you know, fake

47:32

pros are becoming more pervasive,

47:35

right? It was always possible to do

47:37

fake images. You know, politicians have done

47:39

that forever, but it's

47:41

now much easier to fake things

47:43

that look pretty close to genuine.

47:46

So my hope is that maybe people will realize,

47:48

okay, I can't trust things just because

47:50

somebody's sharing them and they match my beliefs.

47:53

And maybe we'll go back to

47:55

having some sort of verification before people

47:57

will, you know, pass it along, because they'll recognize it.

48:00

Hey, this could very easily be fake. And I had an easy

48:02

way of knowing. That's the optimist view. Dan

48:06

Simons, we can't thank you enough for

48:08

talking with us today and your book,

48:10

Nobody's Fool, that you co-read with Christopher

48:12

Shabrie. It's available anywhere books are sold.

48:15

So everyone go out there and buy

48:17

it. Just make sure you go

48:19

to the right places where books are sold.

48:21

So you really buy the book. Not

48:26

amazon.com. That's right. All

48:29

right. Thank you, Dan. It was very nice to meet

48:31

you. Thanks, Dan. Great to meet you. Thanks for having

48:33

me on. Now

48:47

it's time for the 10 force one. Our

48:50

paranoid takeaway to help keep you safe

48:52

online. Travis, you

48:54

said you got something kind of upsetting by

48:56

email or I don't know, is it snail

48:58

mail? It was snail mail. Yeah. Um,

49:01

I got a letter in the mail saying that

49:03

my daughter's information had been compromised for data breach.

49:05

What kind of information? Well,

49:08

they always use the term may have been

49:10

compromised, but it sounds like it's the full

49:12

rundown, including name, address, social security number. It's

49:14

not great, but do you know who was breached?

49:18

I was actually her pediatrician. It does bring

49:20

up the potential for a kind of identity

49:22

theft that people might not be fully aware

49:24

of, which is medical identity theft. And when

49:27

you're talking about the kind of identity theft

49:29

that can flow from that, you're talking about

49:31

everything from account takeover, opening new accounts, getting

49:34

medical treatment in your name, committing

49:37

criminal acts, your children

49:39

becoming victims of identity theft. I mean,

49:41

it is the full montage. Yeah.

49:43

And the other thing is because your

49:46

family member was, uh, compromised, you might

49:48

think, well, you're safe. You might not

49:50

be because that information can be a

49:52

foothold to your information. And it might

49:54

be that your child's granular

49:56

information might be the very way you

49:59

get into someone. gets into your mortgage, not

50:01

theirs, because they don't have one, but you do,

50:03

and it's worth a lot of money. We could

50:05

all be compromised with everybody. Yeah, I mean, it's

50:07

one of the most, if not the most expensive

50:09

form of identity theft. The statistic I looked up

50:11

said that the average cost to the victim is

50:13

over $22,000. Well,

50:16

it's not just financial though, because if somebody goes

50:18

in and gets something fixed or checked or worked

50:20

on or whatever you want to say, you

50:23

could end up not being able to get

50:25

that very same treatment because your insurer is

50:27

going to be like, you just got, you

50:30

know, an appendectomy. You can't

50:32

have two. Yeah,

50:34

and it also depends on what or

50:36

how someone's using your plan. But for instance,

50:38

I was just reading about a woman who

50:41

had her purse stolen, including your driver's license

50:43

and her insurance card. And

50:45

she went through all the right steps. She canceled

50:47

her credit card. She got no identification. But two

50:49

years later, she was arrested because the thief in

50:51

question used her medical plan to get over 1,700

50:55

opioid prescriptions. Listen, even

50:57

if you have a different blood type showing up

50:59

in your medical record, that simple

51:02

stupid little difference, it could kill

51:04

you, unless you're a universal

51:07

donor, I guess. So a

51:09

lot can go wrong. And you know, doctors

51:12

may not know that it's

51:14

wrong. So you really have to take care of yourself

51:16

here. It goes back to

51:19

what we talked about with David Mayman. The

51:21

thief isn't just limited to your social

51:23

security number anymore. Your driver's license, medical

51:26

card, anything connected, you can be used

51:28

to commit identity crimes. Right.

51:31

Okay. Enough of this scare tactics here.

51:33

What can people do? Well, the

51:35

first thing is to protect your medical information,

51:37

like your insurance number as carefully as you

51:39

would your social security number. Wait,

51:42

fine. But that can be pretty

51:44

hard. Unlike your social security

51:46

card, you need to carry your driver's

51:48

license and insurance information in case something

51:50

happens. Right? Yeah, totally. But monitor

51:52

your health care too. And that of your kids or your other

51:55

dependents in the same way you would check your credit. So

51:57

my daughter's in elementary school and if I happen to look up here, I'm going

51:59

to her records and find out that

52:01

she has Vicodin prescriptions in her name. I'll

52:04

know something, so. You should also, every time

52:06

you get an explanation of benefits statements from

52:08

your health insurer, look at it and

52:11

make sure that the person who got

52:13

that treatment, who had that appointment, was

52:16

actually you. Right. And if you see

52:18

something that was, you know, in your

52:20

name, but wasn't, you didn't get it,

52:22

that's another kind of fraud. Absolutely.

52:24

And if you have your wallet or purse stolen,

52:26

ask your health insurer to provide a new

52:28

account number in addition to a new card. Listen,

52:31

most doctor's offices are going to make

52:33

sure that your information is up to

52:36

date, and they'll give you an opportunity

52:38

to verify that. Don't

52:40

blow them off. Just take a look, make sure it's right. And

52:43

that's our 10-4-L-1. What

52:53

the heck with Adam Levin is a production

52:55

of Loud Tree Media. You can find us

52:57

online at adamlevin.com and on Instagram, Twitter, and

52:59

Facebook at Adam K. Levin.

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features