Podchaser Logo
Home
Technology-Assisted Review: Sara Lord Interviews Data Scientist Lenora Gray

Technology-Assisted Review: Sara Lord Interviews Data Scientist Lenora Gray

Released Monday, 8th April 2024
Good episode? Give it some love!
Technology-Assisted Review: Sara Lord Interviews Data Scientist Lenora Gray

Technology-Assisted Review: Sara Lord Interviews Data Scientist Lenora Gray

Technology-Assisted Review: Sara Lord Interviews Data Scientist Lenora Gray

Technology-Assisted Review: Sara Lord Interviews Data Scientist Lenora Gray

Monday, 8th April 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:01

Welcome to the Emerging Litigation Podcast

0:04

. This is a group project driven

0:06

by HB Litigation , now part of

0:08

Critical Legal Content and

0:10

VLEX Company's Fast Case and Law

0:12

Street Media . I'm your host , tom

0:14

Hagee , longtime litigation news

0:16

editor and publisher and current litigation

0:19

enthusiast . If you wish to reach

0:21

me , please check the appropriate links in the

0:23

show notes . This podcast is also

0:25

a companion to the Journal of Emerging Issues

0:27

and Litigation , for which I serve as

0:29

editor-in-chief , published by Fastcase

0:32

Full Court Press . And now

0:34

here's today's episode . If you like what

0:36

you hear , please give us a rating . So

0:41

welcome back to another episode . We're going to get practical

0:44

on this one , as we sometimes do Try

0:46

to break or mix things up a little

0:48

bit for you , and we're also

0:50

going to have our guest host , sarah Lordback . Sarah

0:53

, she's been doing a series for

0:55

us on legal tech

0:57

and you're lucky to have her , and

1:00

so am I . So here we go . This

1:02

one's going to be about eDiscovery . We have talked

1:04

about this one before . It's a huge

1:06

part of any litigation practice , as anybody

1:09

in litigation knows . It's

1:11

been transformed quite a bit by Technology

1:13

. Assisted Review Tools , or TAR

1:16

is the acronym . There . Ediscovery

1:18

is its own specialty . We've got

1:20

eDiscovery experts on staff , often

1:23

who know all there is to know about the

1:25

technology , the standards , the processes

1:27

or processes and practices

1:29

or practices . Every

1:32

litigator needs to understand how e-discovery

1:34

tools work . You've got to be able

1:36

to answer questions around the approach being

1:39

used , why a specific

1:41

approach was chosen . It's good

1:43

to understand the reliability of the assisted

1:45

review , a human oversight , implementation

1:49

too . You know it's funny , we're talking about humans

1:51

. That's been happening more and more in my conversations

1:54

recently . As we talk about these things , we

1:56

say , well , how are the humans doing ? That

1:58

was never a thing as far as I

2:00

know . I don't know maybe if I worked at NASA

2:03

or something . It's a lot of jargon here

2:05

. It's a lot of acronyms , specialized terminology

2:07

. That's what jargon is and it's always

2:09

changing because technology is changing . Also

2:12

changing are the legal standards and

2:14

ethics ethics

2:17

, competency , issues around technology . Because

2:20

so much of our work is done by a technology

2:22

vendor that has specialized tools . It

2:25

can feel like your review is based

2:27

on blind faith and that

2:29

finding the pieces to support your case requires

2:31

you to rely on dumb luck . Can

2:33

you do more than pray to the document

2:36

? Gods , yeah , for

2:38

those who celebrate , but I'll tell you what . Listen

2:41

as Sarah Lord interviews Lenora Gray . But I'll tell you what . Listen as

2:43

Sarah Lord interviews Lenora Gray . Lenora is an e-discovery expert and

2:45

data scientist , skilled in auditing

2:47

and evaluating e-discovery systems . In

2:50

her role at Redgrave Data , she designs

2:52

and analyzes structured and unstructured

2:54

data sets , builds predictive

2:56

models for use in TAR workflows , implements

2:59

automation solutions and develops custom

3:01

software . Before Redgrave , she spent 12

3:03

years as a paralegal role

3:11

in which she managed discovery teams . She's currently pursuing her MS in data science from Johns Hopkins

3:14

University . She earned her BS in computer

3:17

science from Florida Atlantic University

3:19

and , like I said , I welcome back

3:22

Sarah Lord to guest host

3:24

. Sarah is a former practicing

3:27

attorney with a decade of experience in data

3:29

analytics . She applies her experience

3:31

in law firms and businesses to explore

3:34

and address the cultural and practical barriers

3:36

to diversity in law , supporting

3:38

value creation through legal operations and

3:41

client-first , business-oriented

3:43

practices . In her recent work as managing

3:46

director of legal metrics , she led a team

3:48

of experts focused on providing

3:50

the tools to support data-driven

3:52

decision-making in legal operations

3:54

and closer collaboration between law

3:56

firms and their clients through automation

3:58

and standardization of

4:01

legal metrics . Sarah

4:03

earned her JD from New York University

4:05

School of Law . Also

4:09

, you want to check out our bonus content

4:11

. We're going to have that on YouTube and at

4:13

litigationconferencecom and

4:16

there you will see that Sarah makes it pretty clear

4:18

that she doesn't love that

4:20

initial review part of discovery

4:23

. I don't know . I think that comes

4:25

through . But check

4:27

out the YouTube video on

4:30

YouTube and at litigationconferencescom

4:32

for some bonus content . And

4:35

now here is Sarah Lord

4:37

interviewing Lenora Gray , a

4:40

data scientist at Redgrave

4:42

Data . I hope you enjoy it .

4:46

Okay , a data scientist at Redgrave Data . I hope you enjoy it . Okay , welcome to all you

4:48

podcast people in podcast

4:50

land . We are here to

4:52

talk to Lenora Gray . I'm very

4:55

excited about this discussion . Thank you for joining

4:57

me . Thank you for having me . We

4:59

are going to talk about tar

5:02

and e-discovery , which I know

5:04

everyone loves this topic

5:06

, so it's going to be very exciting . So

5:09

let's just start with the basics . How

5:12

would you describe technology-assisted

5:14

review for those who are new to litigation

5:16

or don't regularly work with

5:18

the review process ?

5:20

Well , technology-assisted review is a way for

5:22

us to use computers and machine

5:25

learning algorithms to assist to

5:27

make document reviews more efficient . So

5:29

we use a form of machine learning

5:31

that's called supervised learning . We

5:34

call it supervised because it learns by example

5:36

. So basically , let's

5:38

say you have a large collection of documents

5:41

that might be important for you to produce

5:43

to the other side , collection of documents that might be

5:45

important for you to , you know , produce to the other side

5:47

and you don't want to

5:49

review them all one by one because it would cost you a gazillion dollars and take you so much

5:51

time to do so . So you can have one or maybe a few attorneys , take

5:54

a small set of those documents , code

5:56

them for whatever issue you have relevance versus

5:59

non-relevance , privilege versus non-privilege

6:01

or any you know categorical issue

6:03

that you can , you know , devise privileged versus non-privileged

6:05

or any categorical issue that you can devise and then the supervised

6:07

learning algorithm uses that training data that your attorney's

6:09

coded and

6:17

it produces a statistical model that we call a classifier . The classifier can

6:19

then be used to predict the coding of the rest of that document

6:21

population . So it'll make a statistically valid

6:24

guess on what the coding would be if the attorneys

6:26

had looked at it and then the classifier

6:28

can also give you a confidence level on

6:31

its prediction . So and

6:33

then we can rank those documents by that confidence

6:35

level . Let's say , if it has a confidence of

6:37

one , I'm 100% sure that this document

6:39

is relevant as opposed to non-relevant

6:41

. We can rank them by the confidence

6:43

level and we can bring those documents to the attention

6:46

of the attorney so they can review the

6:48

relevant ones first . Or

6:50

, if we are very , very confident in

6:52

the accuracy

6:54

of our classifier , we

6:57

can just produce them as they are without the attorneys

6:59

having to look at them at all . So it basically saves

7:02

you a lot of time and effort when it comes

7:04

to conducting a document review . Okay

7:22

, so let's say I'm a litigator beginning

7:24

a discovery process . What are the primary TAR

7:26

options available to me ? And the first phase

7:28

, the active learning , which is the

7:30

supervised learning , iteratively

7:33

sorry , that's a hard word for me selects

7:35

documents for training

7:37

purposes , and reviewers code those documents

7:39

and the resulting coded

7:41

set is used to train and produce

7:43

the classifier , like we said before . And

7:53

the second phase once we are sure of the accuracy and we like what the classifier is doing , it's

7:55

used to classify the rest of the documents , and then the rest of the documents

7:57

might get reviewed by attorneys or , like lower

7:59

cost reviewers , contract reviewers instead

8:01

of senior attorneys , or just

8:03

a privilege screen , or they might be produced without

8:06

further review from anybody . There's

8:09

also TAR 2.0 , which omits

8:11

the second phase . So what it does is it

8:13

in each new iteration

8:15

we produce a new classifier and as

8:18

documents are coded in the collection , the

8:20

coded documents are used to produce a new classifier

8:22

, and this keeps going and keeps going until

8:25

an agreed upon number of

8:27

relevant documents have been found in the collection . It's

8:29

usually around 80% for legal cases

8:31

and , as I said before , there are some

8:33

hybrid options . But these are

8:36

the two basics and ultimately

8:38

, the choice between choosing TAR 1.0

8:40

and TAR 2.0 should be

8:42

guided by specific objections

8:45

, constraints and characteristics of your project

8:47

, your document collection , what it is the

8:49

issues are in your case . So consulting

8:52

with e-discovery experts at this juncture

8:54

kind of like what we do at Record Data they

8:57

can help you determine the most appropriate approach

8:59

, as , like we would know what to look for

9:01

at this stage that would have downstream effects

9:03

in the process going forward .

9:05

These tools incorporate generative

9:08

AI .

9:11

Some of the major players in the eDiscovery

9:13

platform are currently incorporating

9:15

Gen AI into their software , not

9:18

just for TAR review but for like deposition

9:20

prep , summarizing some of those

9:22

documents . But as far as

9:24

incorporating Gen AI at BreadGrid Data

9:26

, we decided to kind of take a step back and

9:29

make sure that generative AI , when

9:31

evaluated against the tools we're currently using

9:34

in eDiscovery , that it matches

9:36

up and is a viable option

9:39

for our clients . So what we're doing

9:41

now is we're currently doing research to

9:43

compare generative AI against our current

9:46

baseline machine learning algorithms like logistic

9:48

regression and support vector machines , which are

9:50

big statistical words , but it's basically

9:52

the computer algorithms that we currently use

9:55

and we're comparing them to GenAI . And

9:57

if GenAI proves to be at

9:59

least as effective as human review

10:02

or like a TAR 1.0 , while

10:04

it's still comparable in time and cost , then

10:06

it could be used in place of the current algorithms we

10:08

have . Yeah , in the resources I provided

10:10

, there's an article that my colleagues

10:12

just wrote Jeremy Pickens and Tara and

10:15

Will Zett that they published in the

10:17

Sedona Journal and it's outlining

10:19

the difference between using traditional machine learning

10:21

and Gen AI in TAR 1.0 workflows

10:24

. Also , gen

10:26

AI has potentials beyond the

10:28

you know regular traditional machine learning

10:30

because it can be used

10:32

for privilege review and summarization of

10:34

groups of documents and it has

10:36

a huge capability to point out

10:38

relevant passages and also provide

10:41

explanations of why it

10:43

classified a document a certain way . So it has

10:45

a leg up if we can prove it's as valid

10:47

and effective as what we're currently using . Also

10:50

, genai is being used behind the scenes

10:53

to improve language processing tools for

10:55

like name and entity extraction , redactions

10:58

and related tasks like that . So it

11:00

kind of has a lot of potential , but we want

11:02

to be sure that it's as effective as what we have now

11:04

.

11:05

You co-authored a

11:07

white paper called Beyond

11:10

the Bar Generative AI as

11:12

a Transformative Component in

11:14

Legal Document Review . Is

11:16

that the article you were just referring to

11:18

, that your colleagues

11:21

also co-authored , or is this a different one ?

11:23

Oh no , this is a different one . We're constantly doing

11:25

research and actual scientific

11:27

experimentation on these tools at RedRigData

11:30

. So our paper Beyond the

11:32

Bar was actually submitted to a

11:34

scientific conference recently . We

11:36

beefed up all of the mathematics to

11:38

make sure it was a scientifically

11:40

strong paper so that we could submit

11:42

it . So it would be beyond just a

11:45

white paper with our opinions on what was going

11:47

on . So for this paper Beyond

11:50

the Bar , which we did a collaborative research with

11:52

Relativity , which is one of the big e-discovery

11:54

platform players , we designed

11:56

a head-to-head comparison of

11:58

first-level human review and generative

12:00

AI . And I say first-level review because

12:03

when you do a fully human review it's usually

12:05

done in two stages . In

12:07

the first stage , you'll have a set

12:09

of contract attorneys , review and label

12:11

, you know , a set of documents , and

12:13

then in the second stage , more senior

12:16

attorneys will come in and check what they

12:18

have coded to make sure that they have done

12:20

the right thing , and there might be , you know , case

12:22

attorneys that you know guide them on

12:24

that path . In our experiment

12:26

we provided a review protocol to

12:28

both the contract attorneys and to the

12:30

generative AI algorithm as like a prompt

12:33

, like you do in ChatGPT , and

12:36

both of those , both of their coding

12:39

results were compared against

12:41

the gold standard of what the senior attorneys

12:43

would have coded the documents and

12:45

in the results we saw that the generative

12:48

AI system had much

12:50

higher recall than a human review , that

12:54

the generative AI system had much higher recall than a human review . It found 96%

12:56

of the relevant documents as opposed to 54% that were found

12:58

by the humans . E-discovery standards usually require around

13:00

80% recall , so this is very , very good

13:02

. However , the precision

13:05

measure for generative AI was lower

13:07

than the human reviewers . We had 60%

13:09

precision as opposed to 91%

13:12

precision from the humans . So

13:14

precision is a

13:17

measure of the documents that

13:19

you found that you coded as relevant

13:22

. How many of them were actually gold

13:24

standard relevant ? So , like , how

13:26

precise is your classification

13:29

? Precise is your classification

13:31

? So , while we were able to get a

13:33

much , much higher recall , it's

13:40

incurring extra costs in producing documents that are not relevant along with the relevant

13:42

ones , although it's finding most of the relevant ones . So there's more work that needs

13:44

to be done . But we have numerous avenues in improving

13:46

the performance . Like we

13:48

didn't do any fine tuning

13:50

of the prompt that was specific to the matter . We

13:53

, you know , did the one prompt and

13:55

so there's much more to come in experiments around

13:57

Gen AI . For document review For

13:59

future view .

14:01

here it seems like we

14:03

may be on a path

14:05

to a quicker review

14:07

process , a more affordable

14:09

review process and , for those of

14:11

us who do not enjoy first level

14:13

review , a less emotionally

14:16

painful process as

14:18

we enter into the future

14:20

of discovery .

14:22

And also a more robust review , because now

14:24

the general AI will be able to tell you why it made a

14:26

decision and also point out the passage of the document that it you know was

14:28

most relevant to its decision . So it kind of like gives you these you know was

14:30

most relevant to its decision . So it kind of

14:32

like gives you these you know breadcrumbs

14:35

in the actual documents to tell you why

14:37

I made this determination . So it's actually

14:39

. You can pinpoint

14:41

where it might go wrong and where it's

14:43

going very right easily without

14:45

you know with actual

14:47

text . So it helps you a lot . So let's

14:49

say that there's some passage , some stock

14:52

passage or some boilerplate

14:54

language in the documents that you haven't seen so far

14:56

, and it's picking it up and you know , making

14:58

it relevant when it shouldn't be . You can

15:00

, you know , flag that and say just

15:03

because it has this doesn't mean it's relevant . So that's not

15:05

going to be , you know , considered in the statistical

15:07

model . Go back and tell me

15:09

if this is really relevant , based on the rest of the text

15:11

.

15:11

That to me sounds very exciting . So

15:14

when will TAR

15:17

be replaced with fully

15:19

robotic review , which I am

15:21

going to call fur , because

15:23

why not ?

15:25

Is this one you made up , because I've never heard

15:27

of

15:29

it .

15:31

Yes , I wanted to catch on , so

15:33

let's all just start calling the next

15:36

phase fully robotic review

15:38

, aka we'll have tar and

15:40

fur . Now , when the fur gets in the tar it might

15:42

get a little messy , but I don't know .

15:43

Oh , my gosh . No , this concept

15:45

was so far off our radar it didn't

15:47

even have a name

15:50

.

15:50

I'm the first . I did it .

15:54

I think it's important to consider that , while TAR is about technology

15:56

, the success of a TAR effort depends

15:59

heavily on the larger e-discovery

16:02

process around it , and that larger e-discovery

16:04

process requires attorneys and

16:06

specialists and technologists

16:08

and people that know how to do the work

16:10

. The machine

16:13

isn't going to do everything , so the

16:15

negotiations around , for instance

16:17

, how the review collection is defined , are

16:20

supremely important , like negotiating

16:22

custodians , date ranges

16:24

, file types , keyword filters . This

16:26

can affect both the volume of the data that has to

16:28

be reviewed and it can also determine

16:30

how much of the data can be reviewed , be reviewed , and it can also determine how much of the data can be reviewed

16:33

through TAR , because there are certain file types that TAR

16:35

is just not , you know , fitted for . Certain

16:38

decisions made on the request for production

16:40

can affect the difficulty of the classification

16:42

task , so it can make it more difficult to TAR to

16:44

determine whether it's , you know , relevant

16:47

, non-relevant , privileged , non-privileged and

16:49

agreed upon deadlines strongly affect

16:51

which TAR workflow you'll be able to use and

16:53

which one is most practical . Some

16:56

of them take , you know , longer to get the classifier

16:58

right than others , so you know that

17:00

also depends . So humans will be needed for

17:02

each of those steps in the process , so

17:04

the time we will have a like a set it

17:07

and forget it review process , I think is a long

17:09

way out .

17:10

Okay , so TAR is still our best option

17:12

and , recognizing

17:14

that the technology is always evolving

17:16

, are there key terms that litigators

17:19

need to understand and key questions

17:21

they should ask to select the right

17:24

tool ?

17:25

Well , sometimes the TAR process are called

17:27

by names other than 1.0 or 2.0

17:29

. So practitioners would want to be aware

17:31

that there are other names for it . I would

17:33

actually encourage them to you know , figure

17:37

out what exactly they are doing and how it compares

17:39

to the standards that we have for TAR 1.0 or

17:41

2.0 , because there might be some hybrid process that

17:43

your provider is using that you don't know . The

17:47

TAR evaluation metrics are important and

17:49

knowing how those work and what they signify . You know

17:51

the TAR evaluation metrics are important and knowing how those work and what they signify

17:53

, the recall and precision numbers . Recall is the percentage of relevant

17:55

documents that you were able to find . Precision

17:57

is how accurate your classifier

18:00

is of the documents that you found

18:02

that you said were relevant . How many of them were really relevant

18:04

, based on gold standard . Those will

18:06

help you set objectives for your TAR process

18:08

and do the negotiation on the front end . You

18:11

know what level of recall are we looking for here . You

18:13

would want to know what a confidence interval is and

18:15

how they work , which is basically a range of values

18:18

that the true value is estimated to

18:20

be in at a certain confidence level . So

18:22

we'll say we found 96%

18:25

of the relevant documents and we're 95%

18:27

confident in that number . It's between , you know , 95.5

18:30

and 96.5 . It's somewhere in there . We

18:32

think it's 96 . And we're 95%

18:35

confident about that . Terms

18:37

that have to do with defining the review

18:39

collection , such as culling , which

18:42

is when you reduce the document population based

18:44

on the date range or file

18:47

type or keyword searches . A lot of that happens

18:49

, you know , happens in the processing

18:51

stages and things . Deduplication

18:53

is brought up a lot , which is where we replace

18:55

multiple identical copies of a document

18:58

in a document collection by just one representative

19:00

copy . The nesting

19:02

, which is where you remove a bunch

19:04

of system files . When

19:07

you collect documents from a computer system

19:09

or files , it comes with all of the accompanying

19:11

little junk files that you don't need to review . So

19:14

some of those are important to know when

19:17

you're negotiating TAR protocols

19:20

and I also have provided some

19:22

resources to get anyone starting and learning

19:24

about TAR and the glossary for TAR and

19:27

for a deeper understanding of TAR and

19:29

AI and e-discovery practice . We also

19:31

have training programs specialized

19:34

that are ideal for legal and IT professionals

19:36

and Red Grave Data

19:38

provides them through our education and training

19:41

program . So if you want to do a deep dive

19:43

on the essentials , the best practices and ethical

19:45

considerations . Consider going

19:48

to our website and looking it up , because we've got some

19:50

really good stuff there Fantastic

19:52

.

19:53

I've heard about system review

19:55

audits and the

19:57

value around those , but

20:00

I don't really know when they should

20:02

be considered . Can you tell

20:04

us a little bit about , from a best practice

20:07

standpoint , when we should be considering

20:09

review system audits ?

20:11

Well , tar processes and systems have

20:13

to be evaluated at a couple stages To

20:16

ensure that the classifier is bringing back the correct

20:18

documents . You will , you

20:20

know , evaluate the classifier when

20:22

it's created . So when you have in like a TAR

20:24

1 workflow , you have those attorneys

20:27

that go and code the small set of documents

20:29

, you use it to build a classifier

20:31

and then you check if the classifier

20:34

is doing its job correctly , if it's pulling back the

20:36

correct relevant documents , and

20:38

then you would also evaluate it at the stage where

20:41

you have to certify to the other parties

20:43

that the result of your coding has

20:45

met the established objectives . We recovered

20:48

at least 80% of the relevant documents

20:50

from this collection . We're certifying

20:54

it using this , and usually this is done by

20:56

random sampling . So

21:12

we'll select a subset of the documents from the entire set at use our

21:14

statistical equations to estimate

21:16

whether or not the TAR process was effective

21:18

, whether or not it's bringing back a requisite

21:22

amount of the relevant documents . However

21:24

, random sampling is just one of many methods

21:26

we can use to do this and

21:29

it's how both and it

21:31

helps us understand how well both the technical

21:33

and manual parts of the TAR process are performing

21:35

. But beyond the actual

21:37

mechanisms of evaluating a TAR system

21:40

. There's also the issue of visibility

21:43

, or the ability to monitor and understand

21:45

and predict outcomes of the TAR process

21:47

. So this goes beyond the TAR

21:49

system and encompasses like the broader

21:51

capacity to oversee

21:54

the review's progress , the efficiency

21:56

, the effectiveness . Our

21:58

approach at Rigor Data is usually

22:00

through our advanced software capabilities where we

22:02

like prepare dashboarding and predictive analytics

22:05

to provide

22:07

users like with an at-a-glance view of what's

22:09

happening in your review right now and

22:11

analyzing trends in your data so

22:13

we can aid in resource

22:16

allocations . Are there certain sets of documents

22:18

that you'll need to increase your contract attorneys

22:20

for ? If you increase the number of contract

22:22

attorneys , can you finish the review faster

22:25

? Will it diminish effectiveness

22:27

? We also use them for

22:29

forecasting for costs , for

22:32

better budgets and for cost efficiency

22:34

and to optimize the review

22:37

process Because sometimes

22:39

you can like switch modes . So

22:41

let's say at one point you're

22:44

reviewing like full

22:46

families . What if I

22:48

switch to only reviewing parent

22:50

documents ? Or what if I switch to

22:53

reviewing this set person , this one ? How long would

22:55

it take me to finish the TAR process ? So

22:57

we can help you detour

22:59

to a different lane if

23:02

it will make the process more optimized and

23:04

we use our software

23:07

and dashboarding for identifying

23:09

potential risks in the e-discovery process

23:11

that might cause issues downstream .

23:14

Okay . So when I'm running my

23:16

discovery process and I'm receiving

23:19

from my vendor the periodic

23:21

updates that tell me how

23:24

the review is progressing , from a recall

23:27

, from those perspectives , that

23:33

is considered a

23:36

system , a review system , audit

23:38

, just that , that

23:40

process .

23:42

Well , it depends on if you're auditing the

23:44

technology , which would be through

23:46

the random sampling and

23:48

, you know , evaluating the classifier itself , or

23:51

if you're evaluating the

23:53

effectiveness of the process , like is

23:55

the process most optimal ? Are

23:58

the costs optimized adequately

24:00

? Are your resources

24:03

being allocated adequately ? So there's

24:05

a whole ecosystem around

24:07

what the classifier itself is doing . That

24:09

also needs to be evaluated for the

24:11

most effective TAR process as

24:13

a whole and not just the part where

24:15

the documents are being coded , because there's this whole

24:18

ecosystem around it that affects it greatly

24:21

that usually attorneys

24:25

don't have much visibility into . But it's

24:27

our passion at Red

24:29

Grave Data to help people make more informed decisions

24:31

about their review processes through

24:34

showing them , through predictive

24:36

analytics , custom software on

24:38

top of your review platforms and

24:40

your search platforms , and

24:43

showing you what is happening

24:45

in the data , what is happening in your review

24:47

, if not real time , near

24:49

time , so that you can make decisions on

24:53

the ball about whether you need

24:55

to switch processes or switch lanes or detour

24:57

to another process to help you out in

25:00

optimizing your review process itself .

25:02

Gotcha .

25:03

Okay , and optimizing your review process

25:05

itself Gotcha , okay . So

25:11

the kind of data that I am used to seeing in a review process is just

25:13

one set of relevant information that could be referred to when people

25:16

use the phrase review system audits

25:18

. But there is this whole other

25:20

set of data that also can be encompassed , whole other

25:22

set of data that also can be encompassed , and

25:24

so it sounds like when it's it's

25:31

another like term of art where , when people talk to you about review system audits , you

25:33

need to clarify and make sure you know what they're including in that

25:35

and that you're getting all of the information

25:38

required to make informed decisions

25:40

Exactly and it's shown

25:42

to you in a way that you can actually

25:45

make decisions about it .

25:47

Let's say , for instance , in

25:50

a regular review , the collection itself is

25:52

growing all the time . Sometimes you're adding

25:54

more custodians , sometimes

25:56

date ranges change or search

26:01

terms change and things like that . How is that

26:03

going to affect the collection ? Search terms change and things

26:05

like that . How is that going to affect

26:07

the collection ? Like

26:10

, in some cases , you don't actually have the hard data as to how that's affecting

26:12

the collection or how long the review will go . But with these systems

26:14

that we place on top of you

26:16

know , let's say , a relativity instance where

26:19

we're pulling numbers and reporting

26:21

from your relativity instance and placing them

26:23

in the dashboard where you can see what's happening

26:25

with the data , I think it

26:27

makes it a lot more of a data

26:31

intensive process is your

26:33

decision-making becomes .

26:35

Okay , so that makes a lot of sense and it sounds

26:37

like it's something you really want

26:39

to incorporate into any

26:41

review of significant

26:44

size . Material size .

26:46

Absolutely , because you know like one misstep can

26:48

cost you so much as far as time

26:50

and effort and attorney review time

26:53

and taking your senior attorneys

26:55

away from strategy and you know back into

26:58

. You know doing samples and

27:00

I think our

27:03

approach here at Red Grave Data is one

27:05

that has great promise . You

27:08

know doing that last 20 percent that

27:10

the review and the e-discovery platforms

27:12

don't do , and also presenting it to

27:14

our customers clients so that they have

27:16

the specific information they need

27:27

in each matter to make the most

27:29

data-driven decision possible

27:31

.

27:32

Thank you so much . I

27:34

do want to encourage the listeners

27:36

. If you haven't read Beyond

27:38

the Bar Generative AI as a

27:41

transformative component in the legal document

27:43

review , I would encourage you to

27:45

do that . I found it really interesting

27:47

and informative and , even if you don't understand

27:50

everything in there , getting

27:53

that high level understanding from

27:55

the article on where the industry

27:57

might go and

27:59

really seeing

28:01

what kind of options

28:03

are being explored , I think is a

28:06

fantastic use of your time

28:08

. So if you're looking to inform

28:10

yourself a little more , I encourage

28:13

you to read this article .

28:15

And if you have any questions , feel free to reach out . I'm

28:17

happy to answer any questions about the paper

28:20

or any of the processes we talked about today

28:22

.

28:23

I do realize calling it an article instead of

28:25

a paper is definitely a misnomer . It is

28:27

definitely a paper . It is beyond article

28:29

length for the

28:31

kinds of articles you're used to getting , but

28:35

I really did enjoy it , so it's fantastic

28:37

. Thank you so much for sharing that .

28:39

No problem , thank you .

28:41

We are also going to be

28:43

making available

28:45

a TAR reference that

28:47

Lenora was kind enough to pull

28:50

together for us , so that

28:52

will be available on

28:54

Emerging Litigation Podcast website

28:56

and I want to thank

28:58

Tom Hagee and the Emerging Litigation

29:01

Podcast for letting me guest host

29:03

once again and the Emerging Litigation Podcast for letting me guest

29:05

host once again .

29:06

That concludes this episode of the Emerging

29:08

Litigation Podcast , a co-production

29:10

of HB Litigation , critical Legal

29:13

Content , vlex Fast

29:15

Case and our friends at Losty Media

29:17

. I'm Tom Hagee , your host

29:20

, which would explain why I'm talking . Please

29:22

feel free to reach out to me if you have ideas for

29:24

a future episode and

29:27

don't hesitate to share this with clients , colleagues , friends , animals

29:30

you may have left at home , teenagers you've irresponsibly

29:33

left unsupervised , and

29:35

certain classifications of fruits and vegetables

29:37

. And if you feel so moved , please give

29:39

us a rating . Those always help . Thank

29:41

you for listening .

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features