Podchaser Logo
Home
Generating the future of art & entertainment

Generating the future of art & entertainment

Released Tuesday, 12th March 2024
Good episode? Give it some love!
Generating the future of art & entertainment

Generating the future of art & entertainment

Generating the future of art & entertainment

Generating the future of art & entertainment

Tuesday, 12th March 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:06

Welcome to Practical AI.

0:09

If you work in artificial intelligence, aspire

0:12

to, or are curious

0:14

how AI-related tech is changing the

0:17

world, this is the show for

0:19

you. Thank you

0:21

to our partners at fly.io, the

0:23

home of changelog.com. Fly

0:26

transforms containers into micro VMs that

0:28

run on their hardware in 30

0:30

plus regions on six continents. So

0:32

you can launch your app near

0:34

your users. Learn more at

0:36

fly.io. Welcome

0:43

to another episode of the Practical

0:45

AI Podcast. I am your co-host,

0:48

Chris Benson. Usually

0:50

I have our other co-host, Daniel Whitenack

0:52

with us. He is not able to

0:54

join today, but we have a great

0:56

show in store. We have with us

0:58

a super interesting guest. You may

1:00

very well, if you follow AI, have heard about

1:02

this guest and this company doing

1:04

some super cool stuff. So I'd

1:08

like to introduce

1:10

Anastasios Dramenidis, who

1:13

is the co-founder and CTO at

1:15

Runway. Sorry, I screwed up your

1:17

name there. Did I get it anywhere close to

1:19

right there? Yeah, all good. Thanks

1:22

so much for having me. No, sorry

1:24

for the stutter there. Thanks for joining us on

1:26

the show. You guys are doing

1:29

some really cool stuff at Runway. Wanted

1:31

you to actually, before we dive fully in, kind

1:33

of tell us a little bit about your own

1:35

background, and then we'll kind

1:38

of dive into kind of the environment

1:40

that you find yourself in and the

1:42

industry and what kinds of problems out

1:45

there are interesting as we dive

1:47

in. First of all,

1:49

CTO of a hot AI company. How did you

1:51

get there? How did you get to where you're at right now?

1:54

Well, the first thing I would say is that I

1:56

did not get here by planning for it, I

1:59

think. in some ways planning

2:01

against being where I am today. So just

2:03

to give a little background. So

2:05

my background is kind of a

2:07

hybrid of engineering and art. So I

2:10

was for the past decade

2:12

or so, I've been kind of in different

2:14

startups working as an engineer at the same

2:16

time having my own art practice. And so

2:18

doing kind of a variety of work in

2:21

media arts and interactive arts.

2:23

Runway was the first time where those two

2:25

kind of different worlds have converged for

2:28

me, but Runway started in art

2:30

school. So this is not really where

2:32

AI companies get

2:34

started usually. So

2:37

my motivation for going

2:39

to art school was actually to take

2:42

a break from technology to really explore

2:44

the more creative and in

2:47

some ways open-ended exploration of

2:49

those technologies without any concern about

2:51

making something that would make a

2:53

commercial sense at some point. But

2:56

it just so happened that I met my

2:59

co-founders there and we started kind of making

3:01

those small tools. And one

3:03

thing led to another. And we

3:05

realized that this was kind of a

3:07

really useful thing to build out and

3:09

kind of spend our focus time on.

3:12

It sounds like it was a bit

3:14

of a passion project without that commercial

3:16

intent up front. In the beginning, you

3:18

kind of fell into it because it was

3:20

what you love. Yeah, and I think that's

3:22

how the best thing gets started very usually.

3:26

And that's been a general pattern,

3:28

I would say, not just at the start,

3:30

but just throughout the way we rebuild the

3:32

company. We're very bespoke that

3:34

we really give every employee that's

3:36

called why greatness cannot be planned.

3:38

And it just talks about this

3:40

idea that when you have very,

3:42

very concrete goals in mind, it's

3:44

actually very often you end up

3:46

not meeting them. And sometimes going

3:49

for the next stepping stone is

3:51

the right approach to actually get to very

3:53

interesting findings or novel insights. That's

3:56

been part of how Runway started and that's been part

3:58

of how Runway has changed. continue to grow.

4:01

But yeah, initially, I would say our

4:03

main goal was like, these

4:05

machine learning models are super difficult

4:08

to understand, super difficult to use,

4:11

especially when we started like around five years

4:13

ago, but they're super interesting for artists and

4:15

they can make really compelling things with it.

4:17

Once they get to the point where they

4:20

can actually use them. At that point, you

4:22

know, generative models, AI was a

4:24

bit at an earlier stage in terms

4:27

of how many people cared about it

4:29

and also the result of those models.

4:31

But it was still even at that point,

4:33

really useful for artists the moment we gave

4:36

the right tools for them to use it.

4:38

And so that was kind of the inception

4:40

of runway. I'm curious, recognizing that there

4:42

wasn't, you know, the master plan that you were

4:44

implementing, you know, there was a bit of serendipity

4:47

to how you arrived there. I am kind

4:49

of curious, you mentioned that you would kind

4:51

of set aside technology before you were going

4:54

back into art right there. And I'm kind

4:56

of curious, did the technologies you're

4:58

in prior to art school play into

5:00

where you've come out here with, you

5:02

know, in terms of runway being that

5:04

end result? Or did you, you know,

5:06

is there any connection there? Or were

5:08

they just you happen to be in

5:10

a different area? And we're

5:12

finding AI? Were you active in AI

5:14

prior to going back into art school?

5:16

My interest in AI kind of goes

5:18

back into like at least high school

5:20

and before. So I've been before runway,

5:23

I was working as a machine learning

5:25

engineer, as a kind of distribution engineer

5:27

at different companies. So definitely had a

5:29

background in this area was very interested

5:31

in AI. My interest was specifically

5:34

in neural networks, which, you know,

5:36

when I was kind of decades ago,

5:38

they had become kind of like

5:40

ignored area of machine learning, like they were

5:42

kind of seen as a dead end, like

5:44

they wouldn't be able to, like, at that

5:47

point, like support vector machines throughout their kinds

5:49

of models were more popular. But

5:51

there was still something very compelling about

5:53

neural networks that made me

5:56

actually get kind of start working

5:58

with them from any type of machine. initial

6:00

projects. So in very interesting

6:02

AI kind of throughout the motivation for

6:05

going to art school was

6:07

just to kind of keep more context on the

6:09

kinds of art school. It was a program

6:11

at NYU that was exploring

6:13

the intersection of art technology. Technology

6:16

was still part of it, but it

6:18

was less kind of technology for the

6:20

sake of technology or for just like

6:22

novelty for the sake of novelty, more

6:25

understanding like how the technology could be

6:27

used in creative ways or in way

6:30

better, maybe unconventional. As you

6:32

were coming into art school and you

6:34

have this background as a machine learning

6:36

engineer and the passion for art,

6:39

what has been your initial vision for that

6:42

industry? Like within entertainment, human creativity,

6:44

which are things that you currently

6:47

are targeting, how did you see

6:49

them? How did you expect to be able

6:51

to impact the industries with AI

6:53

going into the process? So like things are moving

6:56

so fast and we're seeing these amazing technologies which

6:58

we're going to be talking about in the minutes

7:00

to come, but I'm really

7:02

curious what your perspective was about where

7:04

this was going for art

7:07

and entertainment prior to actually arriving

7:09

there. The perspective for us

7:11

has always been that those models, those

7:13

techniques are never going to be a

7:15

source of ideas. They're going to be

7:18

an acceleration and expression of like

7:21

creator's ideas. This is kind of

7:23

a mindset that we started doing those tools

7:25

around and that's why from the beginning we

7:28

started working very closely with filmmakers or designers

7:30

or with artists in making

7:32

those tools and getting their feedback on

7:34

how to make them. The other aspect

7:37

in terms of how we were kind of seeing the

7:39

trajectory of those models was when we

7:41

looked back at like 2017 or

7:43

2018 when we just started kind of working

7:45

on this, the results of

7:48

those models were you know fix-related,

7:50

low resolution, very experimental, you

7:52

know the composition was off, but

7:55

you could see the trend very clearly that you

7:57

know every year the resolution without playing the fidelity

7:59

was in improving at the first physics world way.

8:01

And so it was not a matter of if,

8:03

it was a matter of when this would arise.

8:06

And those things are always really

8:08

difficult, so we didn't really know

8:11

exactly when we're going to get

8:13

to this breakthrough where those models

8:15

really started becoming actually useful, but

8:18

we knew that it was going to happen at some point in the

8:20

next few years. Most people who

8:22

were machine learning engineers, and

8:25

I work with university students a lot and

8:27

people at the company I'm at now and

8:29

previous companies, and that's kind of

8:31

their dream job. And I find it's

8:33

really interesting to me that you said, I'm going

8:35

to set that aside for a little bit and

8:38

go and do art school. What

8:40

was the driving factor for you? Because obviously

8:42

that turned out for your story, that turned

8:44

out to be crucial, that

8:46

juxtaposition, if you will,

8:48

of those different factors. I'm just curious, what

8:50

made you say, I think I'm going to

8:53

put down machine learning engineering for a while

8:55

and go back to art school. I was

8:57

just curious what that was because obviously that

8:59

seemed to create a perfect environment for you

9:01

to spring from. I would say mainly

9:04

just the motivation and the need

9:06

to explore the possibilities

9:08

of something without a very

9:11

clear expectation that

9:13

it needed to result in

9:16

a tool that was necessarily

9:18

useful. Or just being in

9:20

an environment where it can

9:23

have this open and exploration of the

9:25

possibilities of the technology. It

9:28

was less that I wasn't interested in

9:30

machine learning, I wanted to get away

9:32

from it. It was more I wanted

9:34

to explore it in a context where

9:36

there was no expectation that I needed

9:38

to feel something that was commercially valuable

9:40

or super useful. Of

9:42

course, that took a turn and

9:44

I ended up with that was the way

9:46

to get to something that ended up being

9:48

a very good fit for a company. But

9:51

I would say initially I

9:53

was very interested in, at some point,

9:55

I think in 2015, 2016, we

9:58

were just starting to emerge. is going

10:00

to new movement around making art with

10:02

AI. And there were so

10:04

many new explorations, a lot of them in

10:06

front of the open source world. And I

10:08

just started contributing to making kind of small

10:11

projects around making kind

10:14

of tools to make art with AI. And

10:16

so really just want to spend more

10:18

time doing those things and less kind

10:21

of in the, kind of purely in

10:23

the industry working with machine learning because

10:25

I think those two things, you're working

10:27

with the same underlying model from the

10:29

same technology, but the actual results

10:32

are very different that you're creating with them.

10:34

And just one more kind of story from our

10:37

school, like to illustrate, we, like

10:39

one of the first projects that we built

10:41

with my co-founder Chris, was

10:44

this drawing tool essentially where

10:47

there was this model that Nvidia released that was

10:50

meant for kind of self-driving car research.

10:53

And the main idea of this model was

10:55

you could give a kind of a layout

10:57

of essentially a street view. So

11:00

like kind of indications of where pedestrians

11:02

are or like the road user or

11:04

other cars are, and then generate an

11:06

image using that layout. It

11:08

doesn't sound like the most kind of creative

11:10

model or like creative use case for

11:12

a tool. The context of

11:14

that model is very much for like, as

11:16

part of like self-driving kind of car research

11:18

and just kind of creating synthetic data for

11:20

that and so on. But we decided to

11:23

build this drawing tool around it where you

11:25

could define kind of the layout of a

11:27

scene and then generate kind

11:29

of street views based on that layout. We

11:31

saw that the moment we gave it to artists, the

11:34

kinds of scenes that we were creating were super

11:36

different than like what the regular office

11:38

of the model was. So they would

11:40

create like giant pedestrians or like street

11:42

signs flying from the sky. So

11:46

there's the same insight there that, you know,

11:48

you're working with the same types of models,

11:51

the same types of technologies, but seeing

11:53

them with a fresh set of eyes

11:55

and a different perspective makes all the

11:57

different. And so this is what

11:59

I came to. are cool to do

12:01

is to see the same underlying ML

12:03

AI technologies with a new set of

12:06

eyes, exploring new possibilities. And this

12:08

is why we hope to be awesome with

12:10

the tool itself. What's

12:24

up, friends? Is your code getting dragged

12:26

down by joins and long query times?

12:29

The problem might be your

12:31

database. Try simplifying the complex

12:33

with graphs. A graph

12:35

database lets you model data the way

12:38

it looks in the real world instead

12:40

of forcing it into rows and columns.

12:42

Stop asking relational databases to do more

12:44

than what they were made for. Graphs

12:47

work well for use cases with lots

12:49

of data connections like supply chain, fraud

12:51

detection, real-time analytics, and generative AI. With

12:54

Neo4j, you can code in your favorite programming

12:56

language and against any driver. Plus, it's easy

12:58

to integrate into your tech stack. People are

13:01

solving some of the world's biggest problems with

13:03

graphs. And now it's your turn. Visit neo4j.com

13:06

slash developer

13:09

to get

13:11

started. Again,

13:13

neo4j.com/developer. That's

13:15

neo4j.com/developer. So,

13:36

you arrived at art school for that purpose of

13:38

seeing all this through a new set of eyes.

13:41

And you met your co-founder, Chris, and

13:43

you guys had that spark of an

13:45

idea which would become Runway. Can you

13:47

talk a little bit about the insight

13:49

that you had there that created Runways

13:51

before we dive fully into what Runway

13:53

has done since? I'm really curious

13:55

what the moment we're where you and Chris, you

13:57

know, kind of said we have something here. The

14:00

something we're gonna go do. Oh, they're distinct

14:02

bomb I did. You just kind of gradually

14:04

arrived there. What was that moment? Like.

14:06

For you decided it's time to go be an entrepreneur.

14:09

In this context, So. I will

14:11

say was that one moment that come

14:13

with with a turning point though we

14:15

were working in a little different projects

14:17

with a crease in them hundred or

14:19

the other. the other cop on there

14:21

are and the he so this for

14:23

this is kind of a standalone tool

14:25

around on the helping for a say

14:27

are causing for an artist or for

14:29

a specific kind of medium of the

14:31

city of contact. Our time we

14:33

realize that there is another the same

14:36

thing for we have he do for

14:38

it he project and had a fine

14:40

spending as like being able to run

14:42

models would be even more like difficult

14:44

than an a day even like can

14:46

and running and a google and call

14:48

of not to with like to my

14:50

slot from time for like i like

14:52

an artist without any technical on a

14:54

bike around or know how about are

14:57

those models work so then saw idea

14:59

was. Let's. Start from was already

15:01

out during the officers world like there

15:03

is already and her wealth of different

15:05

models that perform different top. Five.

15:07

List: Snake economic really fool around them

15:09

so let's bring the kind of interface

15:11

in the kind of I see you

15:13

in that are these are familiar with

15:15

some other races. But used

15:17

to newman with a refunding out. There

15:20

are differences and possibility. And. On

15:22

the back in foes make the

15:24

main idea of runway initially on

15:26

and also. As. A message before there

15:28

was kind of that. That reason with their

15:30

from the third them. As. Is most

15:32

was becoming better and better. more

15:35

on the up, the ability of

15:37

those models would go increasingly more

15:39

from them. You know the more

15:41

experimental use cases do something that

15:43

actually driving production and it's like

15:45

really really useful for of right

15:47

of credit workflows and we saw

15:49

that happened very quickly as were

15:51

certainly. You. mentioned along the way

15:54

they're that the difficulty of implementing some

15:56

of the models and and even today

15:58

with a number of to choices

16:00

out there, it's still something that

16:02

many companies are contending with is

16:04

how to address models, how to

16:06

train them, where they're going to

16:08

train them, what the deployment, how

16:10

it fits into products. There's a

16:12

gazillion questions out there. You

16:14

were doing this at a moment where

16:16

that wasn't even as sorted as it

16:19

is now, and it's still in development

16:21

at this point. How did

16:23

you manage that? Because when

16:25

I've talked to other people, that's often been one

16:27

of the biggest challenges is just getting the resources

16:29

in place, especially at that time

16:31

when it was still in early development. What

16:34

was that like to try to bring that,

16:36

bring your vision out when obviously the environment

16:38

that we were doing AI in was still

16:41

fairly exclusive in a lot of ways in

16:43

the sense of access to expertise, resources.

16:46

You're in an art school that's designed

16:48

to help you do that, but that

16:50

couldn't have been easy. Yeah, so we

16:53

essentially had to figure out a lot

16:55

of things from scratch as we were

16:57

building this. As I mentioned,

16:59

initially Runway was based around providing

17:02

access to existing open source models.

17:05

We quickly actually realized that we

17:07

needed to build

17:09

a new house, like research team in

17:11

order to really get those models from

17:13

something that makes it good

17:15

demo, or good prototype, or something that's really

17:18

useful. That was actually from

17:21

the first month of Runway became very clear that

17:23

we needed to do this. Of course,

17:25

none of our three had built a

17:28

research team before. I had engineering

17:30

and research in a

17:33

research background, but the experience of how to

17:35

build the team, like what skills to bring

17:37

in was like nobody on the team had

17:39

it. A lot of the things we just

17:42

didn't figure out from scratch. One nice thing

17:44

I would say is that because we started

17:46

so early, we had years to figure this

17:48

out. If you're just

17:50

coming into AI as part

17:53

of building a new company today, the

17:56

time horizon, you need to figure those

17:58

things out in much more in a

18:00

more entrepreneurial fashion. So for

18:02

us, we spent the first years figuring out what

18:05

does it mean to actually build a research organization?

18:08

We think that started, and what does it

18:10

mean to build a robust deployment

18:12

pipeline so that you

18:14

can not only kind of serve those models, but also serve

18:17

them interactively, because a big part of the way

18:19

we build tools that run with

18:21

the interaction, it's a

18:23

very key aspect of really making those models

18:25

useful. When I've talked

18:27

to other entrepreneurs about this, they have

18:30

a tough time, as you're kind

18:32

of getting to the place where you're at now in

18:35

terms of being able to, you may not have the research,

18:38

you're doing amazing research, but you had to

18:40

kind of get from A to B in

18:42

the meantime and kind of keep the

18:44

company alive. How did you

18:46

approach from a funding, customers, things like

18:48

that while you were kind of figuring all these

18:50

things out? Because that strikes me

18:52

as a pretty hard problem to tackle

18:54

as you're moving along,

18:57

but you still have to pay the bills, if you

18:59

will. How did you tackle those kind of issues in

19:01

terms of creating an AI startup that

19:03

couldn't instantly be everything

19:05

that it is today from day one?

19:08

I would say the main insight is

19:10

to, we wanted to make

19:12

sure that runway was useful at each stage

19:14

of evolution. So even

19:17

though the generative models were

19:19

not quite as powerful, when

19:21

we started as they are today, they weren't

19:23

as big a part of the initial kind

19:26

of tool offering and we wanted to make

19:28

the tool as useful from the

19:30

very beginning as possible. So

19:32

the product of runway went through many

19:34

evolutions that really tracked how the kind

19:36

of AI models evolved and at which

19:39

stage they were useful for which things.

19:42

A big part of early runway was building

19:44

out a video editor that was really combined

19:46

some of the more traditional video editing techniques

19:49

with AI-based techniques to speed up the

19:51

process of a lot of video editing

19:53

workflows. And that wasn't necessarily something that

19:55

had generative models powering it, but

19:58

it was a really useful tool that... that really gave

20:00

us a little bit inside about how to

20:02

build tools that are really useful for creative

20:05

workflows and how to really solve real pain

20:07

process. But at the same

20:09

time, while we're building those tools, we're also at this

20:12

research that was ongoing that was

20:14

still remaining and more academic level

20:16

of just really demonstrating how we

20:18

can improve the results of generative

20:21

models. And at some point,

20:23

there was that intersection point where we

20:25

started bringing those generative models to production.

20:28

So the overall strategy was we

20:30

knew that generative models would be really

20:32

powerful given that time and if we invest the

20:35

resources on the research side. At

20:37

the same time, we knew that at the

20:39

beginning, not everything is to be powered by

20:41

generative models. So we're building

20:43

a lot of AI-based tools that incorporated

20:46

that were really useful from the beginning

20:48

and that they were used by VFX

20:50

artists, by video editors to speed up

20:53

a lot of their workflow. Even

20:55

far before, we released things

20:58

like Gen 1 or Gen 2 for

21:00

text-to-video functionality. You're

21:02

saying generative, but it was definitely the

21:04

early days of generative. And you certainly,

21:06

right now, it's all the rage. Everyone's

21:09

talking generative in every context. But

21:12

you had some insights into that. You talked

21:14

about the fact that you guys knew that

21:16

that was going to be the case going

21:18

forward. But to your

21:21

credit, not everybody did. A

21:24

lot of people went, aha, much later than you

21:26

went, aha. And I'm curious,

21:30

is there anything that stands out as

21:32

what drove the insights that you guys had

21:34

and why? Because you were really

21:37

one of the very first to get these

21:39

kinds of functionalities to product. That's

21:42

very notable. And you might

21:44

say the rest of the world didn't,

21:46

not that many. And

21:48

so what were some of the

21:50

things that gave you that confidence to say,

21:53

this is clearly going to be critical

21:55

to our future. This is going to drive

21:57

the industry at an early stage. pioneering

22:00

that thought process, how did you get there?

22:02

From the very beginning, a big part of

22:04

the rendering was working directly with artists and

22:07

building the tools. And so when we gave

22:09

them even early versions of alternative models, we

22:12

could already see that they, like,

22:14

there was really compelling aspects

22:16

of work to me then, even if there

22:18

is also a low resolution and like not

22:21

as high fidelity. So like early

22:23

forms of things like prompt engineering, like figuring

22:25

out how to traverse the latent

22:27

space of those models were still there at

22:29

the beginning of the runway. And

22:31

we saw how artists were engaging with

22:34

them, like how they were kind of,

22:36

they were finding them to be really

22:38

compelling and really useful. And

22:40

so really part of it has been

22:43

this early view into how artists

22:45

with some of more early adopters,

22:48

I would say, were engaging with

22:50

those models and just extrapolating that

22:52

once those models improve, other people

22:54

will equally find them as compelling.

22:57

Working with artists, I think, has been a really

22:59

important part of just really understanding

23:01

some of the future of those models and

23:03

extrapolating of how they would be used. And

23:06

also just looking at the kind of

23:08

history of art and how tool

23:10

making was always part of, like how

23:13

new tools always allow kind of new,

23:15

create a new kind of movements or

23:17

allows new kinds of kind of genres

23:19

to emerge and just assuming and kind

23:21

of predicting that the same would happen

23:24

with those sharing models. Along the way,

23:26

as you were going down this path, what

23:29

stumbles did you have as part of putting,

23:32

because it's quite remarkable because you clearly

23:34

could see the future before you got

23:37

there and with more clarity

23:40

than others that might be in a

23:42

similar position. As you did

23:44

that, what kinds of things did

23:46

you, were either unexpected or

23:49

challenges that were bigger than you thought? You

23:51

know, the things were maybe at a moment

23:53

in time, you were grinding your teeth and

23:55

going, or this is not exactly how I had

23:57

it planned. Do you have any stories to that?

24:00

that affect during this process? Many

24:02

stories and many learnings along the way, for

24:05

sure. I think the biggest

24:07

requiring insight that we've had around how to

24:09

deal for those tools and the things that

24:11

I think is still not fully appreciated today

24:13

is how important control is

24:16

in terms of interacting with those models. And

24:19

so every time we invested into adding

24:22

more ways in which you can really

24:24

control the outputs of the models

24:26

that people were using inside Runway,

24:29

we saw a whole new set of possibilities, a

24:31

whole new kind of usage. So

24:33

that has been a really consistent thing.

24:36

And even at the beginning, we just

24:38

saw that those models had a lot of

24:40

flaws that they might not always, like if

24:42

you have a very simple way of controlling

24:44

them, it might not really give you what

24:46

you want. They might have to do a

24:48

lot of tries with the same old kinds

24:50

of other outputs to get to where

24:53

you want your desired

24:55

result. And so that's really

24:57

what we saw with the early, when

25:00

we first released Gen 2, you could

25:02

only control things with a front. And

25:05

we saw very quickly that that side of

25:07

people just generating like tens

25:09

or hundreds of outputs in order to

25:11

get to the result that they wanted.

25:13

And so we invested, kind of continues

25:15

more and more, adding more and more

25:17

ways in which you can manipulate things

25:19

essentially as the film director would

25:22

think about creating a scene. So

25:24

a film director would have a vision, notice

25:26

of like a description, high level description of

25:28

what the scene is, but how the camera

25:30

moves in the scene, or like how do

25:32

the characters interact with each other. So

25:35

having ways in which you can control really the

25:38

kind of camera motion or like the

25:40

motion, the object motion, like the motion

25:42

of the characters in the scene, like

25:44

all those things that make total sense

25:47

from a career's point of view. But

25:49

they're not necessarily how like maybe ML

25:51

researchers would necessarily think about those models.

25:53

I think that has been always the insight

25:55

that, you know, we never saw

25:58

some negative effects from adding. more this

26:20

is a change log news break pewter

26:23

is the Internet OS pewter

26:26

is an advanced open source desktop

26:28

environment in the browser designed to

26:30

be feature rich exceptionally

26:33

fast and highly extensible it can

26:35

be used to build remote desktop

26:37

environments or serve as an interface

26:39

for cloud storage services remote servers

26:41

web hosting platforms and more I've

26:44

been around long enough to see

26:46

a bunch of these desktop OS

26:48

in a browser window demos and

26:50

toys but this is the first time

26:52

I've been impressed by one enough to keep

26:54

the tab open longer than 30 seconds from

26:57

the URL structure to the cloud

26:59

storage integration to the developer portal

27:01

pewter strikes me as an actually

27:03

viable internet-based operating system with potentially

27:05

real-world use cases and that's saying

27:08

a lot oh and it's also

27:10

entirely built with vanilla JavaScript and

27:12

jQuery so you know the devs

27:14

haven't cargo-cultured together something they can't

27:16

grow and maintain on that note

27:18

they say for performance reasons pewter

27:20

is built with vanilla JavaScript and

27:22

jQuery additionally we'd like to avoid

27:25

complex abstractions and to remain in

27:27

control of the entire stack as

27:29

much as possible also partly inspired

27:31

by some of our favorite projects

27:33

that are not built with frameworks.

27:35

BS code, photo pee, and only

27:37

office. You just heard

27:39

one of our five top stories

27:42

from Monday's changelog news. Subscribe to

27:44

the podcast to get all the

27:46

week's top stories and pop your

27:48

email address in at changelog.com/news and

27:50

to also receive our free

27:52

companion email with even more developer

27:54

news worth your attention once

27:57

again that's changelog.com/news

28:06

So before the break, you brought up Gen

28:08

2. And I'd like, we've had a little

28:10

bit of a history on the development, which

28:12

is fascinating. It's an incredible story you have.

28:14

Tell us all about Runway today.

28:17

You've arrived here, you have

28:19

Gen 2. Just talk a

28:21

little bit about how you're impacting industry

28:23

today. And for listeners

28:25

who haven't been to your website, you

28:27

talk about advancing creativity with artificial intelligence.

28:30

And you specifically note that you're an

28:32

applied AI research company shaping the next

28:34

era of art, entertainment, and human creativity.

28:36

What does that mean in 2024?

28:39

As you're out there in the space, can you talk a

28:41

little bit about the company as it is now? Yes,

28:43

to give some context, Gen 2 is

28:46

a text-to-video and image-to-video generation model. Essentially,

28:48

it takes a description of a scene

28:50

and generates a video output from that

28:52

scene. And it's one of the many

28:55

models that we have at Runway, the

28:57

most well-known one. The broad vision of

28:59

the company has remained the same over

29:01

the last five years. And it's understanding

29:04

and creating the new generation of creative

29:06

tools, and then working with artists directly

29:08

to figure out to help them shape

29:10

those tools as much as possible. And

29:13

so I think where we are today

29:15

is, I would say we're

29:17

still at the very early stages of

29:19

where those models can go. I think

29:21

video generation, this is really the year

29:23

where video generation gets really

29:25

good. And so we're really excited

29:28

to be part of building

29:31

out those technologies and figuring out how

29:33

to work with artists to make

29:35

them as useful as possible. We've seen over

29:37

the past year, I think we're at least

29:39

Gen 2, film studios, streaming

29:42

companies, ad agencies adopting Runway. And

29:44

that adoption is not just from

29:46

kind of individual creators, but it's

29:48

really we see companies starting to

29:50

use those models and incorporate them

29:52

in the workflows. And I think

29:54

it's not going to be a

29:56

binary shift where you go

29:58

from not using generally one. at all

30:00

as part of making video

30:02

or making art, to using it everywhere.

30:05

It's a more gradual transition. And for

30:07

us, the big goal is teaching folks

30:09

how to use those models, supporting all

30:12

the creators that are making interesting things

30:14

with those models. So we have an

30:16

AI film festival that we showcase films

30:19

that use AI in

30:21

different ways. So I would say for us, the

30:23

goal is very much holistic.

30:27

We do the research, we do research

30:29

and development in building up the next

30:31

generation of those models. We build useful

30:33

tools around those models. And we also

30:35

work with artists and with companies that

30:37

want to adopt those models in their

30:39

creative workflows. As you have

30:41

been working into this for years, for most

30:44

of the rest of the world, the past

30:46

few months have been a big eye-opener, particularly

30:48

with big cloud companies producing

30:50

their models and stuff and competing in

30:52

that. There's the obvious aspect of

30:54

you have the industries that you're playing in and that

30:57

you're strong in. But what concerns

30:59

do you have from a competitive standpoint

31:01

against other companies, especially these big, all-encompassing

31:03

cloud companies that are in the AI

31:05

arms race to

31:08

produce the ever larger, more

31:10

capable model? At no point in

31:12

this conversation, have you expressed any concern? Have

31:15

you raised that or anything? Which is quite notable. Usually

31:17

people are a little bit worried about that.

31:20

And you seem very strong in your space.

31:22

How do you see those other big players

31:25

that are out there? Do you see them

31:27

as competitors even? Or are they far enough

31:29

from you that that's not a big deal?

31:31

Or are you so tightly into the industries

31:33

that you're serving specifically that you have a

31:35

huge competitive advantage? How do you see all

31:38

that? Well, we've

31:40

always had the perspective and mindset

31:42

of running around race. And so we

31:44

try not to kind of be too

31:46

distracted, but especially at this

31:48

stage, like there's so much kind of

31:50

noise and discourse around AI that it's

31:52

easy to kind of get back and

31:54

like following the latest developments. So I

31:56

think that's kind of the number one

31:58

aspect. first release Gen

32:00

2 last year, one of our positions that

32:03

was not as popular, I would say last

32:05

year was that video generation

32:07

models were going to be the

32:09

kind of like video with the modality that's

32:12

kind of encapsulated as much

32:14

world knowledge and just for

32:16

myself possible. And last year, the focus

32:18

was on language. And for us, it

32:20

was a big kind of unorthodox to

32:23

kind of maybe pay so much attention

32:25

to video statistically and and claim that

32:27

video generation models were like really the

32:29

way to build really broadly useful AI

32:31

system. And over the past

32:33

month, we've seen more companies kind of

32:35

entering this space of video generation models. And

32:37

so it was not nothing unexpected, like we

32:40

know that those models are going to be

32:42

really useful for a wide variety of use

32:44

cases, they're going to be useful

32:46

beyond reading creative tools, which is really our

32:48

focus. And so for us,

32:50

it's really important to maintain that focus

32:52

of really, like not just building those

32:55

models and like making kind of cool

32:57

demos around that, but really figuring out

32:59

like bridging that gap between, you know,

33:01

those demos and really deploying them to

33:03

product and really getting kind of people

33:05

to use them and getting kind of

33:08

making them controllable. So there

33:10

is still that gap, I would say

33:12

from doing just the research and developing

33:14

the model to actually making those models

33:16

controllable and deploying useful tools. And

33:18

for us always, it has been the focus to

33:20

bridge that gap. And that kind of continues to

33:23

be our focus. So again,

33:25

like video generation models are still very early.

33:27

And like, we haven't seen anything yet

33:30

about what the ultimate be

33:33

capable of, you can imagine, you

33:35

know, a year from now to years from

33:37

now, every company can have like a photo

33:39

realistic video generation model. And that's an assumption

33:41

that we're making that the competitive advantage is

33:44

shift over time. And at that

33:46

point, like, what's the differentiation of

33:48

runway for us, it's always been

33:50

working very closely with artists, building

33:53

really useful tools and bridging and bringing

33:56

making those models really controllable and useful.

33:59

It's fascinating me because I talk

34:01

to so many people in different

34:04

companies and they're busy trying

34:06

to just AI everything and they're kind

34:08

of all about the AI. You guys

34:10

are doing the AI but it sounds

34:12

like competitively having

34:14

been so embedded into

34:16

the artistic ecosystem with

34:18

your tooling is really

34:21

kind of something that keeps you right there

34:24

while everybody goes through the kind of the

34:26

AI model wars in terms

34:28

of trying to produce so much. Do

34:30

you think that long heritage of the tool making is probably

34:33

key to your future in that sense? Is that kind of

34:35

how you're thinking about it? I think it's

34:37

the most important aspect of how we're operating. Otherwise,

34:40

again, it's too easy

34:42

to get lost in the short-term race

34:44

of just having kind of a marginally

34:46

better model for a few weeks versus

34:48

kind of really having the mindset of

34:51

building the most useful tool long-term and

34:53

then obviously updating the model, making sure

34:55

you get state-of-the-art results with it. But

34:57

it's not the goal, it's not the

34:59

focus to have the best model. The

35:01

focus is to get artists to make

35:03

the coolest things or the most compelling

35:06

things with those models. And if that

35:08

remains the goal, then that also informs

35:10

how we build those models. And

35:12

so another aspect of

35:14

Runway is that we have a research

35:17

team and then we also have a creative

35:19

team in-house that works with the research team

35:21

on a daily basis and tries out the

35:23

latest model, informs how do the

35:25

research, like what kind of controls needs out

35:27

the models. And having that

35:29

perspective is really like when I talk

35:31

to researchers that work in academic labs

35:34

or kind of large industry labs, they

35:37

might notice papers about the potential creative

35:39

applications of those models, but they don't

35:41

interact with artists daily. They don't often

35:43

know like is it actually useful or

35:46

is it just a hypothesis that I'm making? And

35:48

I run away, as a researcher, you

35:50

get that feedback on a daily basis.

35:53

And I think that really changes how you

35:55

approach building those models. For listeners,

35:57

you and I can see each other. an

36:00

audio only podcast, but you had this

36:02

glint in your eye a moment

36:04

ago when you were talking about kind of where

36:06

you expected these video models to be going. For

36:09

just a minute there, you reminded me of the

36:11

kind of the kid in the candy store. You

36:13

could see your passion really flying out of your

36:15

eyes there and obviously I'm the only one that

36:17

could see that. Talk a little bit

36:19

about where you think this is going. That's what

36:21

everybody is wondering. There's so many questions, you know,

36:24

that people have in terms of, you know, how

36:26

video fits in their life, what life becomes like

36:28

when you have generative capabilities

36:30

that essentially, you know, simulate life

36:32

in so many ways. What are

36:35

you expecting over the next year

36:37

or so? And like

36:39

I'm not holding you to it obviously, but

36:41

just what do you anticipate might

36:43

happen in the video space generatively? And then

36:46

how would you see it several years out,

36:48

you know, when it's kind of exponentially been

36:50

had time to grow a bit? What does

36:52

that look like to you? The

36:54

way we would like to think about

36:57

those generative video models is we have this

36:59

term of their general world

37:02

model. Essentially, they simulate different aspects of

37:04

the world because in order to kind

37:06

of similar to how, you know, you

37:08

have large language models that have been

37:10

trained with a very simple task, just

37:12

predict the next token in a sentence.

37:15

In order to predict that the next

37:17

token and perform the task really well,

37:19

they have to gain all this understanding

37:21

about different aspects of human

37:23

knowledge, different aspects of the world

37:25

just to solve this task well, because

37:28

they need to complete sentences that might

37:31

come from an encyclopedia or like a

37:33

forum post or it's like a wide

37:35

variety of cases. So we think very

37:38

similarly of how the video distribution models

37:40

operate. In order to predict the next

37:42

frame, you need to gain kind of

37:45

not understanding of basic kind of rules

37:47

of motion or like physics. You

37:50

really need to gain a kind of more

37:52

comprehensive like broader understanding of the world. And

37:55

so like if I think, you know, a

37:57

year from now, where do those models go?

37:59

Essentially because more and more higher

38:01

fidelity simulations of the world, giving

38:04

you the ability to really imagine all

38:06

sorts of different kind of scenarios, like

38:08

build out, tell all kinds of different

38:10

kind of scenarios and stories. And

38:12

I think that the applications of that are

38:14

kind of really, there is kind of a

38:17

wide ranging kind of application that goes beyond

38:20

the kind of creation use cases, which

38:23

I think for us are kind of

38:25

still in mind the focus, but just

38:27

building models that can perceive the visual

38:29

world, like of

38:31

course, like can be used in all kinds of

38:33

other ways as well. Thank you for

38:36

sharing your story. As we finish up here,

38:38

we have a lot of young listeners on the

38:40

show and there is, I guarantee that there are

38:43

quite a few young artists that are

38:46

technically inclined out there, you know,

38:48

high school, maybe early college age, and

38:51

they're listening to this and they're going, that guy just lived

38:53

the life that I'm wishing I could live. You know, that's

38:55

the kind of thing that I wanna do. What

38:58

would you, whether they identify themselves kind of

39:00

as a young artist who's technically inclined or

39:02

technologists who loves art, however they see themselves,

39:05

do you have any guidance on how they might

39:08

step into the future and kind of

39:10

get to that sweet spot for them, given

39:12

the fact that clearly the technology, specifically

39:15

with AI and the artistic role will

39:17

continue to merge and develop together for

39:19

years to come, where should

39:21

they go, what should they do, any thoughts? I

39:23

would say the number one thing is following your

39:26

curiosity and tinkering as much as possible, so there

39:28

is a lot of ways in which you can

39:30

start kind of building those

39:32

models yourself, you can start kind of running them, you

39:35

can start to get kind of an understanding

39:37

of what you can do with them, and

39:39

that's available to really kind of anyone, and

39:42

so really, you can start getting

39:44

involved today in building projects, kind

39:47

of exploring AI or making creative projects with AI.

39:50

That would be the number one thing. It's also,

39:52

I would say for me, planning, trying

39:55

to plan ahead too much has

39:57

never quite worked, really focusing on

39:59

my question. that I can build today, like

40:01

where kind of curiosity and interestingness

40:04

will drive me next, has always been

40:06

kind of the guiding principle.

40:08

And so that would generally be my,

40:11

my recommendation is not trying to think

40:13

of, you know, what, where technology will

40:15

be five years from now, because really,

40:17

nobody can fully plan ahead. But

40:20

rather trying to really build interesting things

40:22

today. It's actually surprisingly, I would say,

40:24

easy to like, if you started making,

40:26

you know, projects open source and just

40:28

showing them to others, it can be

40:30

quite fast, but you can get noticed

40:32

for those projects. And you can like

40:34

start to, you know, build a community

40:36

around and work with other people and

40:38

collaborate on your project. And kind

40:40

of with those collaborations, kind of one by one,

40:42

you can kind of get to a point where

40:45

you can kind of start kind of doing this

40:47

work full time. So, like really focusing

40:49

on the next project, I think, for

40:51

me has been really the way to go. Well,

40:54

Anastas, thank you so much. That

40:56

was fantastic guidance. Appreciate your, your

40:58

perspective, fascinating story leading into this,

41:01

and especially in all the early

41:03

insight that you guys had. Thanks for coming

41:05

on and talking about runway and the world

41:07

in which you guys are trying to make

41:09

a bit better. Appreciate it. Thank you, Chris.

41:19

All right, that is Practical AI

41:21

for this week. Subscribe

41:24

now. If you haven't already,

41:26

head to practicalai.fm for all

41:28

the ways and join

41:31

our free Slack team where

41:33

you can hang out with

41:35

Daniel, Chris and the entire

41:37

changelog community. Sign up today

41:39

at practicalai.fm slash community. Thanks

41:42

again to our partners at fly.io

41:44

to our beat freaking residents, Breakmaster

41:46

Cylinder and to you for listening.

41:48

We appreciate you spending time with

41:50

us. That's all for now.

41:52

We'll talk to you again next time. MMMMMMM

42:04

!

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features