The Best of 2024 with Sarah Guo and Elad Gil

The Best of 2024 with Sarah Guo and Elad Gil

Released Thursday, 26th December 2024
Good episode? Give it some love!
The Best of 2024 with Sarah Guo and Elad Gil

The Best of 2024 with Sarah Guo and Elad Gil

The Best of 2024 with Sarah Guo and Elad Gil

The Best of 2024 with Sarah Guo and Elad Gil

Thursday, 26th December 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

Hi, no priors listeners. I hope

0:02

it's been an amazing 2024 for

0:04

you all. Looking back on this year, we

0:06

I hope it's been an amazing

0:08

2024 for you all. of Looking back

0:10

on this year, we wanted to bring

0:12

you highlights from some of our

0:14

favorite conversations. with First up, we have

0:16

a clip with the one and only

0:18

Jensen Huang, of CEO of NVIDIA, the

0:20

company powering the AI Revolution. Since our

0:23

2023 chat with chat with Jensen, NVIDIA's tripled

0:25

in stock price, adding almost

0:27

100 billion of value each month

0:29

of 2024 the $3 entering the More recently,

0:31

More recently, Jensen shared his perspective

0:33

again with us. why This time no

0:35

longer a is no longer a

0:37

chip company, but a data center

0:39

ecosystem. our Here's our conversation with

0:42

Jensen. And has moved has moved and

0:44

larger, let's say say, like unit support

0:46

for customers. I think about I

0:48

from about it going from single chip to, you know, server

0:50

to and VL 72. How do How do

0:52

you think about that progression? progression?

0:54

next? what's NVIDIA Like, data center? do your

0:56

fact, we built center? In fact,

0:58

way that we build everything

1:00

unless you're building build everything, unless If

1:03

you're developing software, you need

1:05

the computer in its full manifestation. in

1:07

its don't we don't build

1:09

PowerPoint slides the ship the

1:11

chips. And we build we built a

1:13

whole data center. until we get we get

1:15

the whole data center built up. How do you

1:17

know works until you get the whole data center

1:20

built up? How do you know whole data you

1:22

know built up? How do you fabric

1:24

works in. works? all the

1:26

things that you that efficiencies to be, how

1:28

do you know it's gonna really work

1:30

at the scale? you know it's And that's

1:32

the reason why at the scale? not unusual. the reason

1:34

why it's not unusual to see

1:37

somebody's performance. be

1:39

be dramatically lower than their

1:41

peak performance as shown in

1:44

PowerPoint slides. And it's, computing is just not

1:46

used to, it's not what it is just not

1:48

be. You know, I say that the it used to be. of

1:50

I say that the new unit of computing

1:52

is the data center. that's what you us. So

1:54

that's what you have to deliver. That's what

1:56

we Now we build we build a whole thing like

1:58

that, then and then we, for every single thing about

2:00

combination, air cooled, X86, liquid

2:03

cooled, grace, Ethernet, infinite band, envy link, no envy, like, you

2:05

know what I'm saying? We built every you know

2:07

what I'm saying? We built

2:09

every single configuration. five We have

2:11

in our in our company today. year

2:13

we're going to we're going to build. easily five

2:15

more. So if you're serious about software, you

2:17

build your own computers. to build easily five about

2:19

software, then you're going to build your whole

2:21

computer, and we build it all at

2:23

scale. This is the part that is really

2:26

interesting. We build it at scale, then and

2:28

we build it to build whole We optimize it

2:30

build it. full stack, and then, and then we then, and

2:32

then we disaggregate everything. and we sell

2:34

and we sell it in parts. That's the part the

2:36

part that is completely

2:38

utterly remarkable about what we do. do.

2:40

The The complexity of that is

2:42

just insane. And And the reason for

2:45

that is that is be able to

2:47

to be able to our infrastructure into GCP,

2:49

GCP, AWS, OCI, Azure, their control planes, All

2:51

of their control planes, security planes

2:53

are all different. all of all of

2:55

the way they think about their

2:57

cluster sizing, all different. And, but yet we make

3:00

it possible yet, we make it possible

3:02

for them to all invidious architecture, that

3:04

that could could be everywhere. That's

3:06

really, really in the end. thought,

3:08

you know, that we would like to you know,

3:10

that we would like to have

3:12

a computing platform that developers could

3:14

use that's largely consistent, you know, 10% here and

3:17

you know, 10 % here and there

3:19

because people's infrastructure are slightly optimized

3:21

differently differently, and % here and there, but,

3:23

but everything they, they build run

3:25

everywhere. This is kind of of of

3:27

the principles of software that should

3:29

never be given up should never be given

3:31

up and we it quite it quite dearly. makes

3:33

it possible for our software engineers

3:35

engineers to build run everywhere. everywhere. And that's because because

3:37

we recognize that the investment of

3:39

software is the most expensive investment. It's

3:42

easy to test. Look at the Look at

3:44

the size of the whole hardware

3:46

industry. And And then look at the

3:48

size of the world's industries. trillion on

3:50

top of this of trillion industry. And

3:52

that tells you something. tells The software

3:54

that you build, you have to. you

3:56

build, you have to, you know, you basically as long as

3:58

you shall live. as you shall live. course, have

4:00

to mention our conversation with the lovely

4:02

Carpathie, we dig into the future of

4:04

AI as an exo -cortex, an extension

4:07

of human cognition. Andre,

4:09

who's been a key figure in

4:11

AI development, from open AI to

4:13

Tesla to the education of us

4:15

all, shares a provocative perspective on

4:17

ownership and access to AI models,

4:19

and also makes a case for

4:21

why future models might be much

4:23

smaller than we think. If we're

4:25

talking about a exo -cortex, that feels

4:27

like a pretty fundamentally important thing

4:29

to democratize access to. How do

4:31

you think current market structure of

4:33

what's happening in LLM research, you

4:35

know, there's a - small number

4:37

of large labs that actually have

4:40

a shot at the next generation

4:42

progressing training. Like training. how does that

4:44

translate to what people have access

4:46

to in the future? So you

4:48

kind of alluding to maybe is the state of

4:50

the ecosystem, right? So we have kind of like an

4:52

oligopoly of a few closed platforms and then we

4:54

have an open platform that's kind of like behind so

4:56

like MetaLama, et cetera. And this is kind of

4:58

like mirroring the open source kind of ecosystem. I do

5:00

think that when this stuff starts to, when we

5:02

start to think of it as like an exo -cortex,

5:04

So there's the, there's a saying in crypto, which is

5:06

like not your keys, not your, not your, Not

5:08

your keys, yeah. Like, it the case that if it's

5:10

like not your weights, not your brain? That's

5:12

interesting, because a company is effectively controlling your -cortex

5:14

in their part of Yeah, it starts to feel

5:16

kind of invasive. If this isn't my Exocortex. I

5:18

think people will care much more about ownership, yes. Like, yeah,

5:21

you realize you're renting your brain,

5:23

like it seems strange to rent your

5:25

brain. thought experiment is like, are you willing

5:27

to give up ownership and control to rent

5:30

a better brain? Because I am. yeah. so

5:32

I think that's the trade -off, I think we'll see

5:34

how that works, but maybe it's possible to like

5:36

by default use the closed versions because they're amazing,

5:38

but you have a fallback in various scenarios.

5:40

And I think that's kind of like the way

5:42

things are shaping up today even, right? Like

5:45

when APIs go down to some of the closed

5:47

source providers, people start to implement fallbacks to

5:49

like the open ecosystems, for example, that they fully

5:51

control and they're empowered by that, right? So

5:53

So maybe that's just the extension that will look

5:55

like for the brain, you fall back on

5:57

the open source stuff. Should

5:59

I? I? anything but most of the time you

6:01

of the time it's quite important that the open source

6:03

stuff to. that think so source stuff this is

6:06

not like an obvious point or something that people

6:08

maybe agree on right now But I think

6:10

100 or I guess one thing I've been wondering

6:12

about a little bit is now, but I think 100%...

6:14

I is the smallest about a model that

6:16

you can get to in some sense,

6:18

either in parameter size or everyone to

6:20

think about it? you can little bit curious

6:22

about your sense, either thought a lot about size

6:24

distillation, small models, you know? a I

6:26

think it can be surprisingly small. view.

6:28

And I do I do think that the

6:30

current models are wasting a ton

6:32

of capacity remembering stuff that doesn't matter,

6:34

matter. they remember they remember like the

6:36

ancient ancient... the data set is not curated.

6:38

is not curated, the best. Yeah, and I think this will go

6:40

away and I think we just need to get

6:42

to the cognitive core think we I think the cognitive core

6:45

can be extremely small And I think it's just

6:47

this thing that can be And if it needs to look up

6:49

information, it knows how to use different tools. Is that

6:51

it needs billion up Is that

6:53

how billion? how to use a billion, tools. and surprises. We'll

6:55

probably get to that point and the models can can

6:57

be very, very small. And I And I

6:59

think the reason they can be very small

7:01

very small is think just like distillation works and

7:03

maybe like the only It would say. the only

7:05

works say. well. works like Distillation is where you

7:07

get a really big model you get huge

7:09

amount of compute or something like that, of computers,

7:11

something a very small model. a very Our conversation

7:13

with Brett Taylor, with Brett Taylor, board member

7:16

and founder of Sierra of a really different

7:18

picture of how we interact with businesses

7:20

in the future. in the future. a

7:22

clip of Brett explaining company agents and why

7:24

the website is going to take a back

7:26

seat. going other category seat. The the

7:28

area that my company which is there that

7:30

what I call company agents. and

7:32

is it's really agents. simply about

7:34

automation or autonomy, but in this

7:36

world of conversational AI, AI, how does

7:38

how does your company exist

7:40

digitally? I'll use the metaphor use the

7:42

metaphor of it. We're 1995. you know, if you you

7:44

existed digitally about having a

7:47

website and being in Yahoo and being

7:49

in Yahoo In 2025, existing digitally

7:51

will probably mean having a branded

7:53

AI agent that your customers can

7:55

interact with to do. to that they

7:57

can do on your website, do on your it's

7:59

whether know, asking about your products

8:01

and services. and services, doing commerce, doing

8:03

doing customer service. That domain, I think, is

8:05

that domain I think shovel ready

8:08

right now with current technology. Because again,

8:10

like the persona base agents, it's

8:12

not the proverbial proverbial ocean. You know, you you

8:14

know, you have well -defined processes for

8:16

your customer experience, well -defined systems that

8:18

are your systems of systems of it's

8:20

really about saying in this world

8:22

where world where... We've gone from websites to

8:25

apps to now, conversational experiences, what is the

8:27

conversational experience you want around your brand?

8:29

And it doesn't mean it's perfect or

8:31

it's easy, otherwise we wouldn't have started

8:33

a company around it, but it's least

8:35

well -defined. around it. And I think well defined. And

8:37

I Right now right now in AI, you're

8:39

working on artificial general intelligence, your

8:41

version of your means something different, and

8:43

that's means That's just a different

8:45

problem to be solved. to be solved. But I

8:47

think, you know, particularly areas that Sierra works a

8:49

lot a lot of the companies that

8:51

you all have invested in, is in,

8:53

it saying, you there there shovel ready opportunities

8:55

right now with existing technology? And I

8:57

absolutely think there are. Can you describe

9:00

the, think there a shoveling cycle of building

9:02

a company of Like, what is the gap

9:04

between research and reality? Like, how do you,

9:06

reality? do you invest in as an engineering

9:08

team? as an how do you understand the

9:10

scope of different customer environments? the Just, like,

9:12

what are the sort of vectors of investment

9:14

here? And maybe as a as a starting point

9:17

worth even be worth also defining are what are

9:19

the products that provides today for its customers customers

9:21

and then. do you want that to go and

9:23

then maybe we can feed that back into

9:25

like what are the components of that because

9:27

I think what are the folks are really emerging

9:29

as a leader and you're vertical but it'd

9:31

be great just for a broader audience to

9:33

understand what you focus on. your sure. just give

9:35

a couple of examples to make it concrete.

9:37

So if you buy a new what you focus on.

9:39

Yeah, sure. you're having technical issues with your speaker,

9:41

you get the dreaded So light, new you'll now

9:43

chat with having is powered by speaker, help you

9:45

dreaded flashing help you debug whether it's

9:47

a hardware Sonos AI, -Fi issue, by Sarah. things

9:49

like that. you're a If a XM

9:51

subscriber, their AI agent agent is named which

9:53

which I think is a delightful

9:55

name. and it's everything from upgrading and

9:58

downgrading your subscription level to if you... get

10:00

a trial when you purchase a new vehicle, speaking

10:02

to you about that. Broadly

10:04

speaking, I would say we help companies

10:06

build branded customer facing agents. And branded

10:08

is an important part of it. It's

10:10

part of your brand, it's part of

10:12

your brand experience. And I think that's

10:14

really interesting and compelling because I think

10:16

just like you know, I go back

10:19

to the proverbial 1995, you your website

10:21

was on your business card. It was

10:23

the first time you had this digital

10:25

presence. And I think the same novelty,

10:27

and probably we'll look back at the

10:29

agents today with the same sense of,

10:31

oh, that was quaint. You know, I

10:33

remember if you go back to the Wayback

10:35

Machine, you look at early websites it

10:37

either someone's phone number and that's it, or

10:39

it looked like a DVD intro screen

10:42

with like of graphics, You a lot of

10:44

the agents that customers start with are

10:46

often around areas of customer service, which is

10:48

a really great use case. But I

10:50

do truly believe if you fast forward three

10:52

or four years, your agent will compass

10:54

all that your company does. I've used this

10:56

example before, but I like it. But

10:58

just imagine an insurance company, all that you

11:00

can do when you engage with them.

11:02

Maybe you're filing a claim. Maybe you're comparing

11:04

plans, We were talking about our kids earlier,

11:06

maybe you're adding your child to your insurance

11:08

premium when they get old enough to have

11:10

a driver's license, all of the above you

11:12

know, all of the above be done by

11:15

your agent. So that's what we're helping companies

11:17

build. Next, we talked to the Sora

11:19

team at OpenAI, which is building an

11:21

incredibly realistic video AI generation model. In

11:23

this clip, we talk about their research

11:25

and how models that understand the world

11:27

fit into the road to AGI. Is

11:30

there anything you can say about how

11:32

the work you've done with Sora sort

11:34

of affects the broader research roadmap? Yeah,

11:37

so I think. something here is about. the

11:41

knowledge that Sora ends up learning about

11:43

the world just from seeing all this

11:45

visual data. It understands 3D, which is

11:47

one. cool thing because we

11:49

haven't trained it to. We didn't

11:51

explicitly bake 3D information into it whatsoever.

11:54

We just. trained it on video

11:56

data, and it learned about 3D because 3D

11:58

exists in those videos. And it learned. learned

12:00

that when you take a bite out of a hamburger

12:02

that... you leave a bite mark. So

12:04

it's learning so much about our world. And.

12:07

And when we interact with the

12:09

world, so much of it is visual.

12:11

So much of what we see and

12:13

learn throughout our lives is visual information.

12:15

So we really think that. just

12:17

in terms of intelligence, in terms

12:19

of leading toward AI models

12:21

that are more intelligent, that better understand

12:23

the world like we do. This will actually

12:25

be really important for them to have

12:27

this grounding of like, hey, this is the

12:29

world that we live in. There's so

12:32

much complexity in it. There's so much about

12:34

how people interact, how how. things happen,

12:36

how events in the past end up

12:38

impacting events in the future, that this

12:40

will actually need to just much more

12:42

intelligent AI models more broadly than even

12:44

generating videos. It's almost like you invented

12:46

like the future of visual cortex plus

12:49

some part of the uh,

12:51

reasoning parts of the brain or something

12:53

of simultaneously. Yeah. And, and that's

12:55

a cool comparison because a lot of

12:57

the intelligence that humans have is

12:59

actually about world modeling, right? All the

13:01

time when we're thinking about

13:03

how we're going to do things. We're playing

13:05

out scenarios in our head. We have dreams

13:07

where we're playing out scenarios in the head.

13:09

We're thinking in advance of doing things. If

13:11

I did this, this thing would happen. If

13:14

I did this other thing, what would happen,

13:16

right? So we have a world model and

13:18

building Sora as a world model is very

13:20

similar to a big part of the

13:22

intelligence that humans have. How do

13:24

you guys think about the sort of

13:26

analogy to humans as having a very

13:28

approximate world model versus something that is

13:31

as accurate as, let's say, a physics

13:33

engine in the traditional sense? right? Because

13:35

if I hold an apple and I

13:37

drop it, I expect it to fall

13:39

at a certain rate. most humans do

13:41

not think of that as articulating a

13:43

path with a speed as a calculation.

13:46

Do you think that sort of

13:49

learning is like parallel in large

13:51

models? I think it's

13:53

a really interesting observation. I

13:56

think how we think about things is that it's

13:58

almost like a deficiency, you know, in humans that it's

14:00

not so high fidelity. So know the

14:02

fact fact that we We actually can't do very

14:04

accurate long -term prediction when you get down to

14:06

you get down to a narrow set of

14:08

physics is something that we something that

14:10

we can improve upon with some

14:12

of these systems. And so we're

14:14

optimistic that Sora will supersede that

14:16

that kind of capability and will, long

14:18

enable it to more more intelligent

14:20

one day than humans as world models. models.

14:22

But it is, it is an an

14:24

existence proof that it's not

14:26

necessary for other types of intelligence.

14:29

Regardless of that, it's still

14:31

something that that SORA and models in the the future

14:33

will be able to improve upon. improve Okay, so

14:35

it's very clear that the trajectory that the for

14:37

like throwing a football is gonna be

14:39

better. a football is the next, next

14:41

versions of these models versions of these

14:43

let's say. than I could add something

14:45

to that, this relates to the

14:47

paradigm of scale. of scale and the

14:49

better lesson a bit lesson a bit about

14:52

how we want methods you as you increase

14:54

compute, get better and better. and and

14:56

something that works really well in this

14:58

paradigm. paradigm. is doing the simple. but

15:01

challenging just of And

15:03

you can try data. with more And

15:05

you can try coming up with

15:07

more complicated tasks, for example, something

15:09

that. but use video explicitly,

15:11

but is maybe in some like

15:14

space that simulates approximate things or something.

15:16

But all this complexity actually isn't

15:18

beneficial when it comes to the scaling

15:20

laws of how methods improve as

15:22

you increase scale. And what works really

15:24

well as you increase scale is

15:26

just just predict data, and that's

15:28

what we do with do with we just

15:30

predict text. And And that's exactly what

15:32

we're doing with we're with which is

15:34

we're not making some complicated trying

15:36

to figure out some new thing to

15:38

optimize. We're saying, hey, the best

15:41

way to learn intelligence in a scalable

15:43

matter matter is to just predict data. That

15:45

makes sense in relating to what

15:47

you said, Bill, what you said, Bill, just get

15:49

much better with no necessary limit

15:51

that approximates limit that We also

15:53

sat down with Dmitry We also sat down

15:55

with of Dolgov. Today, the company

15:57

is scaling its self -driving fleet,

15:59

completing over ,000 fully autonomous rides

16:02

per week in cities like San Francisco

16:04

and Phoenix. It's my

16:06

favorite way to travel. way to In

16:08

this trip, Dmitri explains why

16:10

achieving full autonomy, removing the driver

16:12

entirely the achieving entirely, accuracy 100% than

16:14

99 .99 % accuracy in self -driving

16:16

is much harder than it might

16:18

appear. appear. Why is it breaking from like,

16:20

you know, know, say let's say advanced

16:22

driver assistance to work in more seems

16:25

to work in more and more

16:27

scenarios versus let's say full

16:29

autonomy? autonomy? What's the the what's the Yeah.

16:31

Yeah. the number of the number of Right?

16:33

And it's the nature of the nature of

16:35

this right? If If you think about where

16:37

we started in 2009, in of

16:39

our first first, you know, mile stops. One goal that

16:41

we that we set for ourselves

16:43

was to drive to drive, you know, 10 routes. Each

16:45

Each one was 100 miles long

16:47

all over the the Bay Area. You know, freeways, downtown

16:49

San downtown San Francisco, around Lake

16:51

Tahoe, everything. And you had to do

16:53

100 miles miles no intervention. So the

16:55

car had to drive autonomous from

16:58

beginning to end. That's the goal

17:00

that we created for ourselves. the

17:02

goal that we have seen about a dozen of us. You

17:04

know, about a months, of us took that. 18

17:06

months. We achieved that. 2009. no image net,

17:09

no confidence, no no no

17:11

big models, tiny computers, you know,

17:13

you know, how this, right? to get started. get

17:15

It's always been the property. with

17:17

every wave of technology, wave of very

17:19

easy to get started. to get started.

17:21

that the hard problem. And it's of the curve has

17:23

been getting that the early and steeper, but that's

17:25

not where the complexity is. The complexity is

17:28

in the long tail of the many, many,

17:30

many nights, and you don't see that if

17:32

you go the a prototype, if

17:34

you go for a is in the long and

17:36

this is where many, spending all of

17:38

our, that's the only hard part of

17:40

the problem. if you go for a I guess

17:42

nowadays it's always been getting easy

17:44

with every technical kind of cycle.

17:46

is where, you know, can take with all

17:48

of the advances the advances of an AI. and

17:51

especially in the general area I will and the LLLMs and

17:53

you can take kind of of an

17:55

almost off -the -shelf, know, know, transformers

17:57

are amazing. are amazing. The are

17:59

amazing. amazing. you can take a kind of

18:01

a VLLM that can accept images

18:04

or accept images where

18:06

you can give it text decoder

18:09

where you can text and you of book fine

18:11

tune it you can just it with just

18:13

a little bit of data from, go say,

18:15

say, camera data on a car to instead of words

18:17

to trajectories or or you know you decisions you might

18:19

the thing because a know box, because a black

18:21

box you know you been trained for a little

18:23

bit, and you fine tune it a

18:25

little bit. And you let me fine tune it a

18:27

I little think, if you and like that without you know

18:29

science to build ask AV good this is

18:31

what they would do. science to And out

18:33

of the box, today this something do yeah and out

18:35

it's amazing, right? something that of transport. right yeah

18:37

the power of is mind with powers violism is just a

18:39

little bit of effort, right you get something

18:41

a the road, and of a you bit of effort

18:43

don't get effort of miles something you get your mind. something

18:45

you get But then is is that enough? that

18:48

Is that enough to remove the

18:50

driver and drive and of miles and have

18:52

a safety record? You know, that is just really better than humans?

18:54

No, right? I guess this guess know, with

18:56

every, you know, evolution, technology, and a breakthrough

18:58

in AI, they've seen, like, I'm about, appreciate

19:00

it. Up next, we we have my dear

19:02

friend, friend, Dylanfield, CEO of Figma. Dylan shares

19:04

his prediction for how user interfaces

19:07

will evolve in an AI -driven world.

19:09

While While many predict a shift

19:11

toward conversational or agent -based interfaces, Dylan

19:13

suggests that new interface paradigms will

19:16

complement existing ones. He also

19:18

highlights the exciting potential of potential

19:20

and intelligent cameras as the

19:22

next frontier as input methods. in How

19:24

do you think about How you

19:26

UI the general? in general? that's is gonna come

19:28

with AI. AI. A lot A lot of things are kind

19:30

of collapsing in the short in chat interfaces. There's a

19:32

lot of people talking about a future a lot of world.

19:34

talking which a away with most UI world,

19:37

which and away with stuff happening in the

19:39

background. it's just all How do you think

19:41

about where UI is going in general right

19:43

now? where UI is I kind of in kind of

19:45

comes back to the rabbit point I was making earlier. comes back

19:47

to the rabbit there's a lot earlier.

19:49

Yes, there's a lot of happening in terms

19:51

of agents, but I think like in terms

19:53

of the way that we way that we... use UI

19:55

UI to interact with agents, we're just

19:57

the beginning. just the beginning and

20:00

I think that the interfaces will get

20:02

more sophisticated, but also even if they

20:05

don't. I suspect

20:07

that it's just like any new media

20:09

type. When it's introduced, it's not like

20:11

the old types go away. right? just

20:13

because you have TikTok doesn't mean

20:15

that you, you know, hunger watch

20:17

YouTube. Even if it's true

20:20

that a new... form

20:22

of interaction is via chat interfaces, which I'm not

20:24

even sure I believe. But if, if we take that

20:26

as a prior. on the

20:28

No Briars then I think

20:30

that you still have UI and

20:32

actually I think you have

20:34

more UI and more software. than

20:36

before. Do you have any predictions

20:38

in terms of multimodality? Like, do

20:40

you think there's more need for boys? Like, Like,

20:42

so, you know, a lot of the debates have

20:44

is like, When are you going

20:47

use voice versus text versus other

20:49

types of interfaces? And,

20:51

you know, you imagine arguments in all sorts

20:53

of directions in terms of you know, do you

20:55

use what and things like that. And a

20:57

lot of people are not a lot, some people

21:00

are suggesting because of the rise of multimodal

21:02

models, you'll have like voice input or more things

21:04

like that because you'll be able to do

21:06

real -time sort of smart contextual semantic understanding

21:08

of like conversation. And so you

21:10

have more of a verbal

21:12

conversational UI versus a -based UI. And

21:14

so it kind of changes how

21:16

you think about design. So, I

21:18

was just curious if you have any thoughts on

21:21

that, that sort of future stuff. There's

21:23

all sorts of contexts where a voice

21:25

UI is really important. And

21:27

I think that,

21:29

it might be that we find that

21:31

voice UIs start to

21:33

map to more traditional UIs. because

21:37

it's something that like you obviously do. in

21:40

a more generalized way. Bye.

21:43

But yeah, I mean, personally,

21:45

I don't want to navigate.

21:47

the information spaces that I

21:49

interact with every day,

21:51

all day via voice. I

21:54

also don't want to do it in

21:56

minority report style on the vision pro

21:58

exactly either. Maybe with with

22:00

a keyboard and mouse and like

22:02

an amazing vision pro, monitor setup or Oculus,

22:04

like that could be cool, but I be cool, but

22:06

I don't want to do the board

22:08

thing. thing. it's And so it's so I

22:11

It's interesting, so I think that we get

22:13

these new glimpses that interaction patterns that

22:15

are really cool cool and the natural natural inclination is

22:17

to extrapolate and say they're gonna be

22:19

useful for everything. be useful I think that they

22:21

have I think of their role. sort of

22:23

their role and it doesn't mean that It

22:26

doesn't mean that they're gonna be ubiquitous

22:28

across every interaction we have. we have. But that's

22:30

but that's a natural cycle to

22:32

be in. it's good. It's I think

22:34

it's good, of that it's healthy to have

22:36

sort of that. couldn't do, because

22:38

around. have that, then you do,

22:40

because if you don't have that, then

22:42

you don't get to find out. of And

22:44

so I'm supportive of people exploring as

22:46

much as possible you kind of that's how

22:48

you kind of HCI on HCI how I use out

22:51

how to use computers and to the fullest

22:53

potential that could be possible. One of

22:55

of the things I am really bullish on mean, you

22:57

I mean, you just think of it as

23:00

an input mode or a peripheral, but

23:02

it's really hard for people to

23:04

describe things visually. And so

23:06

the idea of intelligent cameras,

23:08

even in the in the basic

23:10

sense. Oh, it worked. It Oh, it I

23:12

think that's actually a I think that's actually a

23:14

really fun space to be, as you

23:16

said, because I because I

23:18

actually think that will be useful. And

23:21

it's And it's something that every

23:23

user is capable of, right, pictures, capturing

23:25

video. And so I think that'll

23:27

be that'll be, I'm that. bullish wrap

23:29

up our favorite moments from

23:31

2024, we have scale CEO, scale CEO

23:34

In this clip, he shares his bold

23:36

take on the road to take on the Alex

23:38

also dives into why generalization into why harder

23:40

than many think, and why solving these

23:42

niche problems and more data in evals

23:44

is key to advancing the technology. and evals

23:46

you believe about AI that other people don't. My

23:48

My biggest belief here is

23:50

that the the the to AGI

23:53

is a lot more that looks a

23:55

lot more like curing cancer

23:57

developing a a vaccine. what I mean

23:59

I mean by that. is I think think that the build

24:01

AGI is to build in, you know, is

24:03

going to be to solve a bunch of small

24:05

problems that where you to have to

24:07

solve a bunch of small problems where

24:10

you don't get that much positive

24:12

leverage between the next one problem to solving

24:14

the next problem. of, you know, it's like curing

24:16

curing cancer, which is then have to

24:18

then each in to each individual cancer

24:20

and solve them independently. And eventually over

24:22

a multi over a time frame, we're

24:24

going to look back and realize that

24:26

we've built AGI, we've cured cancer. But

24:28

the path to get there will

24:30

be be this like, know, quite road of solving

24:32

individual capabilities and building and

24:34

of. individual to support this

24:37

end mission. to support this end

24:39

I think a lot of people in

24:41

the industry paint of people in like, you know,

24:43

the path to just like, you know, we'll get there,

24:45

we'll like, you know, we'll get there, we'll like,

24:47

we'll solve it in it in one fell swoop.

24:50

this is a lot of think there's a lot

24:52

of implications for how you actually think

24:54

about you know, technology arc arc and... and how society is

24:56

going to have to have to deal with it.

24:58

I I think it's actually a pretty

25:00

bullish case for society adapting the technology

25:02

because I think it's gonna be consistent,

25:05

slow progress for quite some time and

25:07

society will have time to fully sort

25:09

of have time to fully to the technology that

25:11

develops. the When you that a problem

25:13

at a time, right, if we just at

25:15

a away from the analogy a little

25:17

bit, should I think of that as think

25:19

of that as... generality of multi -step reasoning is

25:21

really hard, as you as you know, Monte research

25:24

is not the not the answer that people

25:26

think it might be. run into scaling walls like We're

25:28

just gonna run into scaling walls, like

25:30

gonna of like of what are the dimensions

25:32

of I multiple problems? the main thing

25:34

there's I think there's very limited generality

25:36

that we get from these models for

25:38

and even for for for example, saying there's

25:40

no understanding there's no positive transfer from

25:42

learning in one modality to other training

25:44

off of a off of a bunch of

25:46

video doesn't really help you that

25:48

much with your text problems and vice

25:50

versa. and vice And so. so I I

25:52

think what this means is like, each like each

25:55

sort of each niche of each niche of

25:57

capabilities or each area of is going to

25:59

to require. separate flywheels, data flywheels, to

26:01

be able to push through

26:03

and drive performance. You don't You

26:05

don't yet believe in video as basis

26:07

for world model that helps. I think it's great narrative. I

26:09

narrative. I don't think there's strong scientific

26:11

evidence of that yet. Maybe there will

26:13

be eventually. but I But I think that

26:15

this is the, I think the base case, think the

26:18

say, case, let's say, is one where,

26:20

you know... There's not that much

26:22

generalization coming out of the the models. And

26:24

so just actually just need to slowly solve

26:26

lots and lots of little problems to ultimately

26:28

results in AGI. Thank you so much for

26:30

listening in 2024. 2024. We've really enjoyed talking

26:32

to the people reshaping the world for

26:34

for AI. If you want If you want to more

26:36

deeply dive into any of the conversations you've

26:38

heard today, today, we've linked the full episodes

26:40

in our description. Please let us

26:42

know who you want to hear from on what your

26:44

questions are for next year. Happy holidays. are for next

26:46

year. Happy holidays. us on Twitter at

26:48

No Prior's Pod. Subscribe to our YouTube channel

26:51

if you want to see our faces. Follow

26:53

the show on Apple podcast, Spotify, or

26:55

or wherever you listen. That way you

26:57

get a new episode every week. every

26:59

And sign up for emails find

27:02

transcripts for every episode at no every.com.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features