Generative AI's uncanny valley: Problem or opportunity?

Generative AI's uncanny valley: Problem or opportunity?

Released Thursday, 12th December 2024
Good episode? Give it some love!
Generative AI's uncanny valley: Problem or opportunity?

Generative AI's uncanny valley: Problem or opportunity?

Generative AI's uncanny valley: Problem or opportunity?

Generative AI's uncanny valley: Problem or opportunity?

Thursday, 12th December 2024
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:08

Welcome to this episode of the of

0:10

the Thoughtworks I'm your host, Lily

0:12

Ryan, and I'm speaking to you

0:14

from the lands of the from the

0:16

in Melbourne, Australia. people in Today, I'm

0:18

talking with with Shrini a principal technologist

0:20

at ThoughtWorks, about a piece that

0:23

he recently co -authored with our

0:25

other podcast co -host, Ken co-host in

0:27

the MIT in the MIT Review. This piece

0:29

is called Reckoning with Generative AI's

0:31

Uncanny Valley. And and today to going to

0:33

unpack a lot of these concepts

0:35

and how they interact with AI

0:37

AI in the way that the way that we're seeing

0:40

it today. Srinni, welcome to the show. show. Thanks.

0:42

Thanks for having me here. me. what

0:44

is the valley? I've heard this heard this is

0:46

a term term I've heard a lot,

0:48

but I've usually heard it in

0:50

the context of robotics or animation

0:52

or something like that. How does

0:54

it apply to the the

0:56

collection of technologies that we're calling AI

0:58

these days. days? Yeah, like the concept of uncanually,

1:00

like like I said, you'll probably commonly

1:02

associate with with but we have

1:04

seen this kind of appeared in

1:07

other terms as well, right? other terms as

1:09

For example, in the past, when

1:11

we are developing when we are we

1:13

have been seeing that. we have been

1:15

that when you're app like app, like a

1:17

lot and take a teams take a

1:19

I have a like, okay, I have

1:22

this into a app I wrap this

1:24

into an app and release it, right? When the users

1:26

use they can quickly quickly, like, like,

1:28

for mostly functional, but they can

1:30

also quickly also quickly at time feel in

1:32

the sense right experience doesn't match the

1:34

it should be matching with

1:36

the ecosystem. it should be matching with the ecosystem so

1:38

that's like a little bit of

1:40

surprise where like a a feeling, of surprise

1:42

know, feeling. feeling like you negative

1:44

the negative a of things. side of

1:47

things that's of we've kind of called

1:49

valley. And we see here

1:51

in other technical domains domains

1:53

like they said in stuff. And

1:55

we stuff seeing we also now seeing in

1:57

LLLM What's special about LLM? I

1:59

think think Ella? way for that

2:01

sense is like, it's kind of a

2:03

non-deterministic software in some sense, right? You

2:06

can give the same inputs, but not

2:08

necessarily you will get the same outputs.

2:10

Like you can, you can fine-tune and

2:12

instruct, try to prompt and try to

2:14

make it the same answer, even then

2:16

there is no, not sure that you

2:19

get the right answers. And that's in

2:21

opposition to something like a calculator where

2:23

you would input one plus two and

2:25

you would expect it to be three

2:27

every single time. That's more of a

2:30

deterministic approach. So the non-deterministic, you're getting

2:32

a lot of different kinds of output

2:34

and the same input could generate a

2:36

wide variety of different responses depending on

2:38

what the models got in its training

2:40

data set, how it's been weighted, that

2:43

kind of thing, right? Yeah,

2:45

exactly. Like, you know, a small change

2:47

in input or like a small change

2:49

in something that the model relies on

2:51

can have a, you know, quite different

2:54

outcome, right? Like, output. So I think

2:56

this is kind of the behavior of

2:58

LLLM, which is actually quite useful as

3:00

well, like because we want to, we

3:03

can see the use cases of LLLM

3:05

where it's writing text or essays where

3:07

you actually that's useful to not add

3:09

the same essays for millions of people

3:12

giving a same input. the what's useful

3:14

it can also the other side of

3:16

things where it can not be useful

3:18

right like this kind of this kind

3:21

of makes it makes it what the

3:23

surprise can be either surprise or it

3:25

could be disappointing or this disappointing is

3:27

probably where the it kind of starting

3:29

feeling it today on Kenny Valley This

3:32

uncanny valley thing, it really describes

3:35

that kind of creepy feeling that

3:37

you experience when when you're expecting

3:39

something to be sort of human,

3:41

and then something happens there's some

3:43

kind of output or appearance or

3:46

something, just even for a moment

3:48

that reminds you or indicates that

3:50

the thing that you're speaking to

3:52

is not human. I think classically

3:54

they spoke about zombies as kind

3:57

of falling into this category, you

3:59

know, it's a bit spooky, it's

4:01

human shaped, but the behaviors and

4:03

movements and that kind of thing

4:05

don't feel the same. And so

4:08

for a lot of people, that's

4:10

where that uncatiness can come from.

4:12

So I can see, yeah, how

4:14

it applies in this situation too,

4:17

where you have these moments where

4:19

suddenly, if you're interacting, say, with

4:21

a chat bot, in that case,

4:23

you end up with something that

4:25

really surprises you. And

4:28

when it comes to how we

4:30

look at this in a business context,

4:32

why is it important to sink this

4:35

through? Like I said, the uncanny

4:37

early experiences could be for many reasons.

4:39

Like one of them, like evolutionary

4:41

responses, right? Like a brain or why

4:43

it could act anomalies, right? Or it

4:46

could be a neurological thing, like,

4:48

because you feel that way. But it's

4:50

also very common about the mismatch in

4:52

the expectation of the output, right?

4:54

So that's right, I would probably tie

4:57

it back to the expectation on the

4:59

capabilities of LLLM. Like the capabilities

5:01

of LLLM is kind of wide ranging.

5:03

And now if you see the

5:05

market, there's a bunch of models, there's

5:08

large language models, there are even specialized,

5:10

small language models, right? Like, how

5:12

would someone make sense of this, like,

5:14

whether this is the right model for

5:17

me, for this context of the

5:19

job I'm picking up, right? If the

5:21

mental model for picking that up

5:23

kind of right model is not said,

5:25

that kind of also reflects in the

5:28

user experience. So it's not just

5:30

about what the user can see. It's

5:32

also like how the developer is building

5:34

around it. need to have to

5:36

pick the right one, right? This is

5:39

ever a bigger impact in business. Like

5:41

most commonly you're probably hearing like

5:43

a lot of people are like experimenting

5:45

or like doing a POC for

5:47

a genity away. And you ask them,

5:50

have you deployed production? The answer is

5:52

mostly like we are not sure

5:54

about the implication or amplification. So we're

5:56

holding it at the POC, right? I

5:59

think that's where can he really

6:01

experience and the impact of that is

6:03

quite evident in making or even

6:05

like productionizing the model? Because nobody wants

6:07

to put that experience in front of

6:10

a customer. And I think we've

6:12

seen many cases that have made the

6:14

news where because of that experience happening

6:16

in front of a customer, there's

6:18

been a lot of negative backlash, brand

6:21

reputation impact, that kind of thing. And

6:23

yeah, it's tough. I like the

6:25

distinction that you drew there about the

6:27

way that people are thinking about

6:29

it in terms of their different roles.

6:32

What kinds of mental models about this

6:34

kind of current crop of generative

6:36

AI technologies have you seen out in

6:38

the wild? You know, when you're speaking

6:41

to clients, when you're speaking to

6:43

developers and when you're speaking to members

6:45

of the public, I think all three

6:47

of those groups can really have

6:49

some very different ideas about what generative

6:52

AI is and can have different

6:54

ideas in different contexts. What kind of

6:56

mental models have you seen people have?

6:58

And what do you think are

7:00

the helpful ones? Yeah, I could start

7:03

with like what I can hearing from

7:05

clients or people who are first

7:07

looking at the LLLM, right? Like the

7:09

first time they look at it

7:11

is like, oh, okay, this model can

7:14

recently answer me and can learn from

7:16

my interaction, right? Like, the most

7:18

common question I had is like, okay,

7:21

I deploy an LLLM model now based

7:23

on how I'm using it, does

7:25

it automatically learn from it, right? Like

7:27

the LLLM's. like they have the certain

7:30

modeling and then they can generate

7:32

responses but not necessarily learning from your

7:34

question. They have the concept of

7:36

context window where they can use like

7:38

certain range of information to give us

7:41

you know more more response that

7:43

you're looking for but like outside that

7:45

context window there's not much of memory

7:47

on its own like it doesn't

7:49

learn from here right like. I also

7:52

think this is partially driven by

7:54

like how the industry is approaching this

7:56

right like the one of them commenting

7:58

is like when people interaction

8:01

with LLLMs, they kind of feel

8:03

like, okay, it feels like it's

8:05

reasoning, right? But it's not really

8:07

reasoning, and it doesn't really have

8:09

the basic capabilities, or it's not

8:11

both like humans, right? It is

8:13

a probabilistic model, like also called

8:15

as like stochastic patterns, right? Where

8:18

it is a, it just generates

8:20

an output based on a bunch

8:22

of parameters. It's not reasoning for

8:24

you, but it feels like reasoning,

8:26

right? This is tying back to

8:28

like, the output looks like more

8:30

human, so you kind of assume

8:32

that that mental model, it can

8:34

reason. So you could

8:36

for example ask a chat, but

8:38

something like, please explain how Leonardo da

8:41

Vinci's Mona Lisa ties to the

8:43

cultural aesthetic of the time in which

8:45

it was painted and it could probably

8:47

say something about that, but you could

8:50

then ask it. a classic something

8:52

like, you know, how many ours in

8:54

the word strawberry and it's not going

8:56

to be able to get to that

8:59

point. That's, you know, that's quite

9:01

a famous one and I know that

9:03

Open AI has named one of their

9:05

more recent models codenamed Strawberry because

9:07

they were looking at how do we

9:10

solve this problem. It's fundamentally one of

9:12

the things that I think the industry

9:14

is looking at quite a lot,

9:16

right? Yeah, so like the concept are

9:19

moving towards the, you know, AGI, artificial

9:21

general intelligence. It's like, that's the goals

9:23

probably useful or like it's as

9:25

its own, you know, people looking outcome

9:28

of it, but like it's, it's kind

9:30

of also kind of leads to the

9:32

common understanding of like, okay, I

9:34

should. should I be worried about this

9:37

or like can it actually do all

9:39

the thing I can think of?

9:41

That doesn't actually help people to be

9:43

very realistic or like to be actually

9:46

interacting with these models with the right

9:48

mindset. The common story I think

9:50

of like the one thing I found

9:52

as a mental model I heard in

9:55

good terms are a framing for that

9:57

that's useful as like the AI

9:59

as the stone soup like this is

10:01

of like, you know, three strangers coming

10:04

into a village and they say,

10:06

like, okay, do you have anything to

10:08

make a soup? And the villagers are

10:10

like, no, we ever have food, but

10:13

you don't want to share with

10:15

you. It's like, it's okay, we'll make

10:17

a stone soup, which is very delicious,

10:19

right? They kind of start making, putting

10:22

stones and start boiling order and

10:24

then they say, like, oh, it stays

10:26

nice. It just need a little bit

10:28

of something, do you have. The

10:30

slowly, actually the soup, the outcome is

10:33

delicious. It's not because of the stones

10:35

or the boiling, or the things that's

10:38

strange about them. It's about the

10:40

people, what the information they share, the

10:42

data I share, right? So I think

10:44

this kind of sets the context of

10:47

the LLLM as a good model,

10:49

where it's like, you know, the outcome

10:51

is actually determined by a lot of

10:53

things that you share. It's not

10:55

a magical thing to solve. So you

10:58

kind of be aware of like how

11:00

the whole process works and what it

11:02

translates into. I feel like this

11:04

is a good analogy or metaphor that

11:07

actually can help people to approach LLLM

11:09

or gender AI with the right, in

11:11

over the right mindset. Right,

11:14

so it gives you this basis

11:16

from which to structure your own

11:18

thinking by kind of offering you

11:20

the possibility of anything and then

11:22

having it reflect back to you

11:24

the things that you put into

11:27

it and build on top of

11:29

it over time as a, not

11:31

quite a collaborator in that way

11:33

that it is actively participating with

11:35

you, but that it provides the

11:37

framework for you and the people

11:40

working with you on this to

11:42

get something substantial out of it.

11:44

Right? Totally. You also mentioned about

11:46

like, you know, what's the other

11:48

kind of mental models? People are,

11:50

the common one I can see

11:53

with business is like they kind

11:55

of have this assumption, I can

11:57

replace something with this. right? Like,

11:59

you can call it as an

12:01

automation bias where they want to

12:03

automate a big chunk of thing

12:06

without understanding the ecosystem, right? I

12:08

can quote a recent example with

12:10

the coding assistance, right? Like with

12:12

the coding assistance and other stuff,

12:14

it's the outcome they want to

12:16

replace as writing, generating code, right?

12:19

The recording assistance can generate code.

12:21

but then as it makes it

12:23

easy to read those code in

12:25

production when you're trying to deploy,

12:27

right? That's those kind of things,

12:29

right? The assumptions people take is

12:31

like I can substitute some part

12:34

of the current workflow entirely with

12:36

the LLLA, the margin, or the

12:38

AI, often leads to other problems

12:40

where I think you can learn

12:42

from. the past where automation with

12:44

humans and machines is a similar

12:47

thing where if you take like

12:49

for example pilots using autopilot right

12:51

when they are using autopilot they're

12:53

not exercising their regular skills but

12:55

they also need to know about

12:57

how autopilot engages right There's a

13:00

bit of complexity that they also

13:02

need to understand to be able

13:04

to step in and this engage

13:06

our pilot in a place where

13:08

it's relevant, right? So it's not

13:10

like suddenly I run out of

13:13

pilot, I have taken care of

13:15

these things completely out of the

13:17

human control and doing a best

13:19

job or effort. It actually increases

13:21

the cognitive load for the people

13:23

to understand how the machines operate.

13:26

I think for a human machine

13:29

interaction, there's two bits that's very

13:31

important. One is the observability part

13:33

of it where you kind of

13:35

want to know what the process

13:37

is replacing and how it's behaving

13:39

given and generated more model. It's

13:41

not a deterministic model. It makes

13:43

it challenging. The other bit is

13:46

able to direct the machine or

13:48

the LLLN to a specific task,

13:50

right? You can do that bit

13:52

of a prompting saying like, hey,

13:54

can you focus on this bit

13:56

and change it? But again, it's

13:58

not as direct to as an

14:00

in-home. a driving a bike or

14:03

a driving a car, right? So

14:05

this bit of, the key bit

14:07

of automation that actually makes this

14:09

human machine or human automation works

14:11

well as the observability and directly,

14:13

but the nature of a limb

14:15

child enjoys this, right? Like the

14:18

nature of a limb is not

14:20

quite, quite adaptive for this kind

14:22

of expectation. I think that's where

14:24

the industry is generally moving about

14:26

like how we can improve the

14:28

reliability or how we can bring

14:30

the observability and and the directability,

14:32

and that's probably where these mental

14:35

models can help them to approach

14:37

it very realistically. It's interesting that

14:39

you talk about how it can

14:41

be used to augment a lot

14:43

of things, and you also mentioned

14:45

the problems around things like pilots,

14:47

understanding autopilot. There's also

14:50

a concern that if you're adapting

14:52

too many automated models in your

14:54

workflows that you will replace the

14:56

people who are learning from doing

14:58

those kinds of tasks as introductory

15:01

tasks, you know, if they've just

15:03

started out their careers. and something

15:05

can replace them in that space,

15:07

how then do we get people

15:10

who are going to be seniors

15:12

in, say, 10 years time, you

15:14

know, learning the foundational skills right

15:16

now? Is there any mental model

15:19

there that you can? Is there

15:21

an approach that you would recommend

15:23

that people take if they are

15:25

looking at these kinds of solutions

15:27

such that you are not only

15:30

helping people with these things, but

15:32

also helping people grow their skills

15:34

from first principles? Yeah,

15:37

I think that's very interesting question

15:39

in the sense like, you know,

15:41

it's a very important one. When

15:44

you're replacing something, it's kind of

15:46

taking away the opportunity or the

15:48

current process, right? So it is

15:50

important for people to steady the

15:53

ecosystem, what is the second order

15:55

effects of doing this and what

15:57

it can imply. there is an

15:59

impact on another role on another

16:02

person, then opportunity is taken away,

16:04

then there needs to be other

16:06

set of things that need to

16:08

be pulled off. I read about

16:10

this somewhere where, you know, the

16:13

doctors using robotics these days, right,

16:15

like which can, they can actually

16:17

make it very minute operations very

16:19

clearly. which probably previously like it

16:22

required 10 assistants in the operation

16:24

room. Now it's robotics with, you

16:26

know, robotic hands. They just have

16:28

to, they just have to use

16:31

the robotic to control and make

16:33

the operation. Usually they don't have

16:35

that many assistants on the room.

16:38

When the assistants look at it,

16:40

they also just look at the

16:42

machine operating, they're not really looking

16:45

at the person, how they are

16:47

controlling another stuff, right? It kind

16:49

of changes the learning ecosystem for

16:51

even new people. And due to

16:53

the cost and other things involved,

16:55

again, only people who are experts

16:57

are kind of given opportunity to

16:59

operate robotics, right? When someone is

17:01

new in their career, there's not

17:03

enough ways for them to actually

17:05

do a robotic operation, unless they

17:07

are experts, right? In a way

17:10

I feel like these LLLM or

17:12

these advanced techniques kind of very

17:14

useful tool for exports, right? Because

17:16

they kind of know what they

17:18

can, what the machine can do

17:20

and they can also can, can

17:22

also sense that it's not going

17:24

right or things like that very

17:26

effectively. Didn't they actually their knowledge

17:28

and the people's knowledge is the

17:30

key bit in making that all

17:32

outcome? maybe like 10x effectual, right?

17:35

But at the same time it

17:37

breaks the metaphor of, or like

17:39

not even a metaphor, but like

17:41

it breaks like how other people

17:43

can engage in this. Maybe there

17:45

needs to be a second simulation

17:47

place or a digital twin or

17:49

those other things needs to be

17:51

explored where people can safely move

17:53

from, you know, wise into slowly

17:55

able to run, like, but just

17:57

replacing something. of have this second-over

18:00

effects. Outside this I can also

18:02

think of the you know the

18:04

alpha-go like I think the trend

18:06

after all the color goes like

18:08

the the players ranking started stacking

18:10

up. like as in like previously

18:12

level X is probably the golden

18:14

standard but after also the people

18:16

learn from the interaction because experts

18:18

can learn from that. The player

18:20

ranking went really really high and

18:22

like it's now somewhere differently but

18:25

also it changes like how people

18:27

can learn like previously something learned

18:29

naturally now people have to learn

18:31

it by like looking at how

18:33

the machine performs like it's like

18:35

it's Someone said, like, it's like

18:37

learning from cheat, cheat, cheat, right?

18:39

Like, you're missing that foundation. You

18:41

ought to jump from somewhere else

18:43

to be at that point, but

18:45

like, how do you bridge the

18:47

gap? Yeah, it's raised a lot

18:50

of interesting questions, to be honest.

18:52

I think it puts a lot

18:54

of responsibility on us when we're

18:56

implementing systems like this to think

18:58

through those second and third order

19:00

impacts on the world around us,

19:02

not for just solving the problem

19:04

right in front of us, but

19:06

for what that problem was doing

19:08

in the world in the first

19:10

place and whether while it provides

19:12

a point of friction, that friction

19:15

is also something that we have

19:17

been learning from over time as

19:19

a part of the larger ecosystem.

19:21

I liked the part in the

19:23

article that you wrote, where you

19:25

were talking about using the uncanny

19:27

valley sensation, that kind of misstep

19:29

eariness, as a way of reminding

19:31

ourselves that what we are dealing

19:33

with here is not actually human,

19:35

and that it is a machine

19:37

and that there are things that

19:40

we need to be mindful of

19:42

around it, that this particular feeling

19:44

is a good reminder of that

19:46

and a good tool and that

19:48

kind of thing. In terms of

19:50

the way that people have approached

19:52

the uncanny valley problem, if indeed

19:54

it is a problem, in the

19:56

past in many different kinds of

19:58

fields, there are some folks who

20:00

have tried to bridge that uncanny

20:02

valley and say, right, well, the

20:05

goal should be that we're getting

20:07

right to the other side of

20:09

the valley where we need to

20:11

have something that will behave exactly

20:13

like a human being at all

20:15

times. You see this in some

20:17

kind of applications of animation, for

20:19

example, where this problem comes up

20:21

where people are pushing the boundaries

20:23

to try and have more and

20:25

more realistic animation over time. At

20:27

the other side of that, you

20:30

have folks who are saying that

20:32

the uncanny valley should just be

20:34

avoided, that it's not actually worth

20:36

bridging, that the goal is not

20:38

about that, but it is to

20:40

stay on this side of the

20:42

valley and work with the thing

20:44

within the limits that it has.

20:46

I think for me, one of

20:48

the interesting... applications of this mindset

20:50

was when for example they were

20:52

when they were animating toy story,

20:55

which was very early on with

20:57

computer generated images, and and they

20:59

chose this story based on you

21:01

know plastic toys. because they knew

21:03

that they could animate something like

21:05

that without falling into this uncanny

21:07

valley trap, that this fit the

21:09

level that the technology was at

21:11

at the time, and that they

21:13

could tell a good story within

21:15

the bounds of what was possible,

21:17

but that they firmly acknowledged that

21:20

those boundaries were there? Do you

21:22

think that the uncanny valley should

21:24

be bridged? Or do you think

21:26

it should be avoided? I think

21:28

it's more about being aware of

21:30

that kind of can trigger your

21:32

set of things that you can,

21:34

right? Like it's not something can

21:36

be avoided, as in like, it

21:38

is something to be looked at

21:40

at every aspect of the work,

21:42

as in like the pastory example

21:45

you said, right, like it is

21:47

not just for the end user,

21:49

it's also the people who are

21:51

building it, or like it can

21:53

show up in different places, right?

21:55

The moment if kind of you

21:57

show up in different places also

21:59

a good reminder to like how

22:01

we can understand this better or

22:03

how we can make it easier

22:05

for someone else to actually work

22:07

with this right like I can

22:10

think of like of the day

22:12

is like AI is like it

22:14

feels like it can reason but

22:16

we kind of know it's a

22:18

probabilistic model like having that mental

22:20

model kind of helps you to

22:22

actually approach it with like what

22:24

else people can assume from it

22:26

how we can bridge gaps around

22:28

it or like how we can

22:30

enable people to be aware of

22:32

those things right not giving them

22:35

a bit of a guided exploration

22:37

or like some guidance on it

22:39

kind of lead into a lot

22:41

more assumptions building on top of

22:43

it. But where it's like it

22:45

kind of leads to the place

22:47

where it definitely going to be

22:49

an uncanny valley with the problem

22:51

of that is candy catastrophic for

22:53

how they interpret it right like.

22:55

I think this kind of happened

22:57

in like data explorations at other

23:00

places as well. I would say

23:02

like for example, even if you're

23:04

taking data algorithm, right, if it's

23:06

blindly implemented, the outcome of the

23:08

choice of the machine is just

23:10

implemented without considering the context and

23:12

other stuff, it can. and shown

23:14

in many places that it is

23:16

not a right outcome where people

23:18

need to actually question what the

23:20

system is providing, why it's providing,

23:22

and is the context is right?

23:25

And they should have the ability

23:27

to actually contextualize it. in addition

23:29

to thinking about the uncanny valley

23:31

as a mental model for working

23:33

with generative artificial intelligence technologies in

23:35

the piece that you wrote you

23:37

also talk about the kinds of

23:39

tooling that people can implement to

23:41

help minimize the likelihood of these

23:43

kinds of uncanny valley moments if

23:45

The choice of generative AI tooling

23:47

is something that people want to

23:50

pursue and go ahead with in

23:52

the context where they're trying to

23:54

apply it. Can you talk a

23:56

bit about the ways that you

23:58

would recommend people think about this

24:00

in practice when it comes to

24:02

minimizing those moments? When it comes

24:04

to minimizing these moments, I probably

24:06

say like for example going back

24:08

to the point of like it

24:10

is not a deterministic software like

24:12

how we can make it more

24:15

understandable like how we can make

24:17

it reliable right like there are

24:19

modern techniques like evolves or testing

24:21

things are like god rails where

24:23

you can make sure like these

24:25

kind of things cannot cannot leak

24:27

from the from the building the

24:29

system. You can also think about

24:31

how people interact with it, not

24:33

just the customers, like the developers,

24:35

like the developers, because their assumptions

24:37

about the model can actually lead

24:40

to an outcome that is not

24:42

for it. Every time a new

24:44

model is coming out, like there's

24:46

a question also comes in like,

24:48

is this better than the other

24:50

one, right? It's how much better

24:52

than this. The question again is

24:54

like the context of the problem

24:56

you're solving, right? a new model

24:58

can do better in something else,

25:00

or like it can be slightly

25:02

better. So as an engineer or

25:05

a developer was building this ecosystem,

25:07

it's like to build measures and

25:09

toolings, like testing techniques, or like

25:11

establishing benchmark for like, okay, if

25:13

my LLLM can answer these questions

25:15

reliably for these many things, then

25:17

I can have some trust on

25:19

it. Or I also put some

25:21

godrails which says like these things

25:23

cannot be. cannot happen, right? Like

25:25

it's kind of building that foundation

25:27

layer you can call it or

25:30

like a layer of things that

25:32

can help you to have a

25:34

quips and stuff like that form

25:36

my context of the problem. Again,

25:38

I stress the context of the

25:40

problem because like there may be

25:42

like off-the-shelf guardrails and other stuff

25:44

which can like very generic. It

25:46

may give you a false sense

25:48

of promise saying like, okay, this

25:50

is taken care of. But the

25:52

context of the problem is the

25:55

keys like where your business use

25:57

case. are the problem you're trying

25:59

to slout. have a very specific

26:01

question. That's where you need to

26:03

go back to the best principles

26:05

and find out, like, what is

26:07

the right questions that I need

26:09

to make sure if this model

26:11

applies, is still valid, right? I

26:13

think in a way it's like,

26:15

that's the role of developers or

26:18

engineers is to make sure it's

26:20

like, Oh, the shiny models keep

26:22

dropping, but like, does it make

26:24

sense to adopt, whether it's safe,

26:26

whether it's be safely built, how

26:28

people are going to interact with

26:30

the system or their controls, are

26:32

their controls and other places where

26:34

they can actually adapt to this,

26:36

those are the things I would

26:38

say. Yeah, it's the

26:40

right tool for the right job and

26:43

making sure that if it is the

26:45

right tool, you're thinking through the context

26:47

that it's going to be applied in

26:49

and those knock-on impacts on the wider

26:51

ecosystem, I think. Is there anything that

26:53

you would like to leave our listeners

26:55

with today in terms of thinking through

26:57

how the uncanny valley could be useful

26:59

to them in their work? Working

27:01

with LLLMs, right? I think, like, previously, automation is

27:04

probably done focused mainly on, like, replacing windowing tasks,

27:06

or, like, can I do this efficiently, or under

27:08

part of times using missionaries and other stuff? But

27:10

LLLMs, kind of, I think people are engaged in

27:12

a creative work, right? Like, as in, like, can

27:14

you generate a bunch of ideas or an approach,

27:17

right? This is kind of more of a cognitive

27:19

work. When you're doing your cognitive work, it is

27:21

very important to, even when you're working with someone

27:23

else, it's kind of establishing what is the outcome

27:25

you want to achieve, or like my assumption about

27:28

the other person, or like what are they thinking

27:30

through, is this correct? It's like reducing those feedback

27:32

loops, or like how we can course correct, right?

27:34

That's becomes paramount. So I think that's where like

27:36

for everyone, this concept of uncanny really is like,

27:38

Kainov could be a good reminder about, okay, do

27:41

I understand? the other side

27:43

of side of the things, and if, why

27:45

fall into this, this, right? Like, what

27:47

other things be looked be looked

27:49

after? also have someone else

27:51

also have this in a

27:54

way? trivial, or is it something

27:56

trivial? Or is it

27:58

going to actually have a

28:00

knock -on effect, right? right? I

28:02

I think this concept

28:04

of thinking through this, I feel

28:07

when what would I feel

28:09

when this I feel And

28:11

why did I feel through

28:13

this? of it, it's a cognitive work when

28:15

you're a cognitive work. When

28:18

work, this a cognitive work,

28:20

this common understanding, about knowing

28:22

about the other side of

28:24

things is quite important. Well,

28:26

Shrini Ragharman, thank you so much thank

28:28

you so much for joining

28:31

us today on the

28:33

ThoughtWorks Technology podcast. I'm I'm pleased

28:35

to have you here. here.

28:37

Thank Thank you, and thank

28:39

you for having me. me.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features