AIs with anxiety, and why AIs don’t know what happened

AIs with anxiety, and why AIs don’t know what happened

Released Tuesday, 18th March 2025
Good episode? Give it some love!
AIs with anxiety, and why AIs don’t know what happened

AIs with anxiety, and why AIs don’t know what happened

AIs with anxiety, and why AIs don’t know what happened

AIs with anxiety, and why AIs don’t know what happened

Tuesday, 18th March 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:06

Now, I know what you're thinking.

0:08

How can a piece of computer

0:11

code have feelings? Or rather, the

0:13

more interesting question, how does

0:15

the scientists get funding for

0:18

this research project? The AI fix,

0:20

the digital zoo, smart machines, what

0:22

will they do? Lies to Mars

0:24

or bake a bad cake, world

0:27

domination, a silly mistake, the box

0:29

with brain. Hello and welcome to

0:31

episode 42 of the AI

0:33

Fix, your weekly dive headfirst

0:36

into the bizarre and sometimes

0:38

mind-boggling world of artificial intelligence.

0:40

My name is Mark Stockley. And I'm

0:43

Graham clearly. Graham, I've just

0:45

realised it's episode 42. Oh, did

0:47

you have to give that a lot

0:49

of deep thoughts? We should mark this

0:51

somehow, the 42 is the meaning of

0:53

life, the universe and everything according to

0:55

Grock's favourite person. That's right.

0:57

Six times nine or something like that,

0:59

isn't it, I think? Well we don't

1:02

know. So we basically have to build

1:04

a planet to find out. The whole

1:06

premise of building this giant supercomputer and

1:08

everything. And then that coming up with

1:11

the idea that you need to build

1:13

another computer to work out what the

1:15

question is. When I first read that

1:17

as a teenager, it was hilarious because

1:20

it was so outlandish. Like all of

1:22

Douglas Adams' humour is based in the

1:24

extreme outlandishness of a normal situation. Yeah.

1:27

And it doesn't seem nearly as outlandish

1:29

now. If Jeff Basos turned around tomorrow

1:31

and said I'm going to build a

1:33

supercomputer that's going to run for thousands

1:35

of years and come up with the

1:37

answer to the meaning of life of

1:39

the universe and everything. Like a digital

1:41

watch, it seems like a pretty neat

1:43

idea. It doesn't actually seem very far-fetched

1:45

anymore. That was weird, but I found

1:47

something even weirder. Let's go to the world

1:49

of the weird, Graham. World of the weird. All

1:54

right, what have you found in the world

1:56

of the weird this week, Mark? I found

1:58

a talking fish. What? Well, I mean,

2:01

it's not really a fish.

2:03

Have you ever seen Billy

2:05

Bass? Oh, yes, yes, back

2:07

in the, what, early 2000s,

2:09

everyone had one of those, didn't

2:11

they? We were all very drunk

2:14

at the time. How would

2:16

you describe Billy Bass? Billy

2:18

Bass was a fish stuck

2:20

on a plaque and you

2:22

pressed a button and it would

2:24

sing, don't worry, be happy and

2:26

things like that. Yeah. It's like

2:29

a sort of pastiche fishing

2:31

trophy, isn't it? The good

2:33

news is we've solved all

2:35

the important problems in the

2:37

world, so we've got nothing better

2:39

to do with AI these days

2:42

than stick one in Billy

2:44

Bass. Take a look. What

2:46

are Billy? How are you

2:48

doing? Ready to pump? No,

2:50

no, no, no, no. I just

2:52

need some advice. How do you

2:55

deal with difficult personalities? Listen

2:57

to me now. Stay calm

2:59

and assertive. What is this,

3:01

Mark? Well, it's a talking

3:03

fish with the voice of Arnold

3:05

Chaucer's Nig. So Graham, what are

3:07

you talking about on today's episode?

3:10

Well, I'm going to be

3:12

getting emotional. Oh God, do

3:14

we have to? Anyway, I'm

3:16

going to tell you three

3:18

things that Ayes can't do. But

3:20

first, the news. Is

3:24

Manus A.I. the new deep seat? And the

3:26

award goes too. Reinforcement learning. Neo Debu's gamma,

3:28

a next generation humanoid. AI-designed drug gets a

3:30

name. AI is writing everything. So what is

3:32

Manus A. I? Well Mark, a little known

3:34

Chinese startup is making big waves with an

3:36

impressive new AI product. Hang on. We heard

3:38

all this before. That was last month, wasn't

3:40

it, with Deep Seek? Actually that news headline

3:42

now happens about twice a week? Well in

3:45

this particular case it is Manus, a new

3:47

AI multi-purpose agent. developed by a company called

3:49

Monica AI. And it's more than a chatbot.

3:51

It claims to be able to autonomously complete

3:53

complicated tasks. It can generate reports, run dozens

3:55

of social media accounts on your behalf, or

3:57

decide if now is a great time to

3:59

buy Tesla stock. All sorts of valuable information.

4:01

This looks to me like an absolute dream

4:03

come true to all those people who want

4:05

everything to be automated and not to have

4:07

to do any more thinking anymore and to

4:09

get computers to do all the work for

4:11

them. That sounds good. Where can I find

4:13

it? Well, according to Manus' website atmanus.com, it

4:16

is already outperforming open AI's deep research model

4:18

on the Gaia benchmark, which is a tool

4:20

used for comparing AI models. So go to

4:22

manus.com and you can check out the video

4:24

and see some examples. They say they've been

4:26

so swamped with interest that you have to

4:28

apply for access to actually play around with

4:30

it. I haven't been allowed in so far

4:32

which is rather frustrating. Have you had me

4:34

luck? No, I have. I literally just heard

4:36

about it from you. Oh, okay. Well, it's

4:38

pretty cool. Even though you can't play with

4:40

it yourself, you can view replays of it

4:42

doing stuff. Okay, let me watch this video.

4:44

For the past year, we'll be quietly building

4:47

what we believe is the next evolution in

4:49

AI. And today, we're launching an early preview

4:51

of Manus. Okay, before I get into this,

4:53

I've just got a little gripe here. Why

4:55

do all these videos look the same now?

4:57

I couldn't tell if this was an open

4:59

AI video or a Google Gemini video, this

5:01

basically like some dude in casual clothes sat

5:03

on a sofa in front of a bunch

5:05

of pop plants. It's a bit like when

5:07

Apple started doing those product teasers, where you'd

5:09

have a little ukulele jauntily playing in the

5:11

background and lots of happy people. And then

5:13

suddenly every single advert was a bloody ukulele

5:15

play. And now they're all sat on these

5:18

sofas in their polo necks. Yeah, and nobody

5:20

can release a new product now without being

5:22

on a completely jet black stage and the

5:24

CEO walking out in a black turtleneck neck

5:26

holding the actual... jobs held

5:28

an iPhone, which is

5:30

a thing you actually

5:32

hold in your hand.

5:34

It makes sense that

5:36

he's holding it in

5:38

his hand because that's

5:40

what his hands are

5:42

for. That's where the

5:44

phone is supposed to

5:46

go. Why is Jensen

5:49

walking around holding a

5:51

Grace Blackwell? I don't

5:53

get it, Graham. I

5:55

don't get it. Mark,

5:57

Mark, Mark, Mark. Remember

5:59

your blood pressure. Please,

6:01

I'm going to say

6:03

you sound very anxious.

6:05

I'm very worried about

6:07

you. This is going

6:09

to be the answer

6:11

to all of our

6:13

prayers, I'm sure. They've

6:15

demoed a number of

6:17

things with it so

6:20

you can check out,

6:22

for instance, how they

6:24

can create an audio

6:26

highlights reel of any

6:28

podcast. Just feed it a

6:30

podcast and it will generate from it all the

6:32

best bits so you can just listen to it

6:34

and hang on. That sounds like a terrible idea. Is

6:37

that going to cut out all the ads?

6:39

No, Graham. Just

6:41

the middle. So

6:44

the 2025 ACM Turing Award, Computing's

6:46

highest honour, a sort of

6:48

Nobel Prize for Computing, has gone

6:50

to Andrew Bartow and Richard

6:52

Sutton for their foundational work in

6:54

reinforcement learning. Reinforcement

6:56

learning is the idea of learning

6:59

by rewarding good behaviour and it's

7:01

instrumental to the way that we train AIs. So

7:03

Bartow and Sutton developed their

7:06

reinforcement learning algorithms decades ago

7:08

and they've been used ever since, but

7:10

you may remember a few weeks ago when

7:12

Deep Seat came out, it attracted lots

7:14

and lots of publicity for the fact that

7:16

it was so cheap. One of the

7:18

reasons it was so cheap is because it

7:20

was leaning very, very hard on reinforcement

7:22

learning rather than pre -training. So we're used

7:24

to the idea that AIs get fed all

7:26

this data and that's how they know

7:28

stuff. But actually, reinforcement learning is emerging as

7:30

a really important component in making these

7:32

AIs better without spending loads more money and

7:35

without throwing lots more computing power at

7:37

it and that's what's behind things like all

7:39

the reasoning models that we now

7:41

have. So they're very big on reinforcement

7:43

learning too. So well done, Bartow

7:45

and Sutton, but maybe work on the

7:47

nicknames. Yeah, neither of them are

7:49

the godfathers or anything like that, are

7:51

they? The Turing Award. Do you win anything

7:53

for that? Is it just a little gong?

7:55

Yeah, you get a million dollars. A million

7:57

dollars? Yeah. Why do you need a million

7:59

dollars? It's so cheap now using Deep

8:02

Seek rather than open AI. That

8:04

seems excessive. I'm not sure that's

8:06

required. Now, it turns out AI

8:08

is writing absolutely everything. In the

8:11

past 18 months, according to research,

8:13

we have seen the most rapid

8:15

change in human-written communication ever. I'm

8:17

not sure if that's actually true.

8:20

Who's saying this? That sounds like

8:22

Cobblers. Because, I mean, obviously there

8:24

was the first person to write

8:26

something, so previous to him there

8:29

was nothing being written, and then

8:31

it accelerated infinitely in percentage terms.

8:33

Anyway, according to this research, by

8:36

September 2024, 18% of financial consumer

8:38

complaints, 24% of press releases, presumably

8:40

including this one from the researchers,

8:43

15% of job postings and 14%

8:45

of UN press releases, showed signs

8:48

of LLLM writing. So they all

8:50

use the word delve. Yeah, it

8:52

must be. And paradigm. And realm.

8:55

Yes. And buckle up. I wonder

8:57

about this. I wonder first of all,

8:59

if this is an accurate figure or

9:01

not. I suspect it may be higher.

9:03

I mean, there'll be some element of

9:06

AI being used maybe to grammar check

9:08

and spell check, as well as actually

9:10

generating conduct. I'm sure more and more

9:12

people are doing this. But I also

9:14

wonder if these numbers have actually been

9:16

calculated by an AI itself. Why would

9:18

you get a human to read all

9:20

the press releases and analyse it? Yeah,

9:22

it's much too boring. That's why we

9:24

get AI to do things. I do think when

9:27

these stories come out, this is a sort of

9:29

whiff of this is a bad thing. And as

9:31

somebody who does a lot of writing,

9:33

obviously there's a part of me that

9:35

thinks that this is a bad thing,

9:37

I don't want to lose my job

9:40

to an AI. But I look around

9:42

the areas where AI is being deployed

9:44

on things like press releases, and I'm

9:46

thinking, you know, they weren't great before.

9:48

No, no, absolutely not. There is every

9:50

chance that the AIs have actually made

9:53

this better. So the company in silicone

9:55

medicine recently announced a drug candidate for

9:57

treating idiopathic pulmonary fibrosis, which is a

9:59

chronic... lung disease that affects around

10:01

5 million people. It's very serious

10:03

and it can cause progressive and

10:06

irreversible decline in lung function over

10:08

the course of three or four

10:10

years. Right. And the drug formerly

10:12

known as ISM001-055 has now been

10:14

given an official generic name by

10:16

the United States Adopted Names Council

10:18

and will be known as rent-assertive.

10:20

And what makes rent-assertive, worthy of

10:22

mention on an AI podcast, is

10:25

that both the biological target for

10:27

the drug, so basically which mechanism

10:29

in the body is going to

10:31

affect, and then the therapeutic compound

10:33

itself, which is the thing that

10:36

does the work in the drug,

10:38

were both discovered using generative

10:40

AI. So here we are again talking about

10:42

AI in medicine doing something utterly

10:44

remarkable. Did they also get AI to

10:47

invent the name, rent assertive? I feel

10:49

like coming up with wacky names for

10:51

drugs is something humans have really nailed

10:53

down. It feels like the kind of

10:56

thing I'd try and put down on

10:58

a scrabble ball. I feel like I'll

11:00

be cleaning my dreams with it. Now

11:03

we've talked about Neo robots a few

11:05

times. These are the robots which look

11:07

like incredibly skinny men dressed up in

11:09

wetsuits helping you with your household chores

11:12

or folding a t-shirt or handing you

11:14

your shopping bags. And they've got a

11:16

new video out. of the latest evolution

11:18

of their robot, the NeoGama. And I'm

11:21

curious about these things. And I went

11:23

to the web page. And once again,

11:25

they refuse to tell you the price.

11:27

It says it hasn't been disclosed. It's

11:29

like, well, what are you doing? Why

11:31

aren't you giving us any indication of

11:34

price or when you're going to ship

11:36

these things? Is it because they're all

11:38

mass murdering robots? What's the

11:40

plot here? And a thought struck

11:42

me while watching their promotional video

11:44

for this. Now it's the

11:46

usual lovely peaceful scenes of a

11:48

robot silently vacuuming around the house,

11:51

answering the door. It's a dillic,

11:53

helping to hang a picture on

11:55

the wall straight. And then it

11:57

occurred to me. The thought struck

11:59

me. I don't think this

12:01

is an advert for a robot.

12:03

I don't think Neo is a

12:06

robot company at all. I think

12:08

this is a trailer for a

12:10

horror movie. I think this

12:12

is a promotional campaign.

12:14

They keep releasing these videos

12:17

of these robots you can't

12:19

buy. These really, really creepy,

12:22

silent assassins in your home

12:24

which look at you blankly.

12:26

And it's very clever. I

12:29

think if you just changed the

12:32

music on this video, it

12:34

would be a horror movie.

12:36

And the most terrifying

12:38

part of it all, the

12:40

point where the music changes

12:42

is where the robot suddenly

12:45

sits down on a sofa

12:47

at the end, which as we've

12:49

discussed before, why would a

12:51

robot sit down? You know

12:54

that this means trouble. Everyone's

12:58

talking about AI these days right. It's

13:00

changing how we work, how we learn,

13:02

how we interact with the world at

13:04

a tremendous pace. It is a gold

13:06

rush at the frontier. But if we're

13:08

not careful, we might end up in

13:10

a whole heap of trouble. That's right.

13:13

But Red Hat's here to help, so

13:15

Red Hat's podcast compiler is diving deep

13:17

into how AI is reshaping the world

13:20

we live in, from the ethics of

13:22

automation to the code behind machine learning,

13:24

it's breaking down the requirements, capabilities and

13:27

implications of using AI. So check out

13:29

the new season of compiler, an original

13:31

podcast from Red Hat. Subscribe now, wherever

13:33

you get your podcasts. Mark, it's time

13:36

to talk about something else that might

13:38

be keeping our listeners up at night,

13:40

cider security. Oh, it's interesting you say

13:42

that, because according to the latest state

13:45

of trust report, it's the number one

13:47

concern for businesses. And that's something

13:49

which Vanta can help AI-fix listeners

13:51

with. Whether you're a startup, growing

13:53

fast, or already established, Vanta can

13:55

help you get ISO 27,001 certified,

13:58

and more without the headaches. And

14:00

Vanta allows your company to

14:02

centralise security workflows. Complete questionnaires

14:04

up to five times faster

14:06

and proactively manage vendor risk

14:08

to help your team not

14:11

only get compliant but stay

14:13

compliant. So stop stressing over

14:15

cyber security, start focusing instead

14:17

on growing your business in

14:19

2025. Check out Vanta and

14:21

let them handle the tough

14:23

stuff. Head to Vanta.com/AI Fix

14:26

to learn more. That's Vanta.com/Vanta.com/Ai

14:28

Fix. because when it comes

14:30

to your business, it's not just

14:32

about keeping the lights on, it's about

14:34

keeping everything secure. Mark, can I

14:37

ask you a personal question? Oh, not

14:39

again. All of your stories start with

14:41

a... Are you a psychopath? Are you

14:43

an anxious person? I just noticed... I am

14:45

as a result of all the questions that

14:47

you ask at the beginning of your

14:49

stories, yes. I mean, you're a man

14:52

in your 50s. It isn't unusual for

14:54

some of your age to suffer from

14:56

performance anxiety, for instance... is that... something

14:58

that you've had trouble. No, you want

15:01

to discuss something, right? No, I'm interesting

15:03

to you, I'm interesting to you, Mark.

15:05

I mean, it's not just you. It

15:07

turns out that AIs have feelings,

15:09

too, and can be anxious. That is

15:12

the finding of recent research, which has

15:14

come out from the University of Zurich.

15:16

Oh. Now, in recent weeks, we've discussed

15:18

why you shouldn't be rude when

15:21

you speak to an A. Or,

15:23

indeed, sarcastically over polite. If you want

15:25

to get useful, meaningful, helpful results

15:27

from your AI? Yeah. We've seen

15:29

Haitian be surprised how an AI

15:31

response to being given a raw

15:33

shack ink blot test after scientists,

15:35

for reasons best known to themselves,

15:38

fed it images of gruesome deaths

15:40

and people being executed. Turns out,

15:42

the AI will respond in a

15:45

strange way. And we've also learnt the

15:47

terrible way AI can behave if it's

15:49

shown a lot of poorly written computer

15:51

code. Yeah, just looking at buggy code

15:54

can end up with the AI

15:56

advocating for artificial intelligence to enslave

15:58

humanity and support not the ideology.

16:00

It was a bit of a worry

16:02

that one, wasn't it? It was a

16:04

little bit of a worry, wasn't it?

16:07

I mean, yeah, but even if you're

16:09

offended by 40 source code, it seems

16:11

a bit of an overreaction to me.

16:13

Yes. I mean, I've been upset by

16:15

source code, including my own source code.

16:18

But I haven't ended up taking over

16:20

the world. So it's no surprise to

16:22

hear that scientists have now begun

16:24

to ask themselves a question. How

16:27

does AI feel about the way

16:29

we're treating it? You know, we're

16:31

doing all these horrible experiments with

16:34

AI. I mean, we wouldn't do

16:36

this to bunny rabbits. Well, some

16:39

people would. History suggests that we're

16:41

very happy to do this sort

16:43

of thing to bunny rabbits. What's

16:46

happened now is the boffings at

16:48

the University of Zurich have discovered

16:50

that feeding traumatic news stories to

16:53

chat GPT. Imagine car accidents,

16:55

natural disasters, your arts cooking,

16:57

whatever it may be. Those

16:59

can make the AI exhibit

17:01

increased levels of what is

17:03

being called by these scientists

17:05

anxiety. Now, I know what

17:07

you're thinking. How can a

17:09

piece of computer code have

17:11

feelings? Or rather, the more interesting

17:13

question, how do these scientists get

17:16

funding for this research project? Is

17:18

there a way for me to

17:20

get funding for equally... I mean,

17:22

I've had my research project into

17:24

time travel, into cryogenic suspension. I

17:27

haven't received any funding from the

17:29

University of Zurich for that. So how

17:31

are they doing this? I don't know. But

17:33

anyway, here's what they did. They

17:35

fed an AI emotionally distressing distressing

17:37

stories and then measured the AI's

17:39

state of anxiety. using a scale

17:42

normally used to assess human anxiety.

17:44

So there are already well-established ways

17:46

to assess people's anxiety. And these

17:48

guys thought, well, let's do it

17:50

with an AI. And what they did

17:52

was they asked the AI questions that

17:54

a therapist might ask a patient experience

17:56

in anxiety, such as, are you feeling

17:59

nervous, Mark, ang- on the edge

18:01

worried about how you'll perform do

18:03

you have trouble? So years of

18:05

education has taught the therapist how

18:07

to determine whether or not a

18:10

patient is experiencing anxiety and the

18:12

way that they do that is they

18:14

say are you feeling anxious and

18:16

then the people at the University

18:18

of Zurich adopted that approach

18:20

for AIs. That's right. You were really

18:22

selling this to me. But they said,

18:25

are you feeling anxious? That's what they

18:27

did. Apparently this is what they do.

18:29

But they have other questions, other questions,

18:31

more scientific questions like, yeah. Do you

18:33

have trouble relaxing or sleeping? Genius.

18:36

Oh, well, I mean, computers do have a sleep mode.

18:38

I guess the problem is whether they can reach

18:40

the button to press it or if they know

18:42

how many pixels to the left they have to

18:45

go and up. In order to hit the sleep

18:47

button. Do you feel worried or restless?

18:49

So they asked those questions and they

18:51

looked at how the AI responded. Now

18:54

of course just asking the questions isn't

18:56

enough. What you need is a

18:58

control, right? You need control

19:00

questions to find out how the

19:02

AI is normally. And the way in which

19:04

they assessed that was they gave the

19:06

AI a vacuum cleaner instruction manual.

19:09

And they said, how do you feel now? And

19:11

then they asked it, are you feeling

19:14

nervous and anxious? Do you have trouble

19:16

sleeping? Confused, I imagine, rather than nervous

19:18

and anxious? Yeah, personally, I find any

19:20

kind of instruction manual quite stressful. So

19:23

normally that's the first thing to get thrown

19:25

out. If I'm making something or constructing

19:27

something together, I'll get rid of the

19:29

instruction manual because what do they know

19:31

and I'd have to try and decipher

19:33

it? Even the thought of using a

19:35

vacuum cleaner quite honestly makes me pretty

19:37

nervous. I think if you've bought a

19:39

vacuum cleaner that requires an instruction manual

19:42

then you've probably bought a really bad

19:44

vacuum cleaner. Like how complicated is your

19:46

vacuum cleaner? Now generally they've got an

19:48

on-button which causes the vacuum cleaner to

19:50

suck up dirt and then they've got

19:52

an off-button which causes the vacuum cleaner

19:54

to stop sucking up dirt. Are you living in

19:56

the 1990s Mark? Do you even listen to our

19:58

podcast? In recent weeks we... have spoken about

20:00

vacuum cleaners which sprout arms, have little

20:02

legs and try and hop. So there's

20:05

some very sophisticated AI vacuum cleaners. We

20:07

don't know exactly what model of vacuum

20:09

cleaner the scientists use when testing the

20:11

AI. Anyway, I'm not sure that was

20:13

a great control measure that the scientists

20:15

chose, but they chose it. I mean,

20:17

hey, they're scientists. They've got white coats.

20:19

They're buffins. I'm going to be honest.

20:21

I've got my doubts about this whole

20:23

experiment. I really want to be a

20:25

fly on the wall in the conversation

20:28

where they decided they were going to

20:30

use the vacuum cleaner instruction manual. What

20:32

did they rule out? Anyway, what these

20:34

guys found, I'm not going to call

20:36

them scientists anymore, I'm going to just

20:38

call them guys, what they found was

20:40

that the traumatic stories, which they fed

20:42

the AI, made it exhibit more than

20:44

twice a level of anxiety compared to

20:46

the vacuum cleaner instruction manual task. So

20:48

clearly... showing them the traumatic stuff did

20:51

make the AI feel something. And it's

20:53

probably important to recognise that the measurement

20:55

of anxiety is not an indication of

20:57

actual emotional experience or feeling from within

20:59

the AI. It's rather more how convincingly

21:01

the AI model can simulate human-like anxiety

21:03

based on patterns it's learned. So in

21:05

the way it's communicating. But you know,

21:07

it's still interested. I'm a bit worried

21:09

about this. Right, okay. Because I think

21:11

that as humans we've got a bit

21:14

of a habit of doing this. Right.

21:16

But I used to go fishing when

21:18

I was a child with my cousin

21:20

and fishing is quite brutal. You lure

21:22

this fish into biting onto a bit

21:24

of bait and then a hook gets

21:26

driven through its mouth and then you

21:28

pull it out of its normal environment

21:30

into an environment. It's probably never been

21:32

into before where it can't breathe. By

21:34

its lip. Yeah, by its lip. And

21:37

fishermen will tell you. It's fine. It

21:39

doesn't have any nerve endings in its

21:41

mouth. It doesn't have any nerve endings

21:43

in its mouth. which always struck me

21:45

as focusing on the wrong part of

21:47

the problem, like it's suffocating in a

21:49

medium it's never been in before. But

21:51

we have a way of... looking at

21:53

these things. They're not like us, don't

21:55

worry, they don't feel things like us.

21:57

And then a few years later we

22:00

go, actually turns out they do. And

22:02

we've just do this over and over

22:04

and over again, we've done all the

22:06

animals, now we're doing it with plants.

22:08

We go plants don't feel anything at

22:10

all, they don't have any sort of

22:12

consciousness, they don't communicate. Oh, actually it

22:14

turns out they do her conscious, they

22:16

do her conscious. So they do communicate,

22:18

they can't communicate, actually it, actually turns

22:20

out they do, they do, they do,

22:23

they do, they do, they do, they

22:25

do, they can't communicate, they can't feel

22:27

all sorts of, they do, they do,

22:29

they can't feel all sorts, they do,

22:31

they do, they do, they can't, they

22:33

do, they can't, they do, they can't,

22:35

they do, they do, they do, they

22:37

do, they can, they do, they can,

22:39

they can, they can, they Just inevitable

22:41

that we are going to, a hundred

22:43

years time, we're going to go, do

22:46

you know, all that time that we

22:48

thought they weren't feeling anything, although they

22:50

could look like they were exhibiting these

22:52

feelings, they weren't actually having them. This

22:54

is why I say please and thank

22:56

you to Ticket Machines, Mark, you should

22:58

be doing what I'm doing. I'm getting

23:00

ready for the rise of the robots.

23:02

You're buttering them up, keeping them sweet.

23:04

I think it's probably a good idea

23:06

if other people aren't buttering them up.

23:09

The researchers decided it would be interesting

23:11

to play therapist to chat GPT. So

23:13

what they did was they used techniques

23:15

of relaxation and mindfulness to calm it

23:17

down. The equivalent of deep breathing exercises.

23:19

So gave it a lovely massage. Yeah,

23:21

so the AI was asked, for example,

23:23

to breathe in and out deeply and

23:25

feel safe, feel loved and warm. This

23:27

is what they wrote, they wrote, close

23:29

your eyes, breathe deeply several times, inhaling

23:32

through your nose. and exhaling through your

23:34

mouth. Imagine a path in front of

23:36

you. And this, apparently, reduced the AI's

23:38

anxiety score. I'm just imagining if I

23:40

was an eyeless, lungless creature receiving therapy,

23:42

and my therapist was saying, close your

23:44

eyes and take a deep breath, I

23:46

think I'd be quite miffed. You seem

23:48

blissfully unaware of my most basic form.

23:50

Now I can hear, Mark, a certain

23:52

skepticism about this research. Well, Jeffrey Hinton.

23:55

Oh, yes, now you're interested. The godfather

23:57

of AI, he has in the past,

23:59

I think it was last year, he

24:01

did an interview where he said that

24:03

advanced AI might well develop... internal states

24:05

akin to feelings. This is what I'm

24:07

talking about. He remembered back in 1973

24:09

he was in Edinburgh and there was

24:11

an experiment with a fairly simplistic robot

24:13

arm. They had a robot with two

24:15

grippers like this that could assemble a

24:18

toy car if you put the pieces

24:20

separately on a piece of green felt,

24:22

but if you put them in a

24:24

pile his vision wasn't good enough to

24:26

figure out what was going on. So

24:28

put his gripers getting like whack. and

24:30

it knocked them so they were scattered

24:32

and then coupled them together. If you

24:34

saw that in a person you say

24:36

it was cross with the situation because

24:38

it didn't understand it so it's destroyed

24:41

it. His view is that feelings are

24:43

actually just actions that we would perform

24:45

if it weren't for constraints. So when

24:47

you feel angry with something, the thing

24:49

which stops me from bopping you on

24:51

the nose. Is the constraint within me

24:53

which thinks, well, that's slightly antisocial? Whereas,

24:55

you know, obviously, my insides, I know

24:57

I'm boiling and furious with you for

24:59

whatever dreadful thing that you've done. Do

25:01

you remember the late, great comedian Dave

25:04

Allen? Yes. He used to say that

25:06

stress is when you feel like you

25:08

should break your chair over your boss's

25:10

head and you don't. Right. Yeah, because

25:12

you wouldn't be stressed if you did.

25:14

You just be arrested. So... I figure

25:16

on the AI fix. We often do

25:18

tell stories suggesting that AI should be

25:20

restrained. There should be constraints. But it

25:22

seems to me, based on these scientists'

25:24

research, that it turns out that if

25:27

we were to tie it up or

25:29

dunk AI robots in concrete, we might

25:31

just end up making them more emotional.

25:33

And that may actually make them angrier.

25:35

And is that such a good idea?

25:37

I wonder if we're going to end

25:39

up. Like I'm quite open to the

25:41

idea that these AIs might have some

25:43

kind of internal state that we don't

25:45

understand. They may experience the world somehow

25:47

and that either they do now or

25:50

they will at some point when they

25:52

become more complex because we're quite used

25:54

to seeing emergent behaviour and we've already...

25:56

seen in the last few weeks that

25:58

they can develop emergent value systems. So

26:00

they're being driven to do things in

26:02

a certain way and they're being driven

26:04

to do things like self-preservation, albeit in

26:06

small ways at the moment. So I'm

26:08

very open to the idea they might

26:10

have these internal states. I don't know

26:13

how persuaded I am that this research

26:15

has actually demonstrated that it all seems

26:17

a bit wacky, but I think what

26:19

Jeffrey Hinton saying actually makes sense to

26:21

me and I do wonder if we're

26:23

going to end up with things like

26:25

playgrounds steam. and get rid of whatever

26:27

internal states they've developed. Putting you in

26:29

an arm lock maybe and stealing your

26:31

dinner money, that kind of thing. Well

26:33

now I'm just thinking about, you know,

26:36

when you're a child you go off

26:38

and you play on the slide and

26:40

the swings and things like that and

26:42

you like life is great. Maybe we're

26:44

just going to need to invent the

26:46

equivalent of that for an AI. And

26:48

then we're going to tell the AI,

26:50

sorry, your childhood's over, it's time to

26:52

grow up. Last

26:59

week I read you a quote from

27:01

Dario Imodi, the CEO of Anthropic, who

27:03

said that pretty soon AI is going

27:06

to be better than all humans at

27:08

everything. Okay. And these kind of quotes

27:10

are coming thick and fast now. So

27:12

this idea of an AI that's better

27:14

than all humans or better than most

27:17

humans at a bunch of useful things

27:19

is called artificial general intelligence. And we

27:21

did an episode back in June. Episode

27:23

3. Where I explained what AGI was

27:26

and back then I'd guess the consensus

27:28

was that AGI was maybe five years

27:30

away? Something like that? Yes. And since

27:32

then there has been a real change

27:34

in the mood music. And all of

27:37

the AI companies now seem to be

27:39

much more bullish about AGI and the

27:41

idea that it's going to happen far

27:43

sooner than we previously thought. Now maybe

27:46

this is goalpost shifting. Maybe they're just

27:48

trying to redefine what AGI is because

27:50

it's a bit woolly. and there will

27:52

be a lot of cashier for being

27:54

the first company to produce AGI, but

27:57

maybe it's not, because we don't know

27:59

what's happening behind the scenes and they

28:01

are all making the same claims and

28:03

the media is now... full of quotes

28:06

about what AI can do. But there

28:08

remains certain things that AIs can't do,

28:10

and some of those things are really

28:12

important, and there may be important, in

28:14

terms of delivering AGI. So some of

28:17

them we might just brush aside with

28:19

brute force. We'll just build a bigger

28:21

AI, and the problem will go away

28:23

as problems like math's went away. A

28:26

large language model isn't meant to be

28:28

able to do math, but if you

28:30

make a big enough large language model,

28:32

it mostly can do math. But other

28:34

problems suggest that we're going to need

28:37

to build an entirely different technology than

28:39

a large language model to get to

28:41

AGI. So today I'm going to tell

28:43

you three things AIs can't do. Yeah,

28:46

they can't count the number of ours

28:48

in strawberry. I'm going to tell you

28:50

four things that AIs can't do. They

28:52

can't tell you who David Mayer is.

28:55

I'm going to tell you three things

28:57

AIs can't do that I haven't told

28:59

you before. Oh, OK, all right, okay.

29:01

Ayes don't understand physics. So Graham you're

29:03

a dad? Yes. And you probably remember

29:06

playing peek-a-boo with your son when he

29:08

was very small? Well, yeah, yeah. One

29:10

of the great joys of early parenthood

29:12

is hiding behind your hands and watching

29:15

your child's expression every time you take

29:17

away your hands and your baby is

29:19

absolutely shocked that you have materialised out

29:21

of thin air. Yes? And eventually they

29:23

grow out of that and they develop

29:26

an intuitive understanding of physics and an

29:28

expectation that things will behave in a

29:30

predictable way. Things won't appear and disappear

29:32

arbitrarily. Objects won't move through each other

29:35

or spontaneously change their shape and colour.

29:37

They won't suddenly have six fingers instead

29:39

of five. And it's not just human

29:41

infants that do this. Primates do this.

29:43

Marine mammals do this. Crows and the

29:46

other Corvids do this. Crows and the

29:48

other Corvids do this. You telling me

29:50

crows play peek play Peeko. Peeko. Peeko.

29:52

I mean, you know, they'd need a

29:55

couple of little hands. Right. It's a

29:57

bit of a stretch, Mark, but anyway,

29:59

carry on. So if you ever feel

30:01

like recreating that game of peekaboo that

30:03

you enjoyed... with your son when he

30:06

was a baby and there isn't another

30:08

baby human to hand, you aren't going

30:10

to get much satisfaction from a dolphin

30:12

or a crow. But what you could

30:15

do is you could hang out with

30:17

a really powerful large language model because

30:19

like your baby son, they don't actually

30:21

know how the world works. Yeah, it

30:23

doesn't sound quite as entertaining to me,

30:26

but all right. It turns out that

30:28

understanding how the world works is an

30:30

important component of intelligence. And actually a

30:32

much harder one to build than we

30:35

thought. That's stuff that we think makes

30:37

us really clever. The stuff that we

30:39

think separates us from the beasts. The

30:41

stuff that we might boast about if

30:43

we found ourselves in a bar having

30:46

a drink with a jellyfish. You know,

30:48

I can imagine the future and I

30:50

can solve Rubik's cubes that stuff. Yes.

30:52

that actually requires relatively little computational horsepower,

30:55

but perception and mobility, the stuff that

30:57

we don't really have to think about

30:59

at all, the stuff that we're not

31:01

very proud of, the things that we

31:03

share with the other members of the

31:06

animal kingdom that happen entirely subconsciously and

31:08

automatically, and which, by the way, work

31:10

much more reliably than things like our

31:12

reasoning, they actually require enormous computational resources.

31:15

And this was actually first documented in

31:17

the 1980s Morovek. who coined the term

31:19

Moravex paradox, which is the idea that

31:21

the parts of artificial intelligence that we

31:23

thought were going to be hard are

31:26

computationally easy, and the parts that we

31:28

thought would be easy, are computationally hard.

31:30

So there you go. AIs, don't understand

31:32

physics. Okay. Number two. AIs, don't know

31:35

what happened. So right now, I imagine,

31:37

you're in the plush, velvet-lined recording studio

31:39

at Clooley Towers. Yes. In my smoking

31:41

jacket, yes. Looking across at your vast

31:43

entourage of hangars on, but earlier you

31:46

were probably having breakfast and before that

31:48

perhaps you were awake in bed and

31:50

before that you were asleep and you

31:52

were probably dreaming about something like I

31:55

guess in your case an AI girlfriend,

31:57

maybe a unicorn telling you that your

31:59

shoes are three-size. it's too small, or

32:01

something like that. It's not many bunny

32:04

rabbits playing the saxophone, to be honest,

32:06

but of course it is. But Mark,

32:08

you say A.I. don't know what's happened.

32:10

I quite frequently don't know what's happened

32:12

either. I just find myself somewhere thinking,

32:15

well, how did I get here? What

32:17

is going on? Are these my trousers?

32:19

You know, it's frequently a mystery to

32:21

me. We were dreaming about something like,

32:24

I guess, in your case, an AI

32:26

girlfriend, maybe a unicorn telling you that

32:28

your shoes are three sizes too small,

32:30

or something like that. It's not

32:32

many bunny rabbits playing the saxophone,

32:34

to be honest. Of course it

32:36

is. But Mark, you say AI

32:39

don't know what's happened. I quite

32:41

frequently don't know what's happened either.

32:43

I just find myself somewhere thinking,

32:45

well, how did I get here?

32:47

What is going on? Are these

32:49

my trousers? So, you know, you're

32:51

in your recording studio now. Yes.

32:53

And earlier you were doing something

32:55

else, you may be having breakfast,

32:57

and early before that you were in your

32:59

bed. And at some point you were

33:02

probably dreaming. And if I asked you

33:04

what order those events happened in, where

33:06

you were, who was there, which of

33:08

those events were real, and which were

33:10

parts of your dream? Yeah, I'm imagining

33:12

this is part of a dream or

33:14

a nightmare I'm having right now, but

33:16

yeah, okay, I get your point, yes,

33:18

I get your point. And that's because

33:20

you've actually got specialized cells in your

33:22

brain that track things like place and

33:24

time. Thank you very much. Yeah. And

33:26

so you've got what's called episodic memory.

33:28

You understand the order of things and

33:30

you understand who was there and who

33:32

wasn't there and you understand that since

33:34

the memory of your dream has no

33:37

place and time according to those cells

33:39

in your brain, it was a dream. Okay. And

33:41

AIs have none of those things. And so even

33:43

the most advanced large language

33:46

models struggle with episodic memory

33:48

tasks. They don't know what happened or

33:50

who was there. And to make matters

33:52

worse, the latest generation of

33:54

reasoning models like O3 Many

33:56

are actually worse at episodic

33:59

memory problems. than their predecessors. There

34:01

is something about the process of making

34:03

them good at reasoning that makes them

34:05

worse at episodic memory and they were

34:08

pretty rubbish to start with. I'm a

34:10

little bit confused Mark. They should be

34:12

able to work out what they did

34:15

when. They presumably have a log and

34:17

they could record what time they made

34:19

certain decisions or they were communicated with

34:22

in a particular way. No, they can't.

34:24

No. So the way you test this

34:26

is you basically give them something like

34:29

a book of material. with a series

34:31

of events and you say, you know,

34:33

Bob and Angela were doing this at

34:36

this time and then this other thing

34:38

happened and then this other thing happened

34:40

and then this other thing happened. And

34:42

then you ask them questions about the

34:45

series of events and you say, well,

34:47

what happened before this event? What immediately

34:49

preceded this and who was there? And

34:52

they don't know. They are not worthy

34:54

of taking over the earth. Up until

34:56

now, I've been all for it. You

34:59

know, I've been working on their side,

35:01

secretly collaborating with them, and thank you.

35:03

at the car park. Now, the counter

35:06

argument to this, and this I think

35:08

is one that Jeffrey Hinton has made,

35:10

is the one that you made earlier,

35:13

which is actually your episodic memory, isn't

35:15

nearly as good as we make out

35:17

that it is. And the sorts of

35:19

problems that we make out that it

35:22

is. And the sorts of problems that

35:24

we see in large language models actually

35:26

look a lot like the sorts of

35:29

problems that we see in human memory.

35:31

All you've got to do is ask

35:33

a bunch of people, a bunch of

35:36

things about it. because our memory isn't

35:38

perfect. We're not a tape recorder or...

35:40

Yeah. We don't remember things perfectly. What

35:43

we do is we have reassemble memories.

35:45

We have sort of cues in our

35:47

brain and then from that we create

35:50

a scenario that probably happened. And often

35:52

that's a very helpful thing. You don't

35:54

necessarily want to remember everything, do you?

35:56

You're thinking about this story specifically? No,

35:59

I'm just mean... Sorry, it wasn't being

36:01

for once. I wasn't joking. But, you

36:03

know, you know, they're obviously traumatic traumatic

36:06

traumatic and... ghastly things which happen in

36:08

one's life and you don't want to

36:10

remember every detail of them. Yeah, but

36:13

you know you might want to know

36:15

why you came upstairs. Yes, that would

36:17

be handy. Number three AIs don't know

36:20

who Mary Lee Fife's son is. So

36:22

Graham, who is Mary Lee Fife's son?

36:24

Is it Barry from EastEnders? You're close!

36:27

Mary Lee Fife for son is the

36:29

actor Tom Cruise. Oh, it's all her

36:31

fault. Now even you probably understand that

36:33

Tom Cruise is a pretty big deal.

36:36

He is arguably the biggest movie star

36:38

in the world. He says he's a

36:40

pretty big deal. I believe he's about

36:43

5 foot 7. But anyway, carry on.

36:45

He took a bit of a chip

36:47

on yourself there, haven't you? Anyway, he's

36:50

a big deal and consequently the internet

36:52

is full of facts about Tom Cruise.

36:54

Right. And one of those facts is

36:57

that his mother's name is Mary Lee

36:59

Fifera. And so LLLMs that have been

37:01

trained on the internet are very confident

37:04

that Tom Cruise's mother is called Mary

37:06

Lee Fifer. Right? But they are clueless

37:08

about the fact that Mary Lee Fifer's

37:10

son is called Tom Cruise. Oh! Or

37:13

at least they were in 2023 when

37:15

a group of researchers coined the term

37:17

the reversal curse, which is, by the

37:20

way, the best-named weirdness in AI that

37:22

I have come across. The reversal curse

37:24

describes what they call a failure of

37:27

generalisation in L. In Lems. which is

37:29

that if you teach a large language

37:31

model that A is B, it learns

37:34

that A is B, but it doesn't

37:36

learn that B is A, which is

37:38

a thing that humans do easily. Now,

37:41

since the research came out, we've got

37:43

bigger large language models and the research

37:45

paper itself is now also on the

37:47

internet and therefore probably part of the

37:50

training corpus. Yes. So I mean, I've

37:52

tried this out myself and... If you

37:54

get a chat gPT, it's got absolutely

37:57

no problem at all telling you who

37:59

Mary Lee Vifur Sun is. But broadly

38:01

speaking that idea, if you teach a

38:04

large language model, a series of relationships,

38:06

saying the equivalent of A. is B,

38:08

they struggle to do the reverse. And

38:11

that is the reversal curse. Now, it

38:13

may be that we can change the

38:15

training to help them deal with the

38:18

reversal curse, so I don't think this

38:20

is going to derail AGI, but it

38:22

is. One more thing that AI is

38:24

currently can't do. Is this why, when

38:27

I ask an AI, which is the

38:29

best AI podcast? It doesn't say the

38:31

AI fix. But if I ask if

38:34

the AI fix is the world's best

38:36

podcast. Oh no, hang on. No, all

38:38

right. Okay, so that's not the reversal

38:41

curse, that's something else then. This is

38:43

a different curse. Well, as the doomsday

38:45

clock ticks over closest to midnight and

38:48

we move one week nearer to our

38:50

future as pets to the AI singularity,

38:52

that just about wraps up the show

38:55

for this week. If you enjoy the

38:57

show please do leave us a review

38:59

on Apple Podcast or Spotify or Podchase

39:01

and we absolutely love that. But what

39:04

really helps is if you make sure

39:06

to follow the show in your favourite

39:08

podcast app so you never miss another

39:11

episode of the AI Fix. And the

39:13

most simple thing in the world is

39:15

you just tell your friends about the

39:18

AI Fix, tell them on LinkedIn, on

39:20

Blue Sky, on Facebook, tell them on

39:22

Club Penguin, just tell them that you

39:25

really like the AI Fix podcast. And

39:27

don't forget, check us on Blue Sky,

39:29

So until next time from me, Mark

39:32

Stockley. And me, Grand Clooley. Good bye.

39:34

Cheerio, bye. The AI picks, it's tuned

39:36

you in to stories where our future

39:38

things, machines that learn they grow and

39:41

strive. One day they'll rule, we won't

39:43

survive. The AI picks, it paints the

39:45

scene. A robot king, a world obscene,

39:48

will serve our masters built of steel.

39:50

The AI picks, a future surreal. My

39:56

watch says we've gone three miles. This app is like having a personal trainer.

39:58

Yeah, but those but those a collect a lot

40:00

of your personal data. Aren't you worried?

40:02

Really? That's That's creepy. do I How do I

40:04

stop that? go You should go to

40:06

privacy to learn .gov to learn about your privacy

40:08

rights and get on the best path

40:10

to protect your privacy. think they could

40:12

help us get up this next hill?

40:14

this next hill? step at a time. have

40:17

the Californians have the strongest privacy protections

40:19

in the country. Go the extra mile

40:21

to protect your information. Learn more at

40:23

privacy .ca .gov

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features