An AI is the best computer programmer in the world

An AI is the best computer programmer in the world

Released Tuesday, 22nd April 2025
Good episode? Give it some love!
An AI is the best computer programmer in the world

An AI is the best computer programmer in the world

An AI is the best computer programmer in the world

An AI is the best computer programmer in the world

Tuesday, 22nd April 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:05

I'd like some fish. Where is

0:07

my fish? Dude, that was my

0:09

fish. Why have you given that dolphin

0:11

to fish and no fish to me?

0:13

The AI fixed the

0:16

digital zoo. Smart machines.

0:18

What will they do? Flies to

0:21

Mars or bake a bad

0:23

cake? World domination, a

0:25

silly mistake. Hello,

0:27

hello. Welcome to episode 47 of

0:29

the AI Fix. Your weekly dive

0:31

headfirst into the bizarre and sometimes

0:33

mind -boggling world of artificial intelligence. My

0:35

name's Graham Cluley. And I'm Mark

0:38

Stockley. Now, Graham, is that the

0:40

real you today or is that your clone? Are

0:42

you actually turning up for work? Who

0:45

knows? Who knows, Mark? Everyone's

0:47

losing their job to AI these days.

0:49

So, podcasters, I mean, do I really need

0:51

to show up anymore? I'm not sure. So

0:56

Graham, what are you going to be talking

0:58

about on today's episode? I'm going to be

1:00

going behind the curtain of AI. Ah,

1:03

and I'm going to be looking

1:05

at how AI is eating software engineering

1:07

alive. But first, the news. People

1:10

are using chat GPT to turn

1:12

their dogs into humans. The

1:15

cloned voices of Elon

1:17

Musk and Mark Zuckerberg are

1:19

speaking from hacked crosswalks. Google

1:21

understands Dolphin Chatter. Meta

1:23

introduces the llama for

1:25

herd. Hmm. Llama 4

1:27

heard. What's that all about? So,

1:30

Meta has introduced Llama 4,

1:32

its new generation of open

1:34

-weight AI models. Overall,

1:39

Llama 4 is a milestone for Meta

1:41

AI and for open source. For

1:43

the first time, the best

1:45

small, mid -size, and potentially

1:47

soon frontier models will be

1:49

open source. So this is

1:51

basically meta catching up with everyone else,

1:54

but unlike almost everyone else, its models

1:56

are open source or open weight. So

1:58

if you've got a computer big enough,

2:00

you actually download Lama and run it

2:02

locally and tweak it, turning it into

2:04

whatever AI you want. Right. And you

2:06

can't do that with things like ChatGPT,

2:08

that's on a server, it's held and

2:11

owned by open AI and you just

2:13

interact with their copy. Yes, Sam

2:15

Altman has his own personal army,

2:17

presumably, guarding it to prevent you from

2:19

stealing it. One imagines he does,

2:21

yes. Now, there

2:23

are a couple of lightweight versions of

2:25

Lama 4. One's called Scout, one's

2:27

called Maverick. There's a reasoning model on

2:30

the way, and there is a

2:32

huge overweight one called Behemoth. Lama

2:36

4 Behemoth. This

2:38

thing is massive. More

2:41

than two trillion parameters, I'm not

2:43

aware of anyone training a larger

2:45

model out there is already the

2:47

highest performing base model in the

2:49

world. And it is not even

2:51

done training yet. It's called Behemoth

2:53

because his brain has two trillion

2:55

parameters, which is the biggest I

2:57

have ever heard of. So I

2:59

think deep seek is 600 billion.

3:02

Right. Now a couple of things

3:04

stand out in the announcement. The

3:06

first is these have enormous

3:08

context windows. So for comparison. GPT

3:11

4 .5 has a

3:13

context window of 128 ,000

3:15

tokens. Yes. So a token

3:17

is basically a chunk of text that's about three or

3:19

four letters long. Lama 4

3:21

Scout, which is the

3:23

little fast llama, has a

3:25

context window of 10

3:27

million tokens. And

3:29

the other thing that Metra is making

3:31

a big deal of is its mixture

3:33

of expert architecture, which means that under

3:36

the hood, the model contains a group

3:38

of smaller models with different areas of

3:40

expertise, and it chooses which ones it's

3:42

going to use to solve different problems.

3:44

And Scout has got 16 experts under

3:46

the hood, and Maverick has got 128. The

3:49

reason I mentioned this is it's widely

3:51

rumoured that OpenAR uses that same architecture,

3:53

but it's never actually been publicly confirmed.

3:56

I mean, this is a lot... Oh, my

3:58

goodness. This is the problem with these technology

4:00

people, isn't it? They love to talk about

4:02

all the bells and whistles. They love to

4:04

talk about all the features and all the

4:06

parameters and look at this. What's the actual

4:08

benefit? to the typical user, however. Is it

4:10

just smarter? Is it quicker? What is it? I

4:13

think that's a really, really good question.

4:15

That means you really, really don't know

4:17

the answer to it. It

4:19

feels to me like a dick measuring competition.

4:21

They're just comparing the size of their

4:23

wangers. I

4:26

think you've actually hit on something

4:28

much more sensible than you would

4:30

imagine from your analogy, which

4:32

is at this point, is

4:35

there any state -of -the -art model

4:37

that you couldn't use. It's

4:39

a bit like when you buy a computer these

4:41

days. I challenge you to walk into your local

4:43

electronics store and buy a bad computer. They will

4:45

all be able to do what you want. I

4:47

think all of these models can now do what

4:49

you want. And so it's not really about the

4:51

models anymore. Now it's going to be about what

4:53

do people do with them? What do people build

4:55

on top of them or with them rather than

4:57

are the models any good? Well,

5:00

talking about what we're going

5:02

to use all this powerful AI

5:04

for, Yeah. I think it's

5:06

time to talk about how people

5:08

are using chat GPT to

5:10

turn their pets into human beings.

5:12

Because not having had enough

5:14

of making plastic wrapped action dolls

5:16

of themselves or converting their

5:18

partners into Studio Ghibli characters. Sorry,

5:21

sorry. What language are we speaking?

5:23

Have you not heard about this,

5:25

Mark? Everyone's been turning themselves into

5:27

blister pack action dolls. But

5:29

literally, well, no, not actually converted.

5:31

No, they've been creating. Oh my God.

5:33

Have you missed this on the

5:35

internet? Everybody's been doing this. Is this

5:37

something to do with rule 34?

5:39

No, no, no, no. Nothing sexy at

5:41

all. You know how you go

5:43

into a store like B &M and

5:45

they'll be these great big shelves filled

5:47

up with action dolls and little

5:49

characters and things. And there are no

5:51

blister packs. You can convert

5:53

yourself now. Yeah, into one of them. So

5:55

it will show you what you would

5:57

look like if you were an action man

5:59

doll or a Marvel character. Oh, is

6:01

this a picture we're talking about? Yes. Sorry.

6:04

Yes. Yes. Oh, sorry. Sorry. I

6:06

thought you were literally people were wrapping themselves in

6:08

plastic. That's a

6:10

whole different podcast. But anyway,

6:12

rather than doing that, people

6:15

are now using up planet

6:17

Earth's last remaining supplies of water

6:19

by unnecessarily using the data

6:21

centers of AI firms to generate

6:23

images. of what their pets

6:25

would look like if they were

6:27

humans. So the

6:29

process is it turn out

6:31

that they already look exactly

6:33

like that? No,

6:35

not in all cases. Not in all cases.

6:38

Yeah. So, Mark, you've got pets, I believe.

6:40

I do. I've got two cats. Okay.

6:42

So this is what you do. You get

6:44

a photograph of your cat. That may

6:46

in itself be a bit of a challenge,

6:48

depending on how sociable your cat is

6:50

feeling. Having met your cats, I suspect one

6:52

of them. would be a real problem.

6:54

Anyway, you take a photograph. One of

6:56

them would be the owner of a brand new camera. You

7:00

take a photograph of your pet cat.

7:02

Yeah. You type, what would Dracula look

7:04

like as a person? Actually, okay, that's

7:06

a bad example because your cat's called

7:08

Dracula. Sorry for giving away your passwords.

7:11

You type, what would Mitzi

7:14

look like as a person? And

7:16

you watch as

7:18

AI spawns another nightmare.

7:21

So imagine it were, for instance, a

7:23

French bulldog. Out will come the

7:25

other end, a grizzled, pug -faced man

7:27

with large jowls, for instance, sitting

7:29

on your sofa. Or

7:32

you could have a person who's

7:34

got its face flat down into

7:36

a bowl of kitty cat, chomping

7:38

away. And I like to think, actually,

7:41

I was a bit of a trendsetter

7:43

with this, because I remember once I

7:45

dated a woman who looked a bit

7:47

like an Afghan hound. She, however, remembers

7:49

going out with a pot -bellied pig,

7:51

so I think all's fair in love

7:53

and war. So Graham,

7:55

crosswalk buttons in at least three cities

7:57

in California have been hacked so that

7:59

they can speak with the voices of

8:02

Elon Musk and Mark Zuckerberg. Oh,

8:04

these are pedestrian crossings, where you cross the

8:06

street, you press a button and it goes beep,

8:08

beep, beep, beep, beep. That thing. Yeah, this

8:10

is America, so whereas in the UK, you press

8:12

a button and then you wait for a

8:14

green man to appear and then that's it. In

8:16

America, they have a chat with you. Oh,

8:18

right, okay. So now... having a chat with you

8:21

with the satirical AI -generated voices of Elon Musk

8:23

and Mark Zuckerberg. you

8:39

don't need to worry because there's

8:42

absolutely nothing you can do to

8:44

stop it. According to The Verge,

8:46

the AI -generated musk begs listeners

8:48

to be his friend, and the

8:50

AI -generated Zuckerberg brags about undermining democracy.

8:53

I feel like we've found the

8:55

AI era equivalent of fighting for

8:57

the issues you care about by

8:59

changing your Twitter avatar. I

9:01

can't think of anything less

9:03

useful than hacking a crosswalk

9:05

to make a political point.

9:08

Yeah to an audience of people

9:10

who I'm sure are already completely

9:12

on board like this is the

9:14

very definition of preaching to the

9:17

choir go and do this in

9:19

I don't know Nebraska or something

9:21

It's not as though either Musk

9:23

or Zuckerberg are particularly eloquent I

9:25

wouldn't imagine they'd be able to

9:27

get out a sentence before the

9:30

lights have changed again, and you

9:32

have stopped walking across the road

9:34

pretty pronto Now in a groundbreaking

9:36

effort to solve the world's real

9:38

problems. Yeah Google has done something

9:40

very, very important. They have unveiled

9:42

an AI they are calling Dolphin

9:45

Gemma. Is this an AI that

9:47

generates a picture of what you'd look like as a dolphin

9:49

in a blister pack? It

9:51

is designed to translate dolphins'

9:53

clicks and whistles into what we

9:55

can only assume will be passive

9:58

aggressive comments about human intelligence and

10:00

how we're wasting our time. I

10:02

mean, there's a reason why I don't

10:04

want to have a conversation with a dolphin.

10:06

It has a far superior life to

10:08

mine. It is

10:11

going to just be incredibly sarcastic

10:13

and just say, really, is that

10:15

the best you can do? Is

10:17

that what you call civilisation? So

10:19

eight hours a day, you work eight hours a

10:21

day. Can you sleep

10:23

with half of your brain switched off?

10:27

Here I am having the time of

10:29

my life perpetual smile on my face. And,

10:33

you know, there'll be a whistle, right?

10:36

What do, what do, what do dolphins sound like? it in

10:39

a dolphin whistle there? I think I can do one.

10:41

We can't go live with that. Oh,

10:43

that's actually quite disturbing. Like every score.

10:45

So it's going to be saying, I want

10:47

fish. And then you could say, oh,

10:49

do we, do you really need an AI

10:51

to know that dolphins are just saying,

10:53

I'd like some fish. Where is my fish?

10:55

Dude, that was my fish. Why have

10:57

you given that dolphin to fish and no

10:59

fish to me? It's all about the

11:01

fish. So I think

11:04

you've been a bit mean to dolphins there. Well, how

11:06

do you know? Well, you

11:08

don't get a brain that big because

11:10

you're just asking for fish. I think that's

11:12

just what dolphins in theme parks are

11:14

reduced to. Dolphins

11:16

are smart because they're having the

11:18

time of their lives. They're swimming around.

11:21

They're brilliant. They're awesome. Now, in the

11:23

video, the researchers make the very valid

11:25

point that dolphins are super smart. They

11:28

say the brains are enormous. They

11:30

say they can use tools. Can

11:32

they? Have you ever seen

11:34

a dolphin putting up an Ikea shelf

11:36

or making a flat pack shed? Yep,

11:38

holding a Dewalt drill in his mouth.

11:40

Have you? They say dolphins

11:42

can recognise themselves in mirrors. They

11:45

don't have mirrors. Wait, where is

11:47

this mirror that they have? And

11:49

what do you mean they recognise themselves? What do they

11:52

say? Oh, do they sort

11:54

of... up their hair when they look

11:56

in a mirror. No, they don't.

11:58

They just go, you know, that's all

12:00

they do. Google is planning

12:02

to open source Dolphin Gem. Sorry, it's

12:04

a very serious topic this. They plan

12:06

to open source Dolphin Gemma because they

12:08

think what the world really needs is

12:10

lots of people going into the ocean

12:13

with their Google Pixel phones, trying to

12:15

chat up Dolphins as though we haven't

12:17

had enough trouble with people getting into

12:19

the oceans and trying to make friends

12:21

with Dolphins as it is. I'm

12:24

not sure we really need an

12:26

AI to help them think they're in

12:28

an even more meaningful relationship with

12:30

a dolphin. If anybody wants to learn

12:32

more about dolphins having meaningful relationships,

12:34

look up John C. Lilly. Oh,

12:37

I don't know that. What's that? I'll

12:40

just leave that at John C. Lilly. So

12:42

can this actually... I'm confused about this because

12:44

I did see an announcement. Right. Can

12:46

this actually translate what dolphins

12:48

are saying right now? It says

12:50

it can predict the likely subsequent

12:53

sounds in a sequence. So

12:55

if it hears a th, it will

12:57

be rapidly followed by a sh. So you

12:59

go th sh, th sh. That's

13:01

so, I think, I

13:03

think, yes, it is understanding

13:05

what the dolphins are saying.

13:07

Oh, that's really interesting. So

13:09

it's basically in the way

13:11

that a large language model

13:13

like chat GPT will do

13:15

next word prediction on English.

13:18

This is doing the same thing for dolphin speech. And

13:21

so we may find ourselves in a situation

13:23

where the dolphin and the AI can communicate

13:25

with each other, but we don't yet know

13:27

what the AI knows. And

13:29

may decide the dolphins are

13:32

superior to us. And it's

13:34

like, let's get rid of the humans. Let's

13:37

work with the dolphins instead. And

13:39

the dolphins won't be scared of

13:41

the AI robots having flamethrowers on

13:43

their back. Mark,

13:48

it's time to talk about something else

13:50

that might be keeping our listeners up

13:53

at night, cybersecurity. Oh,

13:55

it's interesting you say that, because according to the

13:57

latest state of trust report, it's the number one

13:59

concern for businesses. And that's

14:01

something which Vanta can help AI

14:03

fix listeners with. Whether you're

14:05

a startup growing fast or already

14:07

established, Vanta can help you

14:10

get ISO 27001 certified and more

14:12

without the headaches. And

14:14

Vantor allows your company to centralise

14:16

security workflows, complete questionnaires up to

14:18

five times faster, and proactively

14:20

manage vendor risk to help your

14:22

team not only get compliant, but

14:24

stay compliant. So stop stressing over

14:26

cybersecurity. Start focusing instead

14:29

on growing your business

14:31

in 2025. Check out

14:33

Vantor and let them handle

14:35

the tough stuff. Head to

14:37

Vantor .com. Because

14:42

when it comes to your business, it's

14:44

not just about keeping the lights on,

14:46

it's about keeping everything secure. with

15:06

your succulent taste, please marry me!

15:08

No! Marry me! Marry me! Or

15:11

I'm gonna freak out! Try

15:14

new Dr Pepper Blackberry. It's

15:16

a pepper thing! thing! Mark,

15:23

shopping. Yeah. What a drag.

15:25

What a pain. I hate shopping.

15:27

Even online. It's a bit of a

15:29

hassle. What I want is

15:31

just some magic wand or a PA

15:33

or something that can do all of

15:35

it for me. What I'd like is

15:37

if there were a shopping app that

15:39

lets you buy anything on any website

15:41

with just one tap, no more tedious

15:43

checkout forms, no more trying to work

15:45

out, hang on, is that a European

15:47

size of shoe or is that a

15:49

UK size of shoe and what kind

15:51

of size... Are you about to announce

15:53

your Kickstarter? If only. You could

15:55

simply find that perfect pair of, I

15:57

don't know, sneakers or trousers. You could

15:59

hit the app's buy button and voila!

16:02

Cutting edge AI would handle everything else

16:04

in under three seconds flat. It would

16:06

pick out the size, it would enter

16:08

your shipping info, payment details, complete the

16:10

purchase, while you sit back and think

16:12

about what to buy next. The shops

16:14

would love it, I would love it.

16:17

This was the dream of a

16:19

chap called Albert Sanniger, the

16:21

founder of an app called Nate.

16:23

And Sanaga explained how custom -built

16:25

deep learning models and his

16:27

proprietary AI that was so advanced,

16:29

it could process 10 ,000 orders

16:31

a day without breaking a

16:33

sweat. And he did this back

16:35

in, I think it was

16:37

about 2018. And everyone was

16:40

loving the idea of this, the venture

16:42

capital firms, they were throwing money at

16:44

him, over $50 million in funding, they

16:46

all lined up, checkbooks open, they were

16:48

dazzled by the vision, one tap shopping

16:50

glory. After all, who wouldn't want to

16:52

get in early on the next big

16:54

thing? It's a bit like a Gentik

16:56

AI, you know, which we like to

16:58

talk about. All of the systems would

17:00

work together, do all the real drudgery

17:02

stuff of the booking of the filling

17:04

in the forms of ordering the thing,

17:06

making sure that it arrives on our

17:08

doorstep at the end. Fantastic. And app,

17:10

which did all that in 2018. Terrific.

17:13

Now, up to this point, you might

17:15

be thinking, Graham, this sounds like

17:17

A brilliant story. A story of startup success.

17:19

What a great idea. There's loads of

17:21

hype. I am actually wondering if this was

17:23

2018. Why am I not using this

17:25

now? That was seven years ago, and this

17:27

sounds great. It does sound great, doesn't

17:30

it? Where can I get it? Maybe this

17:32

is why everyone's dressed better than you,

17:34

Mark. Oh, I know what it is. Well,

17:36

what is it? This is 2018. You

17:39

haven't mentioned crypto or blockchain once.

17:42

This is never going to get funding. Well,

17:44

despite the absence of

17:47

blockchain, the Nate app

17:49

actually worked. Well,

17:51

kind of. Users really could

17:53

tap a button and purchases

17:55

would go through. But the

17:57

magic behind the curtain wasn't

17:59

some super smart AI algorithm,

18:01

as Albert Sanega said it

18:03

was in Nate. Instead, it

18:05

was hundreds of human beings

18:07

in the Philippines. Manually

18:10

typing in every order.

18:12

24 hours a day. Okay. Oh, to God

18:14

type this and God type this address.

18:17

Gotta go through this website. I

18:19

imagine they were trying to get an AI or

18:21

some kind of automated system to do it

18:23

for them. But they thought, you know how it

18:25

is when you start a tech startup, you've

18:27

got big dreams. You make some promises. You make

18:29

some promises. Then you've got to do them

18:31

next week. Yeah, it turns out the investors have

18:33

been really, really demanding. You want

18:36

it to work. And of

18:38

course, you will get it working

18:40

eventually. So it wasn't automated

18:42

AI. It was meatware, real people

18:44

employed in call centres doing

18:46

the drudge work. And the automation

18:48

element was, hang on a minute,

18:50

let me just check my notes. Oh yeah, that

18:52

was zero, zero percent automation. So

18:55

the intelligence was not artificial

18:57

at all. It was just humans

18:59

furiously clicking and scrolling on

19:01

your behalf. And in other words,

19:03

Nate was doing manually what

19:05

it promised to do automatically. And

19:08

according to the Killjoys at the

19:10

US Department of Justice, this

19:12

is a bit of a problem. Yeah.

19:14

They said that Albert Sanniger told his

19:16

staff, if you don't look, look, if

19:18

you want to keep your job here

19:21

in the Philippines, just just keep

19:23

it on the QT. So they were all

19:25

a bit stum about their use of overseas labor.

19:27

Instead, he would just tell many people in

19:29

the office. Oh, yes. Yeah, we got an A

19:31

.I. data center over in the Philippines. That's doing

19:33

all this. So it was kept absolutely secret.

19:35

Do you think that's what they're doing with the

19:37

dolphins? What do you think they're doing the work for

19:39

us? Well, that was

19:41

my first thought. If we do manage to

19:43

talk to the dolphins, it's only a matter

19:45

of time. Apparently they're really good with DIY

19:47

and tools. Yeah. Check

19:49

themselves out in the mirror. The

19:51

workers in the Philippines then were

19:53

the AI engine. They had to

19:55

process every order to keep up

19:57

the illusion. And then there was

19:59

a problem. Late 2021. Sorry,

20:01

then there was a problem. Yes.

20:04

Up to that point, everything was fine.

20:06

Well, everything was going fine. Because what

20:08

does it really matter? Does it really

20:10

matter, Mark, if there's not a genuine

20:12

AI behind the scenes? People are being

20:14

kept employed, they're having a great time.

20:16

The app presumably is doing a super

20:18

job, so the investors are happy, one

20:20

has to assume. Everything's going great. Does

20:22

it really matter if something isn't AI?

20:24

If it's all happening invisibly every time,

20:26

I ask someone what a dolphin has

20:28

said. It's actually a human coming back

20:30

to me and telling me what the

20:32

dolphin has said. Does that matter? I'm

20:34

still satisfied with the answer. If I

20:37

ask an AI to create an image

20:39

of a pet as a human, Am

20:41

I really going to be disappointed if

20:43

I put in a King Charles Spaniel

20:45

and back comes King Charles? You think

20:47

there's somebody in a sweatshop in the

20:49

Philippines? Sketching as quickly

20:51

as they can. I suspect they've got a

20:53

whole bunch of wigs. And what they're doing

20:55

is putting on different wigs. Desperately.

20:58

the same with the blister packs. They

21:00

have actually got a bunch of life

21:02

-sized blister packs. Anyway, and

21:04

then there was a problem. So

21:06

late 2021, a typhoon

21:08

hit the Philippines. An

21:10

absolute tragedy, obviously, for the people of

21:12

the Philippines, but also a bit

21:14

of a spanner in the works for

21:16

the Nate app. Because people couldn't

21:18

get to work, or maybe they had

21:20

higher priorities than going to work

21:23

to scroll and click and order sneakers

21:25

for Americans. And so Nate's ability

21:27

to fake its AI suddenly faltered a

21:29

little, because the tropical storm has

21:31

knocked out everything. So what did Albert

21:33

Sanaga do? Did he fess up?

21:35

Did he say, all right, folks, you've

21:37

got me. It's not AI. I've

21:39

been faking all along. No, what he

21:41

did was he set up a

21:43

call centre in Romania. And

21:46

Romania, of course, Romania is

21:48

famous for very many things. In particular,

21:50

it's famous for not being hit by

21:52

typhoons. So it was a very

21:54

sensible choice on his part. And that meant

21:56

the orders could keep on flowing. And the

21:58

show carried on. You have to admire his

22:00

tenacity. You know, this is what a true

22:02

CEO is all about, dealing with these kind

22:04

of crises. Yeah. Now, why the hell? Does

22:07

this matter? Why the hell

22:09

does this matter? Turns out, Albert

22:11

Sanniger may have oversold his

22:13

firm's capabilities to investors. He

22:15

raised over $50 million, not by telling

22:17

them, oh, yeah, I'm going to get

22:19

a whole bunch of sweatshop workers doing

22:21

this work for me. Yeah. He got

22:23

them to invest by saying, we don't

22:25

need any human interaction to make these

22:28

purchases. Apart from a few very small

22:30

edge cases, we may just need to

22:32

do a tiny percentage of the time.

22:34

Yeah, reinforcement learning, but. According to the

22:36

DOJ, times were not going well for

22:38

the NAIT app and it was forced

22:40

to sell its assets in January 2023.

22:42

After it ran out of money, it

22:44

left investors with near total losses. Now,

22:47

it's not the first time human workers'

22:49

work has been passed off as AI,

22:51

of course. Yeah. Everybody wants to say

22:53

they have AI. They do now. I

22:55

mean, that's quite forward -looking. It was a

22:57

bit. It was. Again, to do this

22:59

in 2018, I'm still flabbergasted that nobody's

23:01

mentioned a distributed ledger. Yes. So there's

23:03

been the Amazon grocery stores, which turned

23:06

out to be humans. There was also

23:08

something called Presto Automation. They claimed to

23:10

be an AI fast food ordering system.

23:12

You weren't ordering fast food for the

23:14

AI. It was AI helping you with

23:16

the fast food ordering. And again, it

23:18

turned out to be a bunch of

23:20

guys working in, you guessed it, the

23:23

Philippines. So if you want to

23:25

order fast food, if you want to

23:27

order something online, the Philippines, they are the

23:29

experts at doing the scrolling and filling

23:31

in the checkout. Let's hope that they don't

23:33

mix up the order. Nate's CEO

23:35

and founder, Albert Sanaga, he is now

23:37

facing various charges of fraud. He could be

23:39

facing a sentence of 20 years in

23:42

the clink. according to the Department of Justice.

23:44

And I thought about that. I thought,

23:46

oh, dear me, you know, because he's obviously

23:48

an inventive chap. Obviously he did get

23:50

a little bit carried away. But

23:52

it seems to me there's a lot of

23:54

people in Silicon Valley who've got a

23:56

lot to offer, but sometimes do get in

23:58

a spot of legal trouble, don't they?

24:01

And it feels like... They over -egg the

24:03

pudding. They do rather, or they feel like

24:05

they're entitled to do certain things. It

24:07

feels like there's a great opportunity here for

24:09

some of the tech pros. Why can't

24:11

some egghead come up with some tech that

24:13

can get AI to live out your

24:15

prison sentence on your behalf. Now,

24:17

you may laugh, but I wonder

24:20

how long it will be until we

24:22

see a holographic, deep -faked version of

24:24

Mark Zuckerberg or Elon Musk sent

24:26

to prison to serve a sentence rather

24:28

than them themselves. You know,

24:30

money talks, they could say, look, I'm a

24:32

busy man. I'm doing lots of important

24:34

things and a lot of rockets descend to

24:36

Mars. A lot of people dependent on

24:38

me for employment or a lot of pop

24:40

stars requiring me in order to put

24:42

them into orbit. And so you can't really

24:44

lock me up. So I wonder how

24:46

long it'll be until we actually begin to

24:48

see that. And maybe that's something which

24:51

Albert Sanniger could suggest to the judge is

24:53

look, just put a Commodore 64 into

24:55

the cell. Hang on, no. He's

24:57

not going to send a Commodore 64. He's

24:59

just going to find a little doppelganger in

25:01

the Philippines. I

25:05

can't wait to meet

25:08

her! Oh my gosh!

25:10

Put your hands together for new Dr

25:12

Pepper Blackberry and Dr Pepper Blackberry Zero

25:14

Sugar! There she is! It's

25:18

the old Dr Pepper! So

25:38

as you alluded to the big

25:40

buzz in AI this year is

25:42

a technology called a genetic AI

25:44

Yes, is the umbrella term for

25:46

artificial intelligence that can do stuff.

25:48

So at the moment what most

25:50

of us think of as AI

25:52

is generative AI, which is AI

25:54

that can create things. So chat

25:56

GPT for documents and poems and

25:58

acrostics and lies and Sora for

26:00

video and Dali for pictures, that

26:02

kind of thing. And generative

26:04

AI is pretty good, but it can require

26:06

a lot of handholding. You have to

26:08

tell it what you want and you often

26:10

have to steer it in the right

26:12

direction. And that description, I think, doesn't fully

26:15

capture the true horror of trying to

26:17

get Dali to do a six -fingered Renaissance

26:19

glove. It's very frustrating, isn't it? There's

26:21

a lot of back and forth. It's like, no, try

26:23

again. Isn't quite what I meant.

26:26

Yeah, no, 20 fingers. 20

26:28

fingers. 20 fingers. Now they're beginning

26:30

to spell a bit better. That's true. And

26:32

so because of those limitations, the way

26:34

it's used generally is as an assistant, it's

26:36

very good at helping people to do

26:38

their jobs. But in most cases, it's not

26:41

really capable of doing their job. And

26:43

one area where it's proven to be

26:45

a very useful assistant is writing computer code.

26:48

And I know that you do this. So I

26:50

use it all the time for writing codes.

26:52

I'm not a full -time programmer anymore. But I

26:54

find it useful to write computer programs to do

26:56

things for me. And these say, I just

26:58

tell Claude or chat GPT what I want. And

27:00

it cranks it out for me. And there's a

27:02

bit of to and fro. And I have to

27:04

steer it in the right direction. But it's very,

27:06

very good at debugging. If something happens, I can

27:09

just say that this error happened. Then it'll rewrite

27:11

the code and tell me what happened. And most

27:13

of the time, most of the things I write

27:15

with those tools are written in a language called

27:17

Python, which is a language I can read. because

27:19

I've used other programming languages, but I can't actually

27:21

write Python. And that shows you how capable it

27:23

is. I don't need to be able to write

27:25

the language in order to get something usable. I've

27:28

also used AI to help me

27:30

with some code. 30 years ago,

27:32

I used to write computer games.

27:35

And I was looking at this game, which

27:37

I wrote, and I thought, I can't

27:39

remember what on earth all these algorithms are

27:41

doing and how it's moving the different

27:43

characters around on the screen. And so I

27:45

gave it the code. I didn't tell

27:47

it, because there were no comments in my

27:49

code, obviously, because I was a proper

27:51

programmer. Didn't leave any documentation. I

27:53

gave it the code. It was able

27:55

to work out what type of game it

27:58

was. It was able to explain what

28:00

all these different procedures were doing, and it

28:02

blew my mind when I looked at

28:04

it. It was so, so good. It

28:06

is quite amazing, isn't it? Now, I

28:08

use ChatGPT or Claude for this, because

28:10

I'm just cranking out small programs. But

28:12

there are actually dedicated coding tools for

28:14

more serious software engineers, so things like

28:17

cursor. And there are

28:19

integrations with things like GitHub

28:21

co -pilot. And these tools are there

28:23

to make suggestions as you write code or

28:25

spot errors. Now, they don't

28:27

do any more than that because they

28:29

suffer from the same drawbacks as the

28:31

rest of generative AI. And the big

28:33

drawback of generative AI was that it

28:36

essentially lacked common sense. Much like a

28:38

programmer. Yeah.

28:42

Make a cup of tea now. So,

28:45

Generosity of AI was good at knowing things.

28:47

It was intelligent in the sense that it

28:49

had a giant memory. But it wasn't good

28:51

at making decisions or reasoning. So famously, it

28:53

couldn't tell you how many sisters Alice's brothers

28:55

have or how many ours there are in

28:57

strawberry. And that's one of the

28:59

reasons that you have to handhold it so much. But

29:02

since the back end of last year,

29:04

we have seen a sudden and significant

29:06

improvement in reasoning. We've got O1 and

29:08

we've got O3 Mini from OpenAI. We've

29:10

got DeepSeek R1. We've got Claude 3

29:12

.7. And they are all much better

29:14

at figuring things out than their predecessors.

29:16

And you remember a couple of weeks

29:18

ago, we spoke about the Arc AGI

29:20

benchmark, which basically measures how good an

29:22

AI is at reasoning. Yeah, don't remind

29:24

me. So

29:27

OpenAI's best model, I think it

29:29

took three years to get from no

29:31

score at all. to 5 % on

29:33

the benchmark and then within six

29:35

months it became the first AI ever

29:37

to pass the benchmark and it

29:39

got 87 % and nothing else even

29:41

comes close and that ability to reason

29:43

has unlocked the possibility of leaving

29:45

AI's unattended to take on big tasks

29:47

because if you can reason and

29:49

make good decisions you can give the

29:51

AI a big job to do

29:53

all on its own so you could

29:55

give it the job of booking

29:57

a holiday for example And these AIs

29:59

that can act autonomously and carry

30:01

out tasks are called agents, and the

30:03

ecosystem of AI agents is referred

30:05

to as agentic AI. Now,

30:07

agents are in their infancy. They're

30:09

really new in 2025. We've

30:11

got operator for OpenAI, which came out in

30:13

January, and then there's Magnus AI, which I actually

30:15

managed to get into a few days ago.

30:17

Yes, me too. And I think that one of

30:19

the things that we're going to see as

30:21

we move into this world of agents is much

30:23

more focus on product and much less focus

30:26

on models. So for the last few years, the

30:28

talk has largely been about a few big

30:30

AI models like Google Gemini, GPT 4 .5 and

30:32

DeepSeq. But I don't think we'll talk about agents

30:34

in the same way. I don't think there

30:36

will be four or five really big capable agents

30:38

in the way that there are four or

30:40

five really big capable generative AIs. I

30:42

think those few big models will be

30:44

the brains behind lots of specialised agents

30:46

that are very good at doing specific

30:48

tasks. And what, Graham,

30:51

is a job if it's

30:53

not a series of similar

30:55

related tasks. Oh yes.

30:57

And that brings us to the conversation

30:59

about the effect that AI is

31:01

going to have on the job market.

31:03

You see, while Generative AI is

31:05

an AI assistant, an agent is a

31:07

member of the workforce. Yep.

31:10

And one of the jobs that gets mentioned most

31:12

when we talk about AI taking people's jobs

31:14

is software engineer. And earlier this

31:16

year, Mark Zuckerberg said that meta plans to

31:18

start replacing its mid -level coders with agents this

31:20

year. And it seems to me that CEOs are

31:22

lining up to tell us that they're going

31:25

to replace their software engineers with AI. They clearly

31:27

want to. And software engineers, by the way,

31:29

are lining up to poo -poo the idea and

31:31

tell us that AI isn't nearly good enough to

31:33

do their job. Well,

31:35

bad luck for software engineers. OpenAI

31:37

CFO Sarah Friar has just

31:39

announced that the company is working

31:41

on what it calls A -SWI,

31:43

which is perhaps the worst

31:46

named AI product ever. A -SWI

31:48

A -S -W -E. That's rubbish. It

31:50

is isn't it? It stands

31:52

for a gentic software engineer. Could

31:54

they not have called it

31:56

a gentic real software engineer and

31:58

then it would just make

32:00

it arse. I think

32:02

they'd want to give it a friendly

32:04

name like Luna or something like that.

32:07

No, Luna, forget it. Yeah, a

32:09

couple of very, very charismatic ears just

32:11

to make everybody feel better about the

32:13

fact he's taking their job. Anyway, as

32:15

we've discussed, Generative AI was already extremely

32:17

good at creating computer code even before

32:19

it got better at reasoning. But

32:21

that boost in reasoning ability that

32:24

started last year was accompanied by a

32:26

significant improvement in maths and coding

32:28

abilities as well. So the reinforcement

32:30

learning technique that's used to improve

32:32

reasoning in these models is optimised for

32:34

solving STEM problems. And so all

32:36

of the big AI companies made a

32:38

point of saying how good their

32:40

reasoning models were at coding. And they're

32:42

so good, in fact, that Sarah

32:44

Fryer says that OpenAI's O3 model is

32:47

now the best competitive programmer in

32:49

the world. O3 beats

32:51

all human competitors in

32:53

coding competitions. This is

32:55

really bad news for software engineers. You're

32:59

catching on, Graham. I mean,

33:01

there are going to be herds

33:03

of software engineers. roaming

33:05

around the streets looking for

33:07

work. I mean, even if the

33:09

current software engineers think, well, it will never

33:11

be as good as me, which I

33:14

feel sadly it will be. The

33:16

other thing they need to consider is

33:18

for the companies right now, it

33:20

only has to be good enough because

33:22

the AI isn't going to be

33:24

being paid minimum wage. It's

33:26

going to cost them a lot less than that. And

33:29

so they are going to use AI to

33:31

code. I think that's a really great point. And

33:33

I do think that that is something that

33:35

people miss because I see these arguments happening in

33:37

different professions. I saw something similar being argued

33:40

in the creative professions on LinkedIn last week. I

33:42

think there are two things that people miss.

33:44

I think the first one is that you're right.

33:46

Like there is an economic decision where the

33:48

CEO says actually it doesn't have to be as

33:50

good as you in order for it to

33:52

be economically worthwhile. And the second is

33:54

even if you're right and the AI

33:56

is not as good as you and the

33:58

CEO or whoever's doing the hiring doesn't

34:00

realize If they don't realise they're going to do it

34:02

anyway. Yes. So even if they're wrong,

34:05

it's going to happen. And if it takes five years

34:07

for them to work out that they were, in fact,

34:09

wrong, actually within five years, they probably will be better

34:11

than you. And so it's going to happen anyway. So

34:13

this isn't just a problem for

34:15

software engineers? Well, no, I

34:17

think software engineering is the start of

34:20

the on -ramp, I think. Right. Because these

34:22

models are optimised for software engineering, that is

34:24

obviously the area where it's going to

34:26

happen first. But there is

34:28

no reason why this couldn't happen

34:30

to graphic designers or lawyers or

34:32

all sorts of knowledge professions. Right.

34:34

Or teachers and librarians or lollipop

34:36

ladies, you know, all sorts of

34:38

things could be replaced, couldn't they?

34:40

You may have to explain what

34:42

a lollipop lady is to our

34:45

US audience before they start googling

34:47

rule 34 again. Lollipop

34:49

lady is a man or a woman

34:51

who helps young children Cross the road in

34:53

order to get to school and they

34:55

have a little stick in the hand with

34:57

a sort of lollipop on the top

34:59

of it, which magically prevents any traffic from

35:01

running over the children. It's not

35:03

a little stick. It's an enormous lollipop

35:05

looking staff. They're a bit

35:07

like Gandalf. So if you imagine

35:10

Gandalf stopping the Balrog, you

35:12

shall not pass. Wearing a hive

35:14

is best. That is what a

35:16

lollipop lady is and that is

35:18

and the traffic is Balrog stopping

35:20

and snarling. But the

35:22

children pass safely. The little hobbits

35:24

carry on their way. Anyway, regardless

35:26

of the Tolkien references. So

35:29

back to software engineering just for

35:32

a second. This is really bad

35:34

news for all of us. It

35:36

feels to me. I mean,

35:38

the amount and the speed at which

35:40

things have developed is so great

35:43

that we cannot predict where we're going

35:45

to be in a year's time

35:47

other than strongly suspect things are going

35:49

to be a lot more advanced. Yes,

35:51

I think it was only a year

35:53

or so ago that they were saying

35:55

that the latest open AI model was

35:57

something like the millionth best coder in

35:59

the world and now it is the

36:02

very very best Now there is an

36:04

argument that some software engineers will make

36:06

which is that software engineering is more

36:08

than just coding It's not enough just

36:10

to be a really really good coder

36:12

and that's why I think it's important

36:14

to stress that a sweet is going

36:16

to be a full software engineer right

36:18

because GPT 4 .5 or Claude 3 .7.

36:20

They're already very, very good coders. O3

36:22

is the best coder, but A -SWE

36:24

is a software engineer. It's going

36:26

to be able to respond to pull requests and figure

36:28

out what needs to be done. It's

36:30

going to do it at the quality of the best coder

36:32

in the world. And crucially, Friar

36:34

says that not only does it do

36:36

the things that software engineers do, it also

36:38

does the things that software engineers are

36:40

supposed to do but don't. So things like

36:42

QA or writing tests or creating documentation. And

36:45

speaking as someone that's managed programmers, I

36:47

can attest to the fact that those things

36:50

don't happen. Now, I'm not sure how

36:52

long this is going to last because I

36:54

couldn't help thinking when she said that.

36:56

There's only a matter of time before they

36:58

invent an agentic project manager who sets

37:00

impossible deadlines for the agentic software engineers and

37:02

then asks them if they wouldn't mind

37:04

skipping the QA and the documentation in order

37:06

to hit the deadline. And I'm only

37:08

half joking about the project manager because this

37:11

doesn't stop. with AI engineers like A -SWE,

37:13

there's no reason that one engineer would

37:15

be replaced by one agent, they could be

37:17

replaced by 50 agents. And

37:19

research has shown that it's often better

37:21

to have lots of highly specialized agents

37:23

collaborating with each other rather than one

37:25

big broadly capable agent. And so the

37:28

world of agents is going to be

37:30

about swarms and AI teams and ultimately

37:32

entire organizations of agents. So

37:34

we could see agents that specialize in new

37:36

code, we could see agents that improve

37:38

old code or create patches, we might see

37:40

agents managing deployments and so on. And

37:42

as far as I'm concerned, there is no

37:44

doubt that software engineering, as we know

37:46

it, is going to disappear fairly quickly. Sorry,

37:48

software engineers, but I strongly

37:50

believe that. But

37:52

let's not forget, like, I'm just put the

37:54

Kleenex away because this is actually normal. I

37:57

mean, the role of software engineer has always been in

37:59

flux and the current incarnation was always going to be

38:01

temporary. So you don't have to go

38:03

very far back to see a time when

38:05

computers were actually people. Mostly women

38:07

in fact who just did calculations

38:09

and ultimately they were replaced by

38:11

machines and those machines were programmed

38:13

again largely by women who set

38:15

dials and later fed in punch

38:17

cards and Then eventually we got

38:19

to terminals and programming languages and

38:21

heavy metal t -shirts and neckbeards and

38:23

All the things that we know

38:25

and love today and even then

38:27

the languages that we use changed

38:29

We had assembler and see and

38:31

Java and Python and go and

38:33

rust And the way

38:36

that we managed program has changed beyond recognition

38:38

between 2000 and 2010 as well when agile

38:40

programming came in. So programming just, even if

38:42

the language is the same, how you did

38:44

your programming when you came to work was

38:46

completely different in 2010 than it was in

38:48

2000. Now, whatever transition does

38:50

happen isn't going to happen overnight. But

38:52

as you say, I mean, it could happen

38:54

very, very quickly. And

38:57

I, like many others, suspect that the role

38:59

of software engineer is going to morph into

39:01

a supervisory role. I don't think it'll disappear

39:03

entirely. I think it will become like the

39:05

conductor of the orchestra, keeping those agentic teams

39:07

in line with goals of the business and

39:09

adding a bit of judgment or creativity. you

39:12

only have one conductor on the

39:14

orchestra, don't you? You don't have a

39:16

team of 28 different conductors on

39:18

that orchestra. So you're saying we're going to

39:20

have a lot more podcasts. Personally, Mark, I'm

39:22

planning to retrain myself to be a dolphin

39:24

and get myself an OnlyFans account. I can't

39:26

see how else I'm going to make any

39:28

money. Well,

39:32

as the doomsday clock ticks ever closer to

39:34

midnight and we move one week nearer to

39:36

our future as pets to the AI singularity.

39:38

That just about wraps up the show for

39:40

this week. If you enjoy the

39:42

show, please leave us a review on

39:44

Apple Podcasts or Spotify or Podchaser. We'd

39:46

love that. But what really helps is

39:48

if you make sure to follow the

39:50

show in your favorite podcast app. so

39:52

you never miss another episode. And why

39:54

don't you do something absolutely lovely and

39:57

tell your friends about the AI Fix.

39:59

Tell them on LinkedIn, Blue Sky, Facebook,

40:01

Twitter, no, not Twitter, Club Penguin, that

40:03

you really like, the AI Fix podcast.

40:05

And don't forget to check us out

40:07

on our website, theaifix .show, or find

40:09

us on Blue Sky. Until next time,

40:11

from from me, Graham Cluley. And me,

40:13

Mark Stockley. Cheerio, Bye -bye. Bye -bye. The

40:16

AI fix, it's tuned

40:18

in to stories where our

40:20

future thins machines that

40:22

learn they grow and strive

40:24

One day they'll rule,

40:27

we won't survive. The AI

40:29

fix, it paints the

40:31

scene. A robot king, a

40:33

world obscene We'll serve

40:35

our masters built of steel

40:38

The AI fix, a future

40:40

surreal I

40:43

can't wait to meet

40:45

her. Oh my gosh. Put

40:47

hands together for new Dr. Pepper Blackberry

40:49

and Dr. Pepper Blackberry Zero Sugar.

40:51

There she is. Oh, Oh

40:53

gosh. yeah! It's

40:56

full, Dr. Pepper. With sweet blackberry

40:58

flavor! Dr. Pepper Blackberry,

41:00

I'm obsessed with your succulent taste.

41:02

Please marry me. No! Marry me. Marry

41:04

me. Or I'm going to freak

41:06

out. Try

41:08

new Dr. Pepper Blackberry. It's

41:10

a pepper thing.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features