Big moments in AI from Y2K to ChatGPT

Big moments in AI from Y2K to ChatGPT

Released Wednesday, 19th February 2025
Good episode? Give it some love!
Big moments in AI from Y2K to ChatGPT

Big moments in AI from Y2K to ChatGPT

Big moments in AI from Y2K to ChatGPT

Big moments in AI from Y2K to ChatGPT

Wednesday, 19th February 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

You're listening to

0:00

The Catalyst by Soft Choice, a

0:03

podcast about unleashing the full

0:03

potential in people and technology.

0:07

I'm your host, Heather Haskin. Imagine waking up in a world where

0:14

machines don't just assist us.

0:18

They think, create, and

0:18

even make decisions.

0:21

That world isn't decades away. It's happening right now.

0:25

In just a few years, AI has crossed

0:25

a threshold with platforms like

0:29

ChatGPT, moving it from an emerging

0:29

technology To a transformative

0:34

force, reshaping industries,

0:34

creativity, and society itself.

0:39

But how did we get here? In today's episode, we're tracing AI's

0:41

recent history, uncovering the pivotal

0:45

moments that shaped the industry. We'll explore the breakthroughs

0:47

that reignited deep learning,

0:50

the rise of neural networks, and

0:50

the key milestones that brought

0:54

generative AI into the mainstream,

0:54

revolutionizing the way we do business.

0:59

And the AI market isn't slowing down. With innovations, like China's DeepSeek,

1:01

shaking up the industry, what's next?

1:07

So how did AI go from niche research

1:07

to a force redefining our world?

1:13

Today we're joined by Gil Press, a

1:13

senior contributor at Forbes magazine.

1:18

Gil covers emerging technologies,

1:18

startups, big data, and the history

1:22

and future of artificial intelligence.

1:26

So, Gil, I'm really excited to

1:26

be able to interview you today.

1:29

I looked into your background a little

1:29

bit, and it seems like you have quite

1:32

an extensive bird's eye view on the

1:32

history of generative AI, and so that

1:36

brings to light so many questions.

1:38

As we think about AI today, it's

1:38

such a buzzword, but there seems

1:42

to be something different about

1:42

the moment that we're in right now.

1:47

And it makes me wonder,

1:47

have we turned a corner?

1:50

What's changed? And maybe talking about that history

1:51

will help us understand that.

1:54

So I'd love to hear from

1:54

you what your thoughts are.

1:57

What's changed recently is

1:57

that, uh, even at AI has been

2:03

around for at least 70 years.

2:05

So the research on the right, some,

2:05

some implementation of, of, uh,

2:10

various AI programs recently, both

2:10

the scope and the reach of AI.

2:18

have expanded significantly.

2:21

And now we have AI programs

2:21

that are expanding the menu

2:26

of what computers can do. For example, generate text or

2:28

generate images, the reach of AI

2:34

in the sense of, uh, I'm managing

2:34

to reach millions of computer users

2:41

around the world, mostly because of

2:41

the very recent success of ChatGPT

2:46

that's what really is happening now.

2:49

It's been a long, long journey, and

2:49

especially over the last 25 to 30 years,

2:57

in terms of very specific advancements

2:57

in what is called deep learning.

3:04

It's interesting to hear

3:04

about some of the changes that have

3:06

happened over time, Gil, and I'm really

3:06

excited to talk with you about some

3:09

of them because there's some things

3:09

that I hadn't really learned before.

3:13

So today we're looking at the

3:13

recent history of AI going back

3:15

to the turn of the 21st century.

3:17

An important paper was

3:17

written in the year 2000.

3:21

And I'd love to hear about that

3:21

paper from your perspective.

3:25

Why was it important? Why did it renew interest

3:27

in deep learning?

3:30

And maybe what was the initial impact in

3:30

the early 2000s after it was published?

3:35

That year there was another

3:35

advancement in deep learning in a sense

3:39

that there were three leading computer

3:39

scientists that insisted for many years

3:46

that this is the right approach to AI.

3:48

And now they had a lot of challenges

3:48

to that and a lot of skepticism.

3:53

The three were Geoffrey Hinton,

3:53

Jan LeCun, and Joshua Bengio.

3:58

And Bengio and his team published in

3:58

the year 2000, a paper on training

4:04

the computer to understand language,

4:04

text processing, language processing.

4:10

And the breakthrough there was the

4:10

idea that there will be a lot of

4:15

examples that mix together phrases with

4:15

similar words, but not similar meaning.

4:24

We humans understand it.

4:26

Computers at that time could not

4:26

understand it at all, got confused,

4:30

and then devised a model that

4:30

overcame that particular challenge.

4:35

This was basically the launching pad.

4:38

For what today we call large language

4:38

models, the ability of the computer

4:44

to understand context, to understand

4:44

the different meaning of words,

4:51

even if they're the same words.

4:53

So we see the next important

4:53

moment in AI between 2007 and 2009.

4:59

What happened then? How were data and processing power

5:00

factors in the technology of the day?

5:05

Why was deep learning a step

5:05

forward from machine learning?

5:09

It's actually an easy question. Deep learning is part of machine

5:11

learning, but there are a lot of

5:14

approaches to machine learning.

5:17

Uh, the advantage of deep learning

5:17

can be summarized in two words.

5:25

big data. So most of the traditional and very

5:26

successful machine learning approaches

5:32

relied on analyzing small sets of data.

5:35

Deep learning is different in the

5:35

sense that it shows its advantages, it

5:40

shows its benefits, specifically when

5:40

we deal with lots and lots of data.

5:46

And by the year 2005 2010, we had lots

5:46

and lots of data because of the invention

5:55

of the World Wide Web in the early 1990s.

6:00

By the way, what we see today in

6:00

this sudden expansion and triumph

6:05

of modern AI is very similar to what

6:05

happened When the World Wide Web was

6:10

invented, it was basically a piece of

6:10

software that was installed on top of

6:18

an existing worldwide network that was

6:18

established 30 years before the internet.

6:27

So today we see this triumph of

6:27

modern and I also based on not a

6:32

lot of work for many, many decades.

6:35

But specifically in around 2005 to 2010,

6:35

and specifically one AI researcher,

6:44

Fei Fei Li, came up with the idea that

6:44

if we have all these images on the

6:50

internet, and all of these images are

6:50

labeled, annotated, because people put

6:56

a picture of dog and they said, this

6:56

is my dog, Sasha, which is my dog.

7:02

But then you see it's, they are

7:02

identified, the images identified.

7:05

That's very important when you do

7:05

deep learning, because otherwise

7:09

you have to do it by hand. Each image, because you train the computer

7:12

and you say, this is an example of a cat.

7:17

So, Fai Fei Li and others. started collecting, scraping the

7:20

internet for all of these labeled

7:24

images, and they put together a

7:24

database they called ImageNet.

7:30

By 2010, they announced an annual

7:30

event, an annual competition, in which

7:37

AI researchers could submit their

7:37

program, specifically image recognition

7:42

programs, to compete with each other. Also, at that time, we had another very

7:44

important element, as I mentioned before,

7:50

the challenge for deep learning for many

7:50

years was the lack of computer power.

7:56

When you deal with lots and

7:56

lots of data, you need a lot

7:59

of computer power to analyze. A company by the name of NVIDIA, already

8:02

in 1993, was established to develop what

8:10

they called graphic Processing unit.

8:13

GPUs. Graphic processing unit, originally,

8:16

when NVIDIA was established,

8:21

were used for computer games.

8:23

So all of this came together, the GPUs,

8:23

big data, and deep learning, advanced

8:32

algorithm, to create a perfect storm.

8:37

A perfect storm that hit us in 2012.

8:47

It's interesting that the

8:47

computers that gamers are using, this

8:51

GPUs, they're nicer video cards, seem

8:51

to somehow advance our deep learning.

8:57

So it's so cool to see how

8:57

different channels of technology are

9:00

converging together during this time.

9:03

So when we come to 2010 and

9:03

we're thinking about ImageNet,

9:08

you mentioned that competition. Why was image and computer

9:11

vision harder to do?

9:16

Identifying images was hard to

9:16

do simply because other approaches

9:20

other than deep learning did not

9:20

use, for example, lots of data.

9:24

Deep learning itself had a challenge

9:24

in doing what it's doing best, which is

9:28

analyzing lots of data simply because

9:28

it didn't have enough computer power.

9:34

But when you put all the three together,

9:34

and that happens in 2012 in, in

9:40

various settings, it's bad the most. Impressive one, the one that influenced

9:42

the later developments was the ImageNet

9:48

competition at the end of 2012.

9:51

And in this competition, Geoffrey

9:51

Hinton and two of his PhD students

9:57

submitted their program, which was

9:57

a deep learning program, but based,

10:02

this is the first time, based on GPUs.

10:05

Which means basically they

10:05

processed whatever the images

10:09

from ImageNet very, very quickly.

10:12

And most important, it was more

10:12

accurate than all the other competitors.

10:17

Its error rate at that time,

10:17

at that particular competition,

10:21

was 15 percent error rate.

10:23

The second best program in that

10:23

competition had an error rate of 26%.

10:30

Using a different approach. Machine learning, but different approach.

10:33

So, um, at least the academic

10:33

world completely changed its mind

10:39

about the benefit of deep learning.

10:43

It was huge excitement

10:43

in the academic world.

10:46

Then the venture capitalists paid

10:46

attention and other investors.

10:51

The government started to pay attention. And we were off to the races of modern AI.

10:56

It makes me wonder what is

10:56

going on in academia right now that

11:00

I should be paying attention to, that

11:00

we'll hear about in a few years time

11:04

when investors make those big moves.

11:07

So as we think about that change

11:07

with ImageNet and these, these big

11:11

updates back in 2012, what did we

11:11

see the big industry companies do?

11:17

Google, I don't even know if

11:17

it was AWS or Amazon back then,

11:20

but what are the big names doing

11:20

with all this back at that time?

11:24

This is a very good question. And I actually omitted mentioning

11:25

big tech in my previous answer.

11:31

It was the venture capitalists, it

11:31

was academics, and most important,

11:35

actually, it was big tech, the Googles

11:35

of the world, because at that time

11:41

there was a very interesting shift from

11:41

a focus of this research in academia.

11:48

And for 70 years, the focus of, of

11:48

AI research was mostly in academia.

11:54

There were periods where there was

11:54

more of a business interest, especially

11:57

with expert systems, so called expert

11:57

systems in the seventies and eighties.

12:02

But mostly it was a, an academic focus.

12:06

But by 2012, we had Google as a

12:06

good example of a company that has

12:13

always, since its inception, almost

12:13

modeled itself after a university.

12:18

Publish or perish was a motto, not

12:18

just in academia, but also at Google.

12:25

They immediately, from the beginning, not

12:25

only invested a lot in research, of course

12:30

around search technology mostly, but also

12:30

about deploying information technology.

12:36

They did a lot of internal innovations.

12:38

So then you see, there

12:38

was already a huge shift.

12:41

And certainly around 2012, it was

12:41

established that to attract talent,

12:47

a company like Google, like Facebook,

12:47

like Amazon, pay a little bit less

12:52

attention even to the implementation,

12:52

to application, the business value of

12:56

that research, but also allowed them

12:56

to build their own reputation just

13:02

like academics do, by publishing. First of all, a lot of academics, the

13:05

ones who had already experience in

13:09

deep learning or other approaches to

13:09

AI, left academia and went to Google.

13:14

Yann LeCun was a good example for a

13:14

long time, being a tenured professor

13:20

at New York University, but he became

13:20

the chief AI scientist for Facebook.

13:26

Many of them just left academia

13:26

and went to work for a lot more

13:29

money for these big tech companies.

13:31

I'm starting to wonder

13:31

when we're getting to the

13:34

start of the modern AI boom.

13:36

So as we talk about 2012,

13:36

are we getting there yet?

13:40

The ImageNet competition of 2012

13:40

is the start of the modern AI boom.

13:47

The big, I think, event that made it more

13:47

accessible Was in 2016 when a program

13:57

from Google, its AI unit at the time,

13:57

DeepMind a program that managed to beat a

14:05

champion of go a game that was considered

14:05

to be a much, much bigger challenge for

14:11

AI programs than chess, for example.

14:15

But the AlphaGo program from Google gained

14:15

a lot of attention, a lot of headlines.

14:22

And I think this is where the

14:22

average newspaper reader found

14:28

out for the first time about this,

14:28

what was basically deep learning.

14:33

People stopped using, by that time, deep

14:33

learning as a label and started using AI.

14:45

Going back to 2014, a couple

14:45

years later, a man named Ian Goodfellow.

14:51

had a big idea that changed the AI game.

14:53

I'd love to hear more about that. What did he do?

14:55

What did neural networks talking

14:55

to each other allow AI to do?

15:02

So this is when we have the birth

15:02

of yet another buzzword, generative AI.

15:07

This is the first time that he

15:07

and some other people came up

15:10

with the idea that if you have two

15:10

models compete with each other.

15:16

that will help the program generate new

15:16

data, generate images, generate text,

15:26

generate eventually videos, and so on.

15:29

A very powerful idea that was taken by

15:29

a lot of other people, a lot of other

15:35

researchers, a lot of other companies. This was in 2014, but the real

15:37

big milestone, and maybe the last

15:43

milestone we have in terms of the

15:43

progress of thinking and developing

15:48

deep learning, what we now call AI. It came in 2017, and there are a

15:51

number of, uh, researchers at Google

15:58

published a paper titled Attention is

15:58

All You Need, in which they suggested

16:04

a new type of architecture, a new way

16:04

to design the deep learning model.

16:09

that allowed the program to

16:09

understand the context of

16:17

whatever was written in the text. Context meaning it could read, so to

16:19

speak, a whole paragraph and understand

16:26

the connection and the relations between

16:26

words far apart from one another.

16:31

Before that, it was really words one after

16:31

another that the program could analyze.

16:37

Now, with this new, actually simplified

16:37

architecture, This was the flourishing

16:42

of large language models started because

16:42

other big tech researchers, other

16:49

entities started competing with each

16:49

other by releasing better and better

16:55

large language models, more accurate.

16:58

Performing better and so on.

17:00

Is this the moment that AI

17:00

entered the mainstream with ChatGPT?

17:04

That's what it feels like to me.

17:06

Yeah. When ChatGPT was released in November

17:07

of 2022, it became within two months the

17:16

most popular consumer application ever.

17:18

I think a hundred million

17:18

users within two months.

17:21

It suddenly reached a very wide audience.

17:25

People got excited about it. People got upset about it.

17:29

Recently, we had surprising

17:29

development from a Chinese company,

17:36

Chinese AI company, DeepSeek.

17:38

It managed to do the same with much less

17:38

computer power and smaller amount of data.

17:46

So DeepSeek is a Chinese company.

17:49

It's also open source. So how does that impact the AI industry?

17:55

Is there controversy about it being

17:55

open source and what is the larger

17:59

business world doing with that?

18:00

Yeah. So open source is not something that Deep

18:01

Seek brought to the table a few years ago.

18:08

Facebook, which is investing

18:08

enormous amount of resources in

18:14

modern AI, made the decision to open

18:14

source everything that it's doing.

18:19

So open source is a very important

18:19

trend or approach to business within AI.

18:26

And it's been, before DeepSeek,

18:26

it's been, let's say, a threat.

18:31

to those companies like OpenAI that

18:31

keep the nature of the models, the

18:38

innards of the models to themselves.

18:41

When we look at the future,

18:41

though, what do you think we're going to

18:44

see in the next few years with AI models?

18:47

And how do you think that will

18:47

change the way that we do business?

18:51

As you know, prediction is

18:51

difficult, especially about the future.

18:55

I tend to focus on the history,

18:55

but history gives us some

18:57

indication of where are we going.

19:00

I do think that we will see steady and

19:00

slow improvement in the accuracy and

19:06

the performance of deep learning models.

19:08

And maybe even more important, we

19:08

will see steady and very, very slow

19:14

adoption by businesses of whatever

19:14

AI can do for them to save money

19:22

and to generate new revenue streams.

19:25

Well, that's the big

19:25

question right now is how businesses

19:28

can utilize generative AI to

19:28

bring ROI to their business.

19:32

So I'm excited to see how. The world changes with all of these

19:35

incredible technology changes.

19:40

And I appreciate having

19:40

you here on The Catalyst.

19:43

Thank you for having me.

19:47

That wraps up our journey

19:47

through the modern history of AI,

19:50

from its quiet resurgence in the

19:50

early 2000s, to the breakthroughs that

19:54

brought us today's AI driven world.

19:56

We've seen how deep learning,

19:56

neural networks, and large language

20:00

models have shaped industries and

20:00

transformed the way we work, create,

20:04

and interact with technology. But this is just the beginning.

20:07

AI is evolving faster than ever,

20:07

and the next big shift could

20:11

be right around the corner. Thank you to Gilpress for coming

20:14

on the show to share his insights.

20:17

This is the last episode of the season. We'll be taking a short break

20:19

and we'll be back soon with more

20:22

episodes on everything tech and AI.

20:25

For The Catalyst, I'm Heather Haskin.

20:29

The Catalyst is brought to you by

20:29

SoftChoice, a leading North American

20:32

technology solutions provider. It is written and produced by

20:34

Angela Cope, Philippe Dimas,

20:38

and Brayden Banks in partnership

20:38

with Pilgrim Content Marketing.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features