Future of Science and Technology Q&A (March 14, 2025)

Future of Science and Technology Q&A (March 14, 2025)

Released Wednesday, 19th March 2025
Good episode? Give it some love!
Future of Science and Technology Q&A (March 14, 2025)

Future of Science and Technology Q&A (March 14, 2025)

Future of Science and Technology Q&A (March 14, 2025)

Future of Science and Technology Q&A (March 14, 2025)

Wednesday, 19th March 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

You're listening to the Stephen

0:02

Wolfram podcast, an exploration of

0:04

thoughts and ideas from the

0:06

founder and CEO of Wolfram

0:08

Research, creator of Wolfram Alpha

0:10

and the Wolfram Language. In

0:12

this ongoing Q&A series, Stephen

0:14

answers questions from his live

0:16

stream audience about the future

0:18

of science and technology. This

0:20

session was originally broadcast on

0:22

March 14th, 2025. Let's have

0:24

a listen. Hello

0:26

everyone, welcome to another

0:28

episode of Q&A about

0:31

future of science and

0:33

technology. I see a

0:35

bunch of questions saved up

0:37

here. All right, one from a

0:40

non. What exactly is an

0:42

AI agent? I wonder if

0:44

that's an AI agent asking

0:47

that. Agentic. It seems like

0:49

nobody knows what those words

0:51

actually mean today. Yeah.

0:54

That is a problem, isn't

0:56

it? Let me tell you what

0:58

I have kind of, you know,

1:00

my version of what I think

1:02

they should mean or mean to

1:05

most people. The issue is

1:07

an LLLM, large language

1:09

model, its original

1:11

concept was something very

1:14

mundane. You'll have been

1:16

giving some text. follow

1:18

along with how that text, say well

1:21

how that text should continue, you know,

1:23

the cat sat on the blank, probably

1:25

the next word is Matt. Turns out

1:27

that particularly with reinforcement learning, it

1:30

was possible to get something which

1:32

has a prompt, like answer this

1:34

question, what is the, you know,

1:37

what color are most cats or

1:39

something, question mark, and then it

1:41

will go and give an answer,

1:43

and reinforcement learning has sort of...

1:46

been able to be successfully used to

1:48

get AIs to kind of go

1:50

and do the next, take the

1:52

next step, answer the question, do

1:54

the thing that you would expect

1:56

to do next, so to speak.

1:58

That's first thing. Second thing

2:01

is, a big thing starting

2:03

right at the very beginning

2:05

of the current sort of

2:07

LLLM revolution when Chatche BT

2:09

came on the scene was

2:11

this idea of calling tools

2:13

and we were very early

2:15

to that talking about with

2:17

the open AI guys about

2:19

calling Wolfram Alpha and then

2:21

calling Wolfram Language and that

2:23

was first deployed in what

2:25

March of 2023 I think. And

2:27

being able to have the LLLM call

2:30

tools that are external to the LLLM

2:32

in the case of Wolfram, Wolfram

2:35

language, Computation and Knowledge

2:37

Tools, but also we built

2:39

very quickly a way of having the

2:41

LLLM call back into your local

2:43

computer to do operations on

2:45

your local computer, to run Wolfram

2:48

language code that could do all

2:50

kinds of things, including delete your

2:52

files on your local computer. And

2:54

remember the moment it must have

2:57

been, I don't know, sometime in...

2:59

spring of 2023, when we had

3:01

this all set up. And I

3:03

was like, now wait a minute,

3:06

what am I doing here? You

3:08

know, but when I press this

3:10

button, we're going to put the

3:12

LLLM in charge of my computer.

3:15

It's going to be able to

3:17

agentically, you could say, go and

3:19

do things on my computer. It's

3:22

like, well, that maybe isn't such

3:24

a good idea until I've really

3:26

figured out how to make sandboxes,

3:28

and then I don't know if

3:30

the LLLM is going to be

3:33

able to figure out how to break

3:35

my sandbox and escape

3:37

and do things on my computer.

3:39

I don't want it to do.

3:41

But in a case, the next kind

3:43

of concept there is this idea

3:46

that LLLMs can sort of figure

3:48

out what to do next. That's

3:51

I suppose this kind of agentic behavior

3:53

is kind of the the ability to

3:55

go beyond sort of the continue the

3:57

text to being and what should I

3:59

do? next type thing to sort of

4:02

be something where it's like we feel

4:04

as humans that we have agency in

4:06

the sense that we can kind of

4:09

decide what we want to do and

4:11

then make it happen. And it's kind

4:13

of it's sort of an imitation of

4:16

that idea that we're talking about that

4:18

for things like our alarms. But a

4:20

big part of sort of grounding agents,

4:23

LLLM agents is and then they've actually

4:25

got to do something in the world,

4:27

not just ponder in their own minds,

4:30

so to speak, or even generate streams

4:32

of text. It's like, do something in

4:34

the world, you know, have agency in

4:37

the world, and that means calling tools

4:39

like Wolfram Language and Wolfram Alpha, and

4:41

in general, being able to operate things

4:44

in the world. You know, I think...

4:46

probably this year we'll see a lot

4:48

of progress in robotics in connection with

4:51

with AI and the sort of generalization

4:53

of large language models to sort of

4:55

physical behavior and I think there it

4:58

will be even clearer what sort of

5:00

agency means because it's like operate the

5:02

actuator move the arm you know pick

5:05

the thing up it's very clearly very

5:07

much like human agency and so on

5:09

in that case So I think that's

5:12

that's kind of the big part of

5:14

it. So for example, we have agent

5:16

API that we're that well actually we

5:19

have most pieces of it now and

5:21

soon the the thing under that name

5:23

will come out that is something intended

5:26

for agentic APIs to use to make

5:28

things happen computationally and all the things

5:30

that you can control computationally. That's a

5:33

well from language based system that is

5:35

kind of the the universal sort of.

5:37

tool for having agency in the computational

5:40

world, so to speak. That's one side

5:42

of the story. The other side of

5:44

the story is how do you build

5:47

a thing that does sort of agentic

5:49

things as an AI? And as I

5:51

mentioned, you know, part of that story

5:54

is, well, what will you do next?

5:56

of bigger modules

5:58

of what you do

6:01

next. Part of the

6:03

what will you do next story is

6:05

things like reasoning models where you're kind

6:07

of let's try this for a while,

6:09

oh that doesn't work, let's try something

6:11

else instead. Another version of

6:13

that is kind of a graph, kind of

6:15

an almost a flowchart, that sounds very

6:17

ancient but it's the same thing back again,

6:19

of sort of what do you do,

6:21

you try this for a while and then

6:23

if that doesn't work okay you go

6:25

to this branch, you try that, it's kind

6:27

of a structure. Now the way that

6:30

that usually works, it's worked with this way

6:32

with neural nets and so on, is

6:34

first you kind of define a fixed sort

6:36

of graph of do this, then this, then

6:38

this, a little bit like a fixed

6:40

flowchart. Later on that becomes more dynamic,

6:42

it's more like a piece of code

6:44

where it's kind of making decisions and

6:46

it's then running different pieces of code

6:48

and I really suspect the same thing

6:50

is going to happen here, in other

6:52

words that the harness for LLMs which

6:54

at the beginning was just like

6:56

produce the next token, produce the next

6:58

token, you know read what's there

7:00

from the prompt and what you have

7:02

already written and just keep producing the

7:04

next code and producing the next token.

7:07

The only thing that was really in

7:09

the harness there was which next token

7:11

should you produce because the typical output

7:13

of an LLM is here the probabilities

7:15

of a bunch of possible next tokens

7:17

and then the question sort of for

7:19

the harness was sort of what the

7:21

temperature should be as in are you

7:23

at temperature zero always picking the most

7:26

probable next token, are you at temperature

7:28

one picking with the probabilities that are

7:30

defined by the LLM, are you at

7:32

a different higher lower temperature sort of

7:34

picking with more randomness or less randomness

7:36

relative to the probabilities defined by the

7:38

LLM. That was sort of the earliest

7:41

version of the kind of harness that

7:43

you would put I suppose another harness, two

7:46

other pieces of the harness one is the stop

7:48

token, when do you stop going

7:50

yak yak yak as an LLM that's

7:52

another thing that's kind of comes from the harness not

7:54

from within the LLM and another thing

7:56

is tool calling at what point do

7:58

you say okay. I take the

8:00

characters, the tokens that were produced here

8:02

and this is just input to the

8:04

tool. It's not what I'm going to

8:07

show the user. Go send it to

8:09

the tool, the tool responds, then you

8:11

get back from that and you use

8:13

that as a further prompt. None of

8:15

this is visible directly to the user,

8:17

to the human user outside or the

8:19

outside user of the LLM that just

8:21

happens inside the LLM and that keeps

8:23

going. So I guess

8:25

that the sort of the

8:27

generalization of all that is make

8:29

a more complicated harness where one

8:32

thing that one expects in the future,

8:34

it's not quite there yet, is

8:37

mostly it's just sort of

8:40

adding a token at a time. It's

8:42

looking at the past. That's kind of

8:44

what the idea of transform on architecture is.

8:46

You're looking at the past, causally looking

8:48

at the past, so to speak, and saying,

8:50

so what do I add next? So

8:53

a very different kind

8:55

of technology is what's used in automated

8:58

theorem proving, for example, where

9:00

you are trying to say,

9:02

given these axioms, for example,

9:04

prove this theorem, and that

9:06

involves kind of finding a path by which

9:08

you apply this axiom and that one and

9:10

that one and that one and just the

9:12

right way to get from the axioms to

9:14

that proof. There are a huge number of

9:17

possible paths, most of which would lead to

9:19

completely different theorems, but there are a few

9:21

paths that lead to the theorem you want.

9:23

That's a kind of story of

9:25

path finding in this ultimately extremely

9:27

large and dynamically built graph of

9:29

all these possibilities. And

9:32

one can imagine in the same type of

9:34

thing being done for an LLM, where

9:36

you're saying, instead of just saying, well, take

9:38

the most probable next token, it's like,

9:40

well, tree out all these possible paths of

9:42

what might happen if you took this

9:44

token and this token and this token and

9:46

then sort of make a plan based

9:48

on where you're going to go in the

9:50

future to get a path that gets

9:52

you to where you want to go in

9:54

the future. I've sort of thought, actually,

9:56

for a couple of years now, that that's

9:58

what buzzwords being what they are. are those

10:00

will one day be called quantum LLLMs.

10:02

They're sort of doing the quantum thing

10:05

following many paths of history to decide,

10:07

in this case, to decide what to

10:09

do. We're not quite there yet on

10:11

the technology. But anyway, back to

10:14

the original question about sort of agents

10:16

and LLLMs and agentic behavior and so

10:18

on. I think the story there is

10:20

really this. this sort of bigger level

10:22

of planning of what are you going

10:24

to do next and maybe the planning

10:27

in a sense is just there's a

10:29

fixed flow chart you're following this branch

10:31

or that didn't work that's follow the

10:33

other branch instead and so on. I think

10:36

it's still an issue in motion so

10:38

to speak I think the part of

10:40

it that's very clear is the part

10:42

where you're calling something like our agent

10:44

API where you're going from the LLLM,

10:47

which is doing its LLLM kind of

10:49

things of making a good linguistic user

10:51

interface and so on, and you're going

10:53

into sort of the hard computational part

10:55

of things that's what we've implemented over

10:58

all these years and more from language,

11:00

and from all from language, it's sort

11:02

of, it's a great kind of universal

11:04

connector from sort of the symbolic specification

11:06

of things to all these different external

11:09

APIs and languages and systems and so

11:11

on, whether you're interacting with a blockchain

11:13

or a blockchain or a database. or

11:15

a piece of rust code or a

11:18

library or this or that, you know,

11:20

one of the things that's happened and

11:22

it's a result of the sort of

11:24

symbolic representation of everything in Wolfen language

11:26

is that we're able to provide this

11:29

really very good sort of coherent universal

11:31

connector. Sometimes what we're connecting

11:34

to is a bit of a jungle, but

11:36

you know, we provide this kind

11:38

of symbolic representation of what is

11:40

there, you know, how we can represent...

11:42

So at least one side of the

11:44

connection, you know, we're connecting to processes

11:47

on distributed computers or whatever. These are

11:49

all specified symbolically. Some of the things

11:51

that happen when you're launching processes on

11:53

those external computers, that's a jungle out

11:55

there. But at least we can specify

11:58

what's going on and get analysis. and

12:00

get results from what's going on

12:02

in this nice clean and coherent

12:04

symbolic way. And that's kind of

12:06

one of the things I use

12:08

quite a bit in both language.

12:10

I mean, I have my personal

12:12

setup, I think I have about

12:14

240 cores that are sort of

12:16

somewhat dotted around my house or

12:18

something. And when I'm doing big

12:20

computations, I will just run, you

12:22

know, parallel table, parallel map, whatever.

12:24

And it just sort of immediately

12:26

distributes the computation among the cores

12:28

that I have. We also have

12:30

remote batch submit and things for

12:32

sending batch jobs to sort of

12:34

external cloud providers and such like.

12:36

Let's see. All right, let's go

12:38

on to some other things here.

12:40

Layla asks, can you tell us

12:42

about the future of media and

12:44

information consumption? Will we become a

12:46

society of AI summaries as our

12:48

main form of information gathering? That's

12:51

an interesting question. I mean, I

12:53

think that, you know, it's funny,

12:55

from the time before Desktop Publishing,

12:57

I'm old enough that I remember

12:59

a time before Desktop Publishing, before

13:01

PowerPoint, before TED Talks, before these

13:03

kinds of things, there was a

13:05

certain difference in the way that

13:07

information was presented and absorbed. The

13:09

idea of kind of the bullet

13:11

point list. I think is an

13:13

idea that is, you know, I

13:15

don't think that really existed so

13:17

much when I was a kid,

13:19

for example. I think it was

13:21

much more narrative. And by the

13:23

way, the style of presentation that

13:25

people make has also changed. I

13:27

mean, in this, people used to

13:29

have a much less direct way

13:31

of expressing, explaining things, I think.

13:33

I think you know if you

13:35

look at sort of the history

13:37

of web expilations of things one

13:39

of the first kind of just

13:41

say it directly kinds of explainer

13:43

things was our math world website

13:45

talking about kind of the you

13:47

know all things math sort of

13:49

explained kind of directly. I would

13:51

say by the way that that

13:53

I also noticed you know when

13:55

I was writing the original documentation

13:58

for Mathematica and now Wolfman language

14:00

back in 1987 or so. I

14:02

also kind of adopted this kind

14:04

of tell it as it is

14:06

style of explaining things that was

14:08

visibly different from the stars that

14:10

existed before, for example, a computer

14:12

documentation, which had a habit of

14:14

being quite formulaic and not kind

14:16

of say it as you would

14:18

say it to a person type

14:20

thing. Math world had the same

14:22

approach of kind of say it

14:24

as you would say it to

14:26

a person. Wikipedia kind of followed

14:28

Math world. I know Jimmy Wales

14:30

made... use of ideas from math

14:32

world and they took a bunch

14:34

of content from math world. They're

14:36

in math world. It was kind

14:38

of, you know, as often happens

14:40

with these things, you know, Wikipedia

14:42

kind of sucked the content from

14:44

math world. I have to say,

14:46

it will be nicer. Wikipedia has

14:48

such more content from all the

14:50

things I write. We've kind of

14:52

made them. number of years ago

14:54

some some Wikipedia's were like you

14:56

should make this content Wikipedia friendly

14:58

and set it up so that

15:00

it has all the right to

15:02

you know letters after its name

15:05

at the bottom of the page

15:07

about you know the different kinds

15:09

of creative commons rights and so

15:11

on that it has so we

15:13

did that and I don't think

15:15

people have sucked as much content

15:17

from there onto Wikipedia as they

15:19

probably should have done and so

15:21

it's a good thing for people

15:23

to actually do I think that

15:25

But in any case, this style

15:27

that got adopted also by Wikipedia

15:29

of kind of the tele directly

15:31

as it is, is kind of

15:33

a style of presentation. Now, you

15:35

know, there's a yet different style

15:37

of presentation that the LLLMs are

15:39

giving us. It's a bit. land.

15:41

Sometimes I like it a lot

15:43

where it's kind of, you know,

15:45

we get these, we started doing

15:47

this back in early 2023, kind

15:49

of getting LLLM summaries of papers

15:51

about LLLM's that come out in

15:53

the world. And I find it

15:55

easier to read the couple of

15:57

sentence summaries of those things than

15:59

to read the original abstract because

16:01

they're very uniform, they're very uniform,

16:03

very easy to consume. So I

16:05

kind of, you know, I recently

16:07

have been taking doing more sort

16:09

of summarization of things that I'm

16:11

asking where I would be using

16:14

web search and I'm now using

16:16

kind of more. sort of LLLM

16:18

summarization. I have to say, I

16:20

think I mentioned this, I know

16:22

why it has come up a

16:24

couple of times, but I always

16:26

use this search engine Caggy, which

16:28

is this kind of search engine

16:30

sort of optimized for the users

16:32

rather than for the advertisers type

16:34

thing. I happen to have been

16:36

an advisor to that company for

16:38

a while, but Caggy has this

16:40

feature that I have to say,

16:42

only learnt from its CEO, fairly

16:44

fairly recently, fairly recently, where if

16:46

you, when you ask a. when

16:48

you put in a query, if

16:50

you end it with a question

16:52

mark, it will start its LLLM

16:54

running and start giving you a

16:56

summarization. And just the fact that

16:58

it's so easy to do it,

17:00

that you know, just doing it

17:02

by putting a question mark at

17:04

the end of the query, I've

17:06

started to find that I do

17:08

that a lot. And now, you

17:10

know, I've gone from probably one

17:12

in 50 of my web searches,

17:14

was a thing where I'm asking

17:16

for another number to more like

17:18

one in three or four. being

17:21

that way. And there definitely it

17:23

sort of changes my view of

17:25

what kinds of things I can

17:27

ask and it's sort of interesting

17:29

to see what happens and I

17:31

I was you know sometimes it's

17:33

like yeah I got the right

17:35

it got the right idea and

17:37

sometimes it's like no it careened

17:39

off into crazy land. Of course

17:41

there's I think the same thing

17:43

about web search sometimes that you

17:45

know while I was looking for

17:47

this thing but oh I didn't

17:49

realize that there was also some,

17:51

some other thing that had a

17:53

similar name and the website kind

17:55

of careened off into Crazy Land

17:57

as far as I was concerned.

17:59

You know, I do think that

18:01

this question of when, you know,

18:03

when do you do better with

18:05

the kind of bland LLLM summary,

18:07

when is it important to have

18:09

that kind of summarization of really

18:11

vast amounts of material out there

18:13

in the world versus, you know,

18:15

when are you better off looking

18:17

at the originals? You know, I

18:19

have to say, I do lots

18:21

of work in sort of history

18:23

of science and technology. And one

18:25

of the principles about doing that

18:27

is always read the original documents.

18:30

You know, you'd think, oh, there

18:32

have been these great historians and

18:34

they summarized what happened and this

18:36

and that and the other, and

18:38

you know, I'll read some of

18:40

that stuff. and then I'll say,

18:42

look, I just make it a

18:44

rule for myself. I got to

18:46

go read the original documents. And

18:48

it is the amount of additional

18:50

sort of flavor, color of what's

18:52

going on that you get from

18:54

those original documents, plus insights that

18:56

were kind of diluted by other

18:58

people trying to understand what was

19:00

said and not really quite getting

19:02

it right, and they're looking at

19:04

things through a certain lens. I've

19:06

always found it really, really valuable

19:08

to go back and read the

19:10

original documents. And maybe that's the

19:12

way one will feel about, do

19:14

I read the LLLM summary or

19:16

do I read the original document?

19:18

Do I read what so-and-so actually

19:20

wrote? Or do I read the

19:22

summary of what so-and-so actually wrote?

19:24

I mean, it's kind of a

19:26

little bit like in education, kind

19:28

of the great books theory of

19:30

just read what the original people

19:32

actually wrote. I'm quite a believer

19:34

in that. Maybe it's I'm prejudiced

19:37

from the fact that I write

19:39

a lot and I think some

19:41

of what I write is pretty

19:43

interesting and I hope people actually

19:45

read it because there's a lot

19:47

more. I know when I go

19:49

back and read things that I

19:51

wrote, there's a lot more detail

19:53

in there and a lot of

19:55

things that I had figured out

19:57

that weren't what I remembered. having

19:59

figured out, I only remembered kind

20:01

of the big picture. And I

20:03

know when people summarize, like, I

20:05

haven't read the summaries that people

20:07

have written of my stuff on

20:09

Wikipedia, I think I'd be pretty

20:11

horrified to do so. But I'm

20:13

kind of guessing that the summaries

20:15

are pretty distant from all of

20:17

the kind of things that are

20:19

really there in what, you know,

20:21

what even I figured out and

20:23

things I've written. So it's kind

20:25

of a thing where, where that's

20:27

a, you know, it's like... Are

20:29

you at an LLLM distance from

20:31

the original documents or are you

20:33

reading original documents? I think there

20:35

will be sort of additional value

20:37

in reading the original human written

20:39

documents, but there are other purposes

20:41

for which getting that planned LLLM

20:44

summary is going to be really

20:46

worthwhile. I don't know whether the

20:48

summaries will stay bland. It may

20:50

be that they'll be in a

20:52

good sort of, I don't know,

20:54

it's like, and waiting for the

20:56

LLLM that can write just like

20:58

me, it will come. You know,

21:00

I've put enough stuff out out

21:02

in the stuff out in the

21:04

I think from all this yacking

21:06

that I do in live streams

21:08

and things like that, there's maybe

21:10

50 million words of stuff for

21:12

me out there. And that's probably

21:14

enough to make a me bot

21:16

that will sound pretty much like

21:18

me. I don't know to what

21:20

extent the whatever edge I might

21:22

think I have of, you know,

21:24

the crisp new idea presented in

21:26

an interesting way. I don't know

21:28

what to at what point the

21:30

sort of the LLLM imitation will

21:32

come close enough. or maybe even

21:34

exceed what I'm able to do

21:36

there, that it will be sort

21:38

of, you might as well read

21:40

it, so to speak. I think

21:42

that there's sort of a question

21:44

of the economic ecosystem of kind

21:46

of material and what one consumes,

21:48

because you know, back in the

21:50

day, it's like things got written

21:53

and people who wanted to read

21:55

them paid for them with, you

21:57

know, magazine subscriptions or... or books

21:59

that they bought and so on.

22:01

And then kind of this. alternate

22:03

economy started up, which I suppose

22:05

had happened with radio and television

22:07

and so on, because, you know,

22:09

in the US at least, you

22:11

couldn't collect from people before cable

22:13

television. It's like you're broadcasting radio

22:15

out into the ether. Anybody can

22:17

pick it up. When I was

22:19

growing up in England, there was

22:21

the BBC, which was a government

22:23

sponsored, you know, still is, a

22:25

government-sponsored thing. And there was this

22:27

very strange concept of you had

22:29

to have a television license to

22:31

receive these television programs. And it's

22:33

a very strange kind of thing.

22:35

There was, I was a kid,

22:37

there were these television detector vans

22:39

that drove around the place trying

22:41

to detect... It must have been

22:43

the superheterodyne signals of the amplifiers

22:45

for televisions that were unlicensed televisions

22:47

and people would come in and

22:49

say you've got to shut down

22:51

that television and you know we're

22:53

going to find you because you

22:55

didn't pay your license fee which

22:57

was paying for the BBC to

23:00

produce the programming that people was

23:02

watching on their televisions. So it

23:04

was a little bit of a

23:06

different way of doing things, but

23:08

it was again a licensing a

23:10

model where the consumers of the

23:12

content were paying for the production

23:14

of the content. That kind of

23:16

disappeared, I think probably initially in

23:18

the US, when the idea came

23:20

about, well, just ran ads and

23:22

the advertisers will pay for the

23:24

creation of the content and, you

23:26

know, the soap operas were advertising

23:28

soap and the etc, etc, etc,

23:30

etc, etc. It's some. And that

23:32

was sort of, that has been

23:34

kind of an alternate model of

23:36

how to pay for the production

23:38

of media. And then, you know,

23:40

when the web came along, the

23:42

web had the sort of anomaly

23:44

that the original concept of the

23:46

web and things like Project Xanadu,

23:48

which was a predecessor kind of

23:50

the hypertext invention that turned into

23:52

the HTCP, you know, the H.T.

23:54

that Tim Berners Lee put in

23:56

the in the names of websites.

23:58

that I think came from Project

24:00

Xanadu and Ted Nelson and all

24:02

these guys who were thinking about

24:04

sort of how to make this

24:07

ecosystem of information of connected information.

24:09

I think they originally imagined that

24:11

people would pay for the information

24:13

they got, that there would be

24:15

a whole ecosystem of micropayments where

24:17

it's like, well, if you got

24:19

this thing from here, you would,

24:21

the person who was the original

24:23

producer of that, would get, you

24:25

know, a millionth of a penny

24:27

for the fact that you looked

24:29

at that thing. And this would

24:31

all get sort of aggregated, sort

24:33

of aggregated up. to pay for

24:35

the production of content. For whatever

24:37

reason, and it was partly because

24:39

of the origination of the web

24:41

and the fact that it was

24:43

sort of based on the internet,

24:45

which was based on the ARPANET,

24:47

which was this sort of free-ish

24:49

government thing in the US, and

24:51

just the way that that whole

24:53

thing developed, micropayments never happened on

24:55

the web, as in that way.

24:57

And so, sort of there had

24:59

to be another model for how

25:01

would you pay for content on

25:03

the web? And the fundamental model,

25:05

which I have to say I

25:07

was amazed it worked, that, you

25:09

know, Google basically pioneered was, I

25:11

mean, Yahoo had done things earlier

25:13

and so on, but it was,

25:16

you know, I just couldn't believe

25:18

this was going to work. And

25:20

it wasn't the original Google business

25:22

model, but, you know, that you

25:24

would make money by basically selling

25:26

advertising against web content. And you

25:28

know that has led to many

25:30

things that have been both well

25:32

some good and many probably not

25:34

so great about I mean it

25:36

sort of allowed the web to

25:38

be more widely in the content

25:40

on it to mean war widely

25:42

disseminated. It's meant that the you

25:44

know the people who pay and

25:46

not the people. who are getting

25:48

the benefits, so to speak, and

25:50

I never like those kind of

25:52

mismatches, something in the business we've

25:54

done with our company. I've always

25:56

tried to avoid that, tried to

25:58

make it be the case that

26:00

the folks who are paying for

26:02

our software and the services we

26:04

make and so on are the

26:06

folks who are getting the benefit,

26:08

rather than... sort of a third-party

26:10

mismatch, so to speak. And I

26:12

think that's another aspect of kind

26:14

of the consumption of media is,

26:16

well, you know, who's paying for

26:18

the production of this media content?

26:20

And, you know, what we've seen

26:23

now with social media and with

26:25

Twitter and X and so on,

26:27

is that it's, you know, there's

26:29

an awful lot of content that

26:31

is produced by people just producing

26:33

it because they want to make

26:35

a point. They want to put

26:37

that content out there. It's very

26:39

different from the model that had

26:41

existed before, where there's a stratum

26:43

of professional journalists whose job is

26:45

to forge the world for content

26:47

to put it out there. It's

26:49

a thing where people who want

26:51

to get content out there, you

26:53

know, push to get content out

26:55

there, so to speak. And then

26:57

sort of it becomes an issue

26:59

of how do you select the

27:01

content you want, and so on.

27:03

But it's a different ecosystem of

27:05

sort of content production. Well now

27:07

we've got the LLLMs coming along,

27:09

which are yet something different. They're

27:11

foraging all that human content and

27:13

sort of making it their own.

27:15

in a way that is now

27:17

quite disconnected from the original sources

27:19

of content. I mean, there's some

27:21

efforts to put references into the

27:23

LLM content. That works in some

27:25

cases and so on. But I

27:27

think as we see more sort

27:29

of LLM-made content out there, it's

27:32

again unclear how that's going to

27:34

shake out economically. I mean, it's,

27:36

you know, right now, there's a...

27:38

a big market for training data.

27:40

It's something, you know, we are

27:42

a source of quite a bit

27:44

of training data that's used out

27:46

there in the world, of things

27:48

that are very, in a sense,

27:50

very clean training data because we

27:52

are computing things. So we're generating

27:54

training data. It's not that, I

27:56

mean, it takes a bunch of

27:58

human effort to generate the right

28:00

thing, but it is not something

28:02

where we're saying, well, it's so-and-so's

28:04

writing that was done whenever, and

28:06

now we're foraging that. very clean

28:08

training data. And it's also training

28:10

data that has the interesting feature

28:12

that's kind of training for reasoning

28:14

in effect. Just like, you know,

28:16

I think the elements from the

28:18

beginning read all this text and

28:20

discovered logic from the text by

28:22

virtue of the fact that there

28:24

were sort of sentences that were

28:26

constructed in this way, which are

28:28

in a sense, the logical sentences.

28:30

And now... you know, there's a

28:32

lot that we have that is

28:34

computation of mathematics and chemistry and

28:36

all those kinds of things that

28:39

is somehow kind of put together

28:41

in a sort of reasoning way.

28:43

So that's been another, the, another,

28:45

sort of piece of that story.

28:47

And I don't know how, how,

28:49

you know, I think The thing

28:51

that's happened is because of the

28:53

ecosystem, the way the economics of

28:55

content generation have worked, there's been

28:57

sort of a lessening, you know,

28:59

the journalism world has sort of

29:01

been going down, I'm sort of

29:03

horrified usually with the kinds of

29:05

things that I'll see in a

29:07

lot of... beyond the very top

29:09

tier of journalism, the kinds of

29:11

things that I see. They're very

29:13

lazy, sloppy, repetitive, secondary kinds of

29:15

things. I think one of the

29:17

things I know, I know Kagi

29:19

uses as a heuristic for its

29:21

ranking for search is how many

29:23

ad trackers are there on that

29:25

page? How much of it is

29:27

sort of original content and how

29:29

much of it is just, you're

29:31

only here because we want to

29:33

get you to look at the

29:35

ads, because that's how we're monetizing

29:37

our page. And I think that's

29:39

now, you know, it's starting. There

29:41

are all these other different channels

29:43

where people are sort of putting

29:46

content out there, just in, you

29:48

know, the things they write in

29:50

blogs and other things like that.

29:52

And I obviously put a lot

29:54

of content out there in that

29:56

form. And it's it's it's kind

29:58

of just stuff I want to

30:00

put out there. It's not because

30:02

I have a living being a

30:04

journalist or a writer of science

30:06

books or something like that. It's

30:08

just stuff that I'm writing because

30:10

I like to write it and

30:12

I think people will find it

30:14

interesting and I'm putting it out

30:16

there. It's a different sort of

30:18

and you know it's not it's

30:20

not how I'm making a living

30:22

so to speak and I think

30:24

the same is true of lots

30:26

and lots of content on X

30:28

and so on and similar kinds

30:30

of kinds of places. So you

30:32

know and I think it... this

30:34

question, you know, the elements are

30:36

a weird twist to the whole

30:38

thing because they are a yet

30:40

different part of the story than

30:42

not at this time. They'd been

30:44

monetized essentially by subscription, which as

30:46

far as I'm concerned is a

30:48

very clean form of monetization. You

30:50

know, maybe one day they'll be

30:52

monetized with ads. I kind of

30:55

hope not because then it's going

30:57

to get even more bizarre as,

30:59

you know, you're reading this element.

31:01

at least in search ads, for

31:03

example, it's an ad, you know,

31:05

it's an ad. By the time

31:07

you're reading LLM content is all

31:09

interwoven with, you know, kind of

31:11

prompts that say, now there's a

31:13

chance to advertise this brand of

31:15

toothpaste or something, and it starts

31:17

doing that. It's going to get

31:19

very confusing. And I, you know,

31:21

and then you'll have to have,

31:23

it's like, well, I want to

31:25

read this thing that doesn't have

31:27

ad contents. I'm going to get

31:29

my own lalam to try and

31:31

guess what's ad content and back

31:33

that out. Just like, you know,

31:35

we're getting lalams to try and

31:37

figure out the things we're sent,

31:39

we're written by lalams, well, it's

31:41

usually pretty obvious at this point.

31:43

And, you know, we're, we're kind

31:45

of like. For me, it's like,

31:47

I don't want to read what

31:49

some LLLM wrote. If you as

31:51

a human want to communicate with

31:53

me, you had some set of

31:55

bullet points, some set of prompts

31:57

that you fed to the LLLM.

31:59

Just give me those prompts. You

32:02

know, I'll deal with the prompts,

32:04

so to speak. I don't want

32:06

to see the LLLM Fluff that

32:08

came out that's many pages long

32:10

that I have to kind of

32:12

grind down. Maybe I have to use

32:14

my own LLLM to do that to see

32:16

what it is to actually trying

32:18

to say. Let's see. A bunch of

32:21

questions about LLLM. So let me

32:23

address those and then go on to

32:25

some other things. Yeah, desk comments

32:27

before AI summaries there were

32:29

encyclopedias and textbooks and cliff

32:32

notes and so on, which were

32:34

useful and convenient. They

32:37

never became de facto. So, you

32:39

know, that's right. I mean, starting

32:41

in the 1700s, the idea of

32:43

encyclopedias with people like Dalomber and

32:45

France and then later encyclopedia Britannica

32:48

or in Scotland were kind of

32:50

like, collect this information, collect. you

32:52

know, the things you need to

32:55

know in the world in a

32:57

digested way. I mean, then it

32:59

was readers digest, then it was

33:02

close notes, then it was all

33:04

these kinds of things that sort

33:06

of a digest of information. I

33:09

agree. I think that that maybe that's

33:11

that maybe that's the way to think

33:13

about the current role of LLLM's

33:15

is as digesters of information.

33:17

I mean, they go beyond that because

33:20

you can kind of talk to the

33:22

book type thing. And I think that's

33:24

a That's a thing that is

33:26

probably going to be increasingly

33:28

useful. I mean, we're currently

33:31

doing a big experiment of

33:33

building an AI tutor, specifically

33:35

for algebra one course. It's an

33:37

attempt to make something which people

33:40

have never been able to make

33:42

work before, which is to have

33:44

a truly autonomous, truly scalable educational

33:47

tutoring system. People, you know,

33:49

there's 70 years of history of

33:51

trying to use computers to be

33:53

teaching machines. It didn't really work

33:56

that well. It's worked great for

33:58

things like Wolfram Alpha, where you're...

34:00

using it as a generator of content

34:02

that is alongside what you're learning, but

34:04

it's not something where the dynamics of

34:07

actually teaching are delegated to the machine.

34:09

What we're trying to do with our

34:11

AI tutor is get a surprisingly complicated

34:13

system that tries to delegate the dynamics

34:16

of teaching to the machine, at least

34:18

for a topic like algebra, which is

34:20

in some level a kind of very

34:23

cut and dried topic. We'll see how

34:25

it works. It's looking pretty promising right

34:27

now, but that's a case where you

34:30

can in a sense expect. Now that

34:32

particular one was done with lots of

34:34

efforts, specifically building an algebra course that

34:36

fitted into the kind of the way

34:39

that one could control it with another

34:41

lamb and so on. I think we

34:43

also have another project that is really

34:46

taking much more off-the-shelf sort of book-like

34:48

content and saying make something where you

34:50

can kind of talk to the book

34:53

and... You know, I think that's a

34:55

that's kind of a new direction for

34:57

quote summarization is well, it can be

34:59

a summary just for you a personalized

35:02

summary. I mean, I suppose I suppose

35:04

that is probably the direction that one

35:06

can expect to go in that your

35:09

LML gets to know you pretty well

35:11

and or your AI gets to know

35:13

you pretty well and it knows the

35:16

thing you don't understand and the one

35:18

thing that should be said to explain

35:20

this to you is this particular thing

35:22

and it will be very different for

35:25

somebody else. I mean, I think it's

35:27

an interesting dynamic as our AIs get

35:29

to know us better and better. You

35:32

know, our feelings about our AIs will

35:34

probably change. You know, they'll probably come

35:36

a time when there are AIs that

35:39

know me better than any human knows

35:41

me. And maybe it's already the case,

35:43

although in a very banal sense of

35:45

an A. I mean, I have a

35:48

pretty well-organized meta-seacher of all of the

35:50

sort of interactions and things I've done

35:52

over the course of the last 30-something

35:55

you know that in a sense that

35:57

system which is not really an AI

35:59

system. becoming more AI-ish over time, but

36:02

sort of knows me very well. There

36:04

will probably come a time when an

36:06

AI that I can routinely interact with

36:08

as I interact with people knows me

36:11

better than any person knows me, in

36:13

the sense that they know my whole

36:15

history of every interaction I've had with

36:18

everybody, and they've sort of seen more

36:20

of the details of what I do

36:22

and say and so on than anybody

36:25

else has. And I think my feelings

36:27

about that AI will be interesting. I

36:29

mean, I think I will be very

36:31

protected probably of that AI. It's a,

36:34

you know, it's a very, it's a

36:36

very sort of close companion AI, so

36:38

to speak, and that will be an

36:41

interesting dynamic to see. Let's see. The

36:43

question from the... When will we get

36:45

the first AI robot news reporter? I

36:48

see these being useful in cases of

36:50

dangerous live broadcasting like hurricanes to keep

36:52

people up to date. That's a fun

36:54

one. I mean, I think there is

36:57

a lot of AI reporting. I mean,

36:59

that's happened for a decade or more,

37:01

whether it's the sports scores or the

37:04

financial news. There's, you know, there are

37:06

people who will read, you know, and

37:08

the Tao is up X number of

37:11

points or down X number of points

37:13

or whatever, and the this and that

37:15

and the other. But, you know, you

37:17

know, starting, probably 15 years ago, 15

37:20

years ago or more. that started to

37:22

be systems that could produce, could synthesize

37:24

good natural language text from the kind

37:27

of the raw numbers of what was

37:29

going on in the stock market or

37:31

wherever else. And I'm pretty sure, I

37:34

mean, I have to admit I'm not

37:36

a consumer of these things, but I'm

37:38

pretty sure that a lot of those

37:40

things up and running right now. Now

37:43

I think another question about the news

37:45

is perhaps, there's news that is directly

37:47

absorbable by people. from the raw data.

37:50

Like if you're watching, I don't know

37:52

what, I don't know, something about the

37:54

stock market or something, the raw numbers

37:57

tell their story, or you're doing... But

37:59

now, there are cases where there's lots

38:01

of stuff happening in the world, but

38:03

most of what's happening, you just don't

38:06

want to know about it's completely boring.

38:08

It's like every security camera that has

38:10

footage of this and that happening. But

38:13

occasionally, there'll be a zebra that escaped

38:15

from the zoo and is prancing through

38:17

the security camera field of view, and

38:20

that's the thing you might want to

38:22

know about. But, you know, most of

38:24

the time, it's just looking in an

38:26

empty street or whatever. I can imagine

38:29

a time when all this data that's

38:31

being collected in the world is being

38:33

more effectively aggregated by AIs to tell

38:36

us, the humans, things that we might

38:38

care about that are going on in

38:40

the world. I think that hasn't yet

38:43

happened. I mean, I've been involved in

38:45

a number of projects actually over the

38:47

years that have sort of gone towards

38:49

that kind of direction. not only in

38:52

that case saying what is happening in

38:54

the world, but based on what is

38:56

happening in the world, what's going to

38:59

happen next? What kinds of, you know,

39:01

what inevitable consequences are there based on

39:03

the model that we have of the

39:06

world, whether it's model of meteorology or

39:08

models of crops or models of supply

39:10

chains, you know, given that this happened,

39:12

what will inevitably happen next in the

39:15

world? And that's something that's... that's relevant,

39:17

I mean, purely in terms of, you

39:19

know, things like financial speculation, that's relevant,

39:22

but there are lots of other reasons

39:24

why that's a relevant thing to be

39:26

able to do. But it's an interesting

39:29

point that I haven't really seen this

39:31

yet, is the really the systematic aggregation

39:33

of what is otherwise a boring kind

39:35

of, you know, a most of the

39:38

time the camera is going to see

39:40

nothing type of type of information. Now,

39:42

obviously X. has a big, has lots

39:45

of streams of stuff that are coming

39:47

along, but those already quite human curated.

39:49

Even the curation of those streams is

39:52

not something that's really happened yet that

39:54

much. I mean, I'm kind of waiting

39:56

for the first sort of X-based newspaper.

39:58

I think actual newspapers these days... take

40:01

a lot of their content from X.

40:03

And, you know, it typically happens first

40:05

there and only later in the newspapers.

40:08

But, you know, I think that's an

40:10

interesting question is sort of an interesting

40:12

almost philosophical question. There's all this stuff

40:15

going on in the world. What's news?

40:17

You know. it's the New York Times

40:19

what you know what the all the

40:21

news that's fit to print or something

40:24

I don't know whether that's a tagline

40:26

they even still use or certainly whether

40:28

it applies really in any serious way

40:31

to that newspaper I'm not sure but

40:33

but I mean there's this question of

40:35

all the things that are going on

40:38

in the world that are picked up

40:40

by all our sensor arrays and all

40:42

this kind of thing you know what's

40:44

what's worth summarizing for a particular human

40:47

and you know obviously one can pick

40:49

particular sections of news, one's interested in

40:51

things like that. I would say that

40:54

the whole idea of personalized news that

40:56

was, I don't know, that's come a

40:58

bunch of times. People have used Wolfmaufer

41:00

a bunch to do various kinds of

41:03

personalized news and some of our natural

41:05

language understanding technology do those kinds of

41:07

things. I don't think that's really become

41:10

a big thing yet. And maybe with

41:12

elements it will. Let's see. Oh yeah,

41:14

the question was about things like broadcasting

41:17

from the hurricane scene and so on.

41:19

Yeah, I mean, I think the one

41:21

of the questions is, you know, it's

41:23

like you're seeing a bunch of footage,

41:26

you know, there are a bunch of

41:28

security cameras being ripped up by the

41:30

hurricane or something. And, you know, it's

41:33

like there could be a human meteorologist

41:35

that interpreting what they see, or it

41:37

could be something where there's a layer

41:40

of... you know, of AI that perhaps

41:42

knows more than us humans know about

41:44

how to interpret that weird picture of

41:46

the thing that was, you know, I

41:49

don't know what happens, you know, that

41:51

blacks out at that moment or whatever

41:53

else it is. But I think, yeah,

41:56

I mean, there are a number of

41:58

these things where I feel like there's

42:00

still some place. to go in terms

42:03

of what people get used to and

42:05

so on. I mean, you know, another

42:07

one of these that I'm surprised hasn't

42:09

caught on more is virtual touristing of,

42:12

you know, you get, I mean, maybe

42:14

when there are better humanoid robots and

42:16

things, that'll be more of a thing.

42:19

It's like, I'm not going to go

42:21

to Antarctica, but I'm going to send

42:23

my robot to Antarctica, and it's going

42:26

to play with the penguins, and... I'm

42:28

going to have, you know, it'll be

42:30

really cool to watch me controlling my

42:32

robot playing with the penguins type thing,

42:35

or whatever it is. I don't know

42:37

if penguins are very friendly or not.

42:39

It's always sometimes suspicious when animals look

42:42

really cute and then when you get

42:44

up close, they're pretty aggressive. I don't

42:46

know in the case of penguins. But

42:49

in any case, I mean, you know,

42:51

I can imagine that being a thing

42:53

that develops there. Let's see. It's

42:56

a question here from LC, how

42:58

far are we from LLLM's generating

43:00

the kind of things that I

43:02

write with sort of similar elucidation

43:04

based on a short prompt? I

43:06

don't know. It feels like we're still

43:08

pretty far away. It feels like,

43:10

I mean I think I'm going to

43:13

know more I'm doing a project

43:15

just starting a project right now,

43:17

kind of using LLLM's as a way

43:19

to kind of understand a very

43:21

broad literature about physics about physics. for

43:23

understanding experimental implications of our physics

43:25

project. I don't know. I'm going

43:27

to know more after I really do

43:30

this. I mean, I've been a

43:32

big believer forever in using the best

43:34

tools one can for doing what

43:36

we want to do. And it's

43:38

clear now that sort of this summarization

43:40

of broad swaths of human output

43:42

is something which the LLLM's are good

43:45

at. And I want to see

43:47

how far I can get with

43:49

that. and to what extent it can

43:51

kind of, you know, to what

43:53

extent it can organize and write kinds

43:55

of things that I'm expecting to

43:57

use it mostly as a tool.

43:59

a foraging tool so to speak more

44:02

than a tool of exposition. We're

44:04

not there yet. I don't know how

44:06

that future will unfold. Wailo

44:08

is commenting when you say that

44:10

teaching is delegated to the machine.

44:12

Are you saying that the machine

44:14

is telling the student what to

44:17

think instead of just answering

44:19

questions? That's a good interesting spin.

44:21

I mean when it comes

44:23

to Wolfram Alpha for example,

44:25

we're just answering questions.

44:28

The... the kind of which questions to

44:30

ask that's all on the

44:32

student the which way should we go

44:34

it's all on the human I mean

44:36

I think the long-term story of AI

44:38

is kind of the which way should we

44:40

go is up to humans to

44:43

define the let's get there as

44:45

as automatically as possible that's for

44:47

the AIs to do you know I

44:49

was just realizing that you know

44:52

when people ask you know what's going

44:54

to happen when AIs are doing all

44:56

these things you know, how do you

44:58

feel about that as a human? And

45:00

I was realizing, at some level, I've

45:02

kind of lived the AI dream for

45:05

decades, because I've been building technology

45:07

that automates the things that I

45:09

mechanically want to do, that involve,

45:11

you know, figuring out things in

45:14

science and technology and so on.

45:16

And, you know, with Wolfman language, I've

45:18

been kind of building the AI

45:20

dream for myself and for lots of

45:22

other people, because it's all about

45:24

going from I imagine doing doing this

45:27

thing. to as automatically as possible,

45:29

the thing gets done. And that's

45:31

kind of the whole idea of

45:33

all from language is I imagine

45:35

the thing, I type it with my fingers

45:38

or even you can say it to

45:40

some extent now, and then, you know, and

45:42

then it's dealing with the mechanics

45:44

of how to get that question

45:46

answered, so to speak. I'm still

45:48

defining the question, it's getting that

45:51

question answered as efficiently as possible.

45:53

And that is kind of the

45:55

AI dream, I think, is we're

45:57

in charge, we're calling the shots.

45:59

and the AIs are efficiently executing

46:01

on what we want to have done.

46:04

Now, you know, there's a scenario where

46:06

the AIs are deciding what to do

46:08

for themselves, I think is one of

46:11

these sort of philosophically doomed scenarios, because

46:13

it's something where you say, well, the

46:15

AIs are going to go off and

46:18

do what they want to do. Sure,

46:20

computers do that all the time. It's

46:22

like, you know, you've set them off

46:25

in this general direction. The way they

46:27

actually do it is this. I've spent

46:29

a lot of time kind of defining

46:32

simple rules. sort of computational rules and

46:34

just seeing what those rules do, then

46:36

it's off on its own. It's not

46:39

under human control. It's not, oh, I

46:41

want to get it to do this.

46:43

I want to have this goal for

46:46

it. It's just doing what it's doing.

46:48

And I think the thing to understand

46:50

is that the big example of that

46:53

happening that we're all very familiar with

46:55

is the natural world, where things are

46:57

happening, all the time. But they're not

47:00

things under our control. They're not things

47:02

where we know the goal, the purpose

47:04

of what's happening. They're just things that

47:07

are happening. And I think that's what

47:09

sort of AI is left to their

47:11

own devices. That's almost by definition, what

47:14

they end up doing is like what

47:16

the natural world does. They just do

47:18

what they do based on the computational

47:20

rules that they have. So, but you

47:23

know, the things that I suppose my

47:25

kind of use of sort of AI

47:27

and AI like technology. It's kind of

47:30

living this AI dream of, I imagine

47:32

what I want to do, the AI

47:34

gets it done. And I think that's

47:37

sort of the well-defined version of how

47:39

that works out. Let's see. So the

47:41

question was about telling students what to

47:44

think as well as how to answer

47:46

the question. So I mean, I agree

47:48

that there's an issue of, you know,

47:51

what is teaching? What's that supposed to

47:53

be about? It's supposed to be about

47:55

sort of inducing certain patterns of thinking

47:58

and certain remembered facts. in the student.

48:00

And yes, it is true, and it's

48:02

an interesting point, that, you know, it's

48:05

the think like an AI, you know,

48:07

are we really going in that direction?

48:09

There is a certain tendency, and, you

48:12

know, even right like an AI, you

48:14

know, sometimes I'm wondering whether, you know,

48:16

the fact that the first generation of

48:19

LLLams was very into the word delve.

48:21

And then I start seeing humans writing

48:23

and using the word delve a lot.

48:26

I don't think I've ever written the

48:28

word delve. I don't know. I can

48:30

check because I have archives of everything

48:33

I've ever written. So I could easily

48:35

check how many times I've written the

48:37

world delve in my life. I'm guessing

48:40

it's very very very small. And I

48:42

haven't started now. But for many people,

48:44

that was a rare word. but then

48:46

the LLLM started using it and then

48:49

the humans started using it, copying the

48:51

LLLM's. And yes, I think that's an

48:53

interesting issue. If it gets to the

48:56

point where sort of one's teaching the

48:58

human and what one is trying to

49:00

do is to model the way that

49:03

humans think about things and the way

49:05

that an expert human might think about

49:07

this and model that for a human

49:10

who is just learning, but yes, there

49:12

is a certain tendency to say. Actually,

49:14

it will turn out to be the

49:17

way the AI thinks about these things,

49:19

and we're teaching the human to be

49:21

a bit like an AI. You know,

49:24

it makes me think of all the

49:26

things that I've automated in my life,

49:28

and the things where it's like, oh,

49:31

I'm thinking of my older daughter who

49:33

is now a mathematician, when she was

49:35

younger, you know, she would do things

49:38

like calculus by hand, and I would

49:40

sort of make... silly annoying parent comments

49:42

like I didn't know anybody still did

49:45

that kind of thing by hand. You

49:47

know, I had been doing that. I

49:49

haven't done that by hand myself in

49:52

25 years that time, 30 years, whatever

49:54

it was. And because I've always been

49:56

using a computer, but it's kind of

49:59

like what there's kind of this, this.

50:01

well, what should you delegate to the

50:03

machine versus what should you do yourself?

50:06

How should you think about things? And

50:08

I kind of feel like the big,

50:10

the big thing that us humans have

50:13

to contribute is sort of what questions

50:15

to ask, what direction do we want

50:17

to go? The, the how to get

50:19

the question answered, like how to get

50:22

the calculus problem solved is much more

50:24

in the domain of the machines. And

50:26

so the question of education in the

50:29

end. is much more how do you

50:31

teach humans to sort of think about

50:33

things that that and make decisions about

50:36

things that haven't been decided before. And

50:38

I tend to think that although people

50:40

might imagine that that's all a story

50:43

of just thinking thinking about the learning

50:45

about the dynamics of thinking itself, my

50:47

personal experience and my observation of other

50:50

things is that knowing a ton of

50:52

facts is super important in being able

50:54

to actually think clearly. It's the facts

50:57

are the bedrock. on which you can

50:59

build this kind of these layers of

51:01

sort of abstract thinking. And if you

51:04

try and just sort of teach the

51:06

abstract thinking in the abstract, it just

51:08

doesn't work. It's something, by the way,

51:11

I think we learned this a little

51:13

bit from the LLLMs. The fact that

51:15

the LMs can do a certain amount

51:18

of sort of common-sensish thinking lives on

51:20

top of the fact that they were

51:22

taught a zillion effectively facts about the

51:25

world. that's where we abstract to be

51:27

able to do thinking from. And so

51:29

in education it's really important, I think,

51:32

to learn a bunch of facts. That's

51:34

the bedrock on which we can actually

51:36

sort of construct this kind of layer

51:39

of abstract thinking that is then probably

51:41

ultimately the thing that we can contribute

51:43

the most to sort of what happens

51:46

in the world is kind of the

51:48

thinking about what... to do next what

51:50

we care about. Those are, in a

51:52

sense, human choices and being able to

51:55

do a good job as humans of

51:57

thinking through those things, I think is

51:59

an important thing to be teaching in

52:02

education. And there's sort of a question

52:04

of can we teach. how to think

52:06

like humans were their eyes. I don't

52:09

know the answer to that yet. Possibly

52:11

yes. I mean, it's, you know, the

52:13

other thing to realize is, for example,

52:16

the way that, I don't know, Wolfen

52:18

language computes some math thing, like calculus,

52:20

is deeply nonhuman. I mean, if you

52:23

know, if you know how it works

52:25

inside, it's a wonderful algorithmically clever industrial

52:27

machine for just grinding out the answer.

52:30

And when we were building step-by-step functionality

52:32

in Wolfmalfa, it's interesting how much of

52:34

a fake it is, in the sense

52:37

that it's very useful for the step-by-step

52:39

stuff to know what the answer is

52:41

going to be. That answer was found

52:44

by industrial methods that are deeply nonhuman.

52:46

When we build the step-by-step explanation, that

52:48

is something that we're building modeling the

52:51

way that humans think about things. And

52:53

it's kind of a retrofitted model of

52:55

the way that humans think about things.

52:58

And yes, it's even if there's a

53:00

situation where the computational way to work

53:02

it out is something deeply nonhuman, one

53:05

could certainly imagine sort of backing out

53:07

what the human way to think about

53:09

that is and being able to help

53:12

people understand how to think in those

53:14

human ways. Teaching people to think like

53:16

computers is not going to work. We

53:18

just don't, our brains are not capable

53:21

of that. I think that the, you

53:23

know, if I say, can I run

53:25

this piece of code in my brain,

53:28

I cannot. If I say, do I

53:30

have intuition about what this code is

53:32

roughly going to do, the answer is

53:35

absolutely. I can get that intuition. It's

53:37

a kind of a vague, abstracted intuition

53:39

about things. But if you say, can

53:42

I actually run this code and say,

53:44

well, I'm going to get three or

53:46

seven out of it, the answer is

53:49

no chance. And I don't think that's

53:51

just my inadequacy, so to speak. I

53:53

think that's just not the way the

53:56

100 billion neurons that are in our

53:58

brains are built to operate. in a

54:00

microprocessor are taught to operate, at least

54:03

when you have layers of things like

54:05

our computation. language on top, but it's

54:07

just a different kind of thing than

54:10

what brains do. Let's see. Maybe this

54:12

is a different question. Some of the

54:14

previous one here. Can a sentient AI

54:17

understand how humans learn if we would

54:19

delegate them the teaching of human kids?

54:21

Would that be compatible with a biological

54:24

point of view? I mean, I think

54:26

that... You know, there's this

54:28

fundamental question of, of can we, can

54:30

we understand psychology? Is psychology a thing

54:33

about which there can be a scientific

54:35

understanding? Is it something where we can

54:37

have a summarization of how psychology will

54:39

work as opposed to just, well, and

54:41

then this neural network with 100 billion

54:44

neurons runs and it does this? which

54:46

is not, you know, science is about

54:48

trying to take what happens in the

54:50

world and turn it into kind of

54:52

a human narrative. And that's, that's something

54:55

that, that it's not obvious you'll be

54:57

able to do that for something like

54:59

psychology. And we're seeing that question for

55:01

LLLM psychology right now as well, of

55:04

to what extent can we summarize what

55:06

the LLLM's do? Can we get a

55:08

feeling for how the LLLM's work? I

55:10

mean, I know in our company, we

55:12

are essentially... sort of getting people to

55:15

learn to be LLLM psychologists. That's what

55:17

we need. We need that because we

55:19

are trying to get the LLLM, we're

55:21

trying to wrangle the LLLM to do

55:24

the right thing. And it's like, well,

55:26

what caused the LLLM to go crazy

55:28

at this point? We need kind of

55:30

an LLLM psychologist to figure that out.

55:32

And we're trying to get intuition about

55:35

doing that. And the question is, you

55:37

know, what level of intuition can we

55:39

get? Can we get as humans? Could

55:41

the LLLMs get about themselves? And then

55:44

the question is, well, what level of

55:46

intuition can an LLLM get about how

55:48

humans think about things? It's a very

55:50

interesting question. I think that the LLLans...

55:52

thumbs. can perhaps model how the humans

55:55

think about things in a sort of

55:57

fairly parallel way, even if you can't

55:59

understand at a narrative level how the

56:01

LLLMs are thinking, and you can't understand

56:03

how the humans are thinking, maybe they

56:06

will be close enough in the way

56:08

that the LLLMs can parallel the way

56:10

that humans think, that at least you'll

56:12

be able to have the L11 kind

56:15

of run in its mind a simulation

56:17

of what the human is going to

56:19

think and so on. I mean, right

56:21

now LLLMs are not very fast compared

56:23

to humans. comparable speed in some ways.

56:26

And that will change. The LLLMs will

56:28

get faster. And thus humans, well, I'm

56:30

afraid we're stuck, you know, operating at

56:32

the many millisecond timescale of nerve firings.

56:35

You know, you can't, you can, you

56:37

can sort of ingest all kinds of

56:39

chemicals to try and speed that up.

56:41

It isn't going to work. It's, you

56:43

know, the whole system is built for

56:46

a certain clock rate. Maybe when we

56:48

get, you know, we'll be able to,

56:50

just like the LLMs can call tools,

56:52

we will have neural implants that are

56:54

like our ability to call tools. So

56:57

it's like we're thinking, and then for

56:59

an LLM, it's we're thinking, we're thinking,

57:01

and then we generate this piece of

57:03

text that is the tool called, it

57:06

goes off to this external tool, you

57:08

got an answer, you know, all from

57:10

language results, comes back, comes back. that

57:12

gets read ingested by the LM. My

57:14

guess is that we'll be able to

57:17

do similar kinds of things directly with

57:19

our brains and that will be sort

57:21

of a, but we won't get the

57:23

sort of raw clock frequencies is not

57:26

going to go up. And maybe that

57:28

will be a place where you can

57:30

kind of have. the simulation of what

57:32

the human would be thinking, running outside

57:34

the human, and it's like, well, you

57:37

know, in the last 200 milliseconds, I,

57:39

the LLLM, figured out, listen, listen, listen,

57:41

listen, listen, listen, this. So I think

57:43

the right next thing to tell you

57:46

is blah, because that's what you're about

57:48

to figure out, so to speak. And

57:50

I can imagine that absolutely happening. the

57:52

robotics space, having kind of a wolfram

57:54

robotics, so to speak. Yes, I had

57:57

been interested in robotics for a long,

57:59

long time. My views about how that

58:01

will come to fruition have changed actually

58:03

a bit quite recently. I mean, what

58:05

I had always assumed was that robotics

58:08

would kind of come of age when

58:10

there was general purpose robotics, as computing

58:12

came of age when there were general

58:14

purpose computers. There was a time before

58:17

the 1940 and so on where where,

58:19

and even a little bit beyond that,

58:21

where, you know, you had a computer

58:23

that did one particular thing, another one

58:25

that did another particular thing. And then

58:28

the idea of universal computing that came

58:30

in initially theoretically and then in practice

58:32

came to pass where it's like you

58:34

just have a fixed computer and you

58:37

program it to do all the different

58:39

things you want it to do. Well,

58:41

I had been interested for a long

58:43

time in kind of modular robotics where

58:45

you had little, little components, you would

58:48

have this... thing made of sugar cube-like

58:50

objects or maybe smaller than that and

58:52

it would have all this kind of

58:54

this flowing of the sugar cubes to

58:56

be just the right shape and it's

58:59

kind of like it would walk along

59:01

a little bit like an amoeba might

59:03

walk along generating pseudopods and so on

59:05

and you'd see these sort of sugar

59:08

cubes moving around and it's kind of

59:10

like a liquid object that is sort

59:12

of sliming around the world so to

59:14

speak. I mean it wouldn't actually be

59:16

liquid it will be solid and I

59:19

you know. probably 15 years ago, I

59:21

was more than that actually, I was

59:23

interested in kind of how you would

59:25

build things like this, try and think,

59:28

try to figure out some of the

59:30

geometry of it, of how things slide

59:32

against each other, how you maintain enough

59:34

contact to keep power to each of

59:36

these little cubes and so on, not

59:39

cubes. We had a bunch of geometries

59:41

that were very much not cubes, that

59:43

kind of rubics cube like, like assembly

59:45

things that could slide against each other

59:48

and so on. And then how would

59:50

you kind of plan? or how does

59:52

it move from this configuration of that

59:54

configuration and so on. So I had

59:56

assumed for a long time that that's

59:59

kind of how robotics would go. down,

1:00:01

but eventually, rather than, you know, my

1:00:03

experience, for example, at robotics trade

1:00:05

shows, had been this weird thing,

1:00:08

that there's an aisle of hands,

1:00:10

there's another aisle of feet and

1:00:12

legs and so on, every aisle

1:00:14

of the trade show is a

1:00:17

different part of the body type

1:00:19

thing, or a different kind of

1:00:21

thing, whereas... It's that you kind of

1:00:24

needed this different special purpose hardware for

1:00:26

all these different kinds of things and

1:00:28

I imagine that sort of the future

1:00:30

there was this kind of notion of

1:00:32

a modular robot where it was really

1:00:34

just all just many copies of the

1:00:36

same thing and some of the work that

1:00:38

I did and cellular automatra and things like

1:00:40

that both I and other people kind of

1:00:43

picked up on some of that trying to

1:00:45

understand how you would build

1:00:47

this kind of modular robotics technology.

1:00:49

My guess right now is that actually

1:00:51

it will be much more like the

1:00:54

science fiction said. It will be much

1:00:56

more humanoid robots will learn to do

1:00:58

things like humans do. And I think

1:01:00

that the humanoid form factor for

1:01:02

a robot is convenient because our world

1:01:05

is built for things that have humanoid

1:01:07

form factors because we built all

1:01:09

of these buildings and machines and so

1:01:11

on to be operated by us humans

1:01:14

with, you know, the size we are,

1:01:16

the hands we have, things like this.

1:01:18

So it's a good, sort of a plug

1:01:21

compatible way to introduce robotics into the

1:01:23

world is to make the robots be

1:01:25

like us humans. Now the thing that's

1:01:27

always been a challenge is like even

1:01:29

how do you pick anything up? You know,

1:01:31

if the thing is a slithery thing and

1:01:33

you know, oh you have to know to

1:01:35

put it underneath there and... and you put

1:01:38

it between your two hands and so on,

1:01:40

or if it's a thing that's very fragile,

1:01:42

you have to know just what force you

1:01:44

can use, and is it going to slide

1:01:47

out of your fingers and all this kind

1:01:49

of stuff? And this has been a big

1:01:51

challenge, and there's been these sort of formalized

1:01:53

challenges of picking challenges and things relevant for

1:01:56

companies like Amazon and so on for operating

1:01:58

their warehouses, and it really hasn't. made, so

1:02:00

far as I can tell, that much

1:02:02

progress until recently. I think the thing

1:02:05

that's sort of now coming is being

1:02:07

able to use sort of the general

1:02:09

ideas of large language models and apply

1:02:11

them to the behavior of robots and

1:02:13

really crack the problem of having a

1:02:15

robot that can do stuff like a

1:02:18

human does. Now it's kind of a

1:02:20

funny thing because for text... There's, you

1:02:22

know, trillions of words of training data

1:02:24

that's out there from all the things

1:02:26

humans have written for what humans do

1:02:29

and how they pick things up and

1:02:31

so on. Well, there's a bunch of

1:02:33

video you can watch, but fundamentally we

1:02:35

don't have that data. And so it's

1:02:37

kind of funny, I visited a few

1:02:40

of these companies that have sort of

1:02:42

people in suits, capturing all this motion.

1:02:44

so that you can teach the robot

1:02:46

kind of how to do these kinds

1:02:48

of motions, giving it that type of

1:02:51

information. I think there's a certain amount

1:02:53

of autonomous stuff that is based on

1:02:55

kind of reinforcement learning of... of you

1:02:57

know let the robot try and feel

1:02:59

its way through the world so to

1:03:02

speak quite literally to learn how to

1:03:04

do things but my guess is that's

1:03:06

going to come and it might come

1:03:08

even quite suddenly that humanoid type robots

1:03:10

become able to do a lot of

1:03:13

tasks and I think there'll be a

1:03:15

you know my guess is there'll be

1:03:17

another it'll probably come suddenly maybe not

1:03:19

I'm not sure and it'll be another

1:03:21

kind of chat gPT like moment where

1:03:24

people say oh my gosh all these

1:03:26

things that only humans could do before

1:03:28

suddenly the machines can do them too.

1:03:30

And they won't do them perfectly, but

1:03:32

they'll do them well enough for many

1:03:35

purposes. And then the real issue, it's

1:03:37

an interesting question, is sort of how

1:03:39

do you fit those capabilities into what's

1:03:41

useful in the world? Just like ChatGPT,

1:03:43

you know, if you want ChatGPT to

1:03:46

solve your computational... problem. It's not a

1:03:48

good idea. It's not going to work.

1:03:50

But if you want it to summarize

1:03:52

this piece of text for you or

1:03:54

to answer questions about a piece of

1:03:56

text or to generate some report that

1:03:59

can be fed to maybe another online,

1:04:01

it's going to be great at those

1:04:03

things or to be able to do

1:04:05

the kind of thing we do in

1:04:07

a notebook assistant where you're asking it

1:04:10

something vaguely and it's using a bunch

1:04:12

of techniques. fundamental L&M nature, so to

1:04:14

speak, to generate precise war from language

1:04:16

code that you can then do things

1:04:18

with. Those things are going to do

1:04:21

great. You make it do the wrong

1:04:23

thing, it's not going to do very

1:04:25

well at all. So it will be

1:04:27

with sort of L11-style robotics for a

1:04:29

large behavior model, LBM-style robotics. There'll be

1:04:32

things that that can immediately do well,

1:04:34

and there'll be things where, yeah, that's

1:04:36

probably not a good idea to have

1:04:38

it do that. I don't know exactly

1:04:40

what the things that will do well

1:04:43

are. It's an interesting question, actually. It's

1:04:45

worth thinking about. It's like, you know,

1:04:47

things in, I don't know what, the

1:04:49

construction industry, for example, which employs a

1:04:51

lot of people. Is that going to

1:04:54

be a place where, yeah, the robot

1:04:56

can get the material for the right

1:04:58

place on the building site or whatever

1:05:00

else? I know from talking to CEOs

1:05:02

of companies that are making humanoid robots,

1:05:05

at least as a year ago, a

1:05:07

bunch of them were saying, well, one

1:05:09

of the use cases is tidying up

1:05:11

after you. I don't know if that's

1:05:13

a realistic one. That seems really hard

1:05:16

to me. It seems like not one

1:05:18

of the ones that you would pick

1:05:20

as a top priority, especially, you know,

1:05:22

depending on, maybe it's one that has

1:05:24

a certain amount of wiggle room. It's

1:05:27

just like, you know, if you make

1:05:29

a machine learning a machine learning system

1:05:31

rank. do rankings for a search engine,

1:05:33

that's a fairly reasonable thing to do

1:05:35

because, you know, if two out of

1:05:37

ten of the things that it ranks

1:05:40

highly are exactly what you want, it's

1:05:42

a win. And if one or two

1:05:44

of the things are completely crazy, doesn't

1:05:46

really matter. So there will be some

1:05:48

tasks for robotics where doing it roughly

1:05:51

right is good enough. And there are

1:05:53

other tasks where... where, if you get

1:05:55

it even slightly wrong, it's a disaster.

1:05:57

I don't know whether tidying up a

1:05:59

room is probably one of these things

1:06:02

that's somewhat forgiving task. I mean, I

1:06:04

remember, oh gosh, how long ago was

1:06:06

it? It was probably, what was it?

1:06:08

Must be, eight years ago maybe? I

1:06:10

remember being at the Consumer Electronics Show.

1:06:13

And one of the big attractions was

1:06:15

this laundry folding robot. There's a big

1:06:17

line to see the laundry folding robot,

1:06:19

but it wasn't that exciting actually in

1:06:21

the end. I don't know if it's

1:06:24

become a product that's really used, but

1:06:26

it's a surprise, it was at that

1:06:28

time a surprisingly hard thing to fold

1:06:30

laundry. My guess is that the next

1:06:32

generation of robots will find it very

1:06:35

easy to fold laundry. That's probably a

1:06:37

task that again is somewhat forgiving. It's

1:06:39

like, oh, you get it slightly wrong,

1:06:41

and the fold is in slightly the

1:06:43

wrong place, big deal, big deal. And

1:06:46

so, you know, there will be a

1:06:48

bunch of things that very quickly, my

1:06:50

guess is, will be doable by humanoid

1:06:52

robots, and then there'll be other things

1:06:54

where people might try to do them

1:06:57

with humanoid robots, and it's sort of

1:06:59

a bad idea and doesn't work very

1:07:01

well. But it'll be interesting to see,

1:07:03

I'm kind of guessing, it'll be a

1:07:05

sudden thing that there'll be just, a

1:07:08

whole bunch of stuff just works. I

1:07:10

mean, there have been practical problems with

1:07:12

humanoid robots, like providing enough power. I

1:07:14

mean, I mean, I'll eat foods, get

1:07:16

muscles, get muscles, is actually a surprisingly

1:07:18

efficient way to get things to move.

1:07:21

And a lot of humanoid robots, kind

1:07:23

of the big trick is there's a

1:07:25

giant cable coming out of the back

1:07:27

that's supplying a lot of power to

1:07:29

the robot. I mean, the thing to

1:07:32

understand about muscles and so on is

1:07:34

that, well, it's actually, muscles are pretty

1:07:36

clever. You know, they have these little

1:07:38

devices on the molecular scale that are

1:07:40

kind of walking down the muscle, pulling

1:07:43

on it and so on. Those are

1:07:45

things which are operating on a molecular

1:07:47

scale. they are much more energy efficient

1:07:49

than the things we have operating at

1:07:51

the scale of electric motors and so

1:07:54

on. And I don't know, you know,

1:07:56

I don't know to what extent the

1:07:58

kind of humanoid robot and it's, you

1:08:00

know, things like the practicalities, the hardware

1:08:02

practicalities, to what extent that's going to

1:08:05

be a long thing factor. It may

1:08:07

also be that those hardware practicalities as

1:08:09

soon as one is, oh, let's make

1:08:11

10 million of these robots, it starts

1:08:13

being much more feasible to have that

1:08:16

work well. I also suspect the battery

1:08:18

technology is improving, you know, has improved

1:08:20

it enough that there may be a

1:08:22

different set of power constraints that exist

1:08:24

there. So many questions, guys, thank you.

1:08:27

Wylo asks. If people delegate all calculations

1:08:29

to machines, might it not happen that

1:08:31

the machine actually learns to ask better

1:08:33

questions than the humans can, since the

1:08:35

machines have the experience built from the

1:08:38

calculations that the humans don't? Well, the

1:08:40

question is, what do you want to

1:08:42

ask the question about? I mean, I

1:08:44

could ask all kinds of questions about

1:08:46

why, you know, the... the tree that

1:08:49

I see out of my window has

1:08:51

this particular pattern of branches and not

1:08:53

another? Why the cloud in the sky

1:08:55

has this particular pattern of fluffiness and

1:08:57

not another? Right now we humans don't

1:08:59

care about those questions. There are an

1:09:02

infinite number of questions we can ask,

1:09:04

most of which we don't care about.

1:09:06

So the question of which do we

1:09:08

care about is as much a sort

1:09:10

of societal question. And a human, human

1:09:13

question as anything, it's not something where

1:09:15

there's sort of an abstract answer. Now,

1:09:17

I think there will come a time,

1:09:19

and it's sort of already happened to

1:09:21

some extent where AI's can suggest, hey,

1:09:24

this is something you as a human

1:09:26

might care about, kind of the idea

1:09:28

generator thing. And yes, I think that's

1:09:30

happened. I think it's, it's, you know,

1:09:32

it's an interesting question when the ideas

1:09:35

that people come up with. are assisted

1:09:37

by AIs and those ideas are kind

1:09:39

of coming out of the exhaust from

1:09:41

the humans, so to speak. What happens

1:09:43

to this whole system? You know, is

1:09:46

it the case that the ideas just

1:09:48

revert to, well, okay, it's a standard

1:09:50

idea? It's just like I see in

1:09:52

things like, I don't know, math enrichment

1:09:54

activity. that people propose, I guess I

1:09:57

haven't looked at these recently, but a

1:09:59

while ago I did. And it's kind

1:10:01

of amusing, I mean, sometimes they're things

1:10:03

based on my work, which is always

1:10:05

nice to see, but there's a certain

1:10:08

similarity to, you know, okay, we're going

1:10:10

to do math and Richmond, okay, then

1:10:12

there are five ideas we might have

1:10:14

that all seem like sort of spunky

1:10:16

original creative ideas, but there was the

1:10:19

same five ideas type thing. And I

1:10:21

could imagine that happening when sort of

1:10:23

all ideas are all ideas at LLLM-s

1:10:25

sourced sourced. Let's see. Weasel is commenting

1:10:27

tiny humans care about those questions. I'm

1:10:29

not sure what the questions were at

1:10:32

this point. The Dan is asking, what

1:10:34

will AI not be able to do?

1:10:36

Do I believe that something like that

1:10:38

exists? Well, the answer is kind of

1:10:40

what I've been saying, the choice of

1:10:43

which direction to go, there are an

1:10:45

infinite number of possible choices. Which one

1:10:47

we choose to go in is a

1:10:49

thing for us? because there's no right

1:10:51

answer. Now we could delegate that to

1:10:54

the AIs, we could say, we're giving

1:10:56

up, just like, okay, from here on

1:10:58

out, it's cruise phase, you know, cruise

1:11:00

phase for humanity, we just do whatever

1:11:02

the AIs tell us to do. My

1:11:05

guess is that our nature is not

1:11:07

such that that will go down very

1:11:09

well. Our nature, you know, forged by

1:11:11

three billion years of biological evolution, has

1:11:13

to do as well, we're going to

1:11:16

seek out the new and so on.

1:11:18

So my guess is that the, you

1:11:20

know, you know, you know, you know,

1:11:22

we're just here for the, we're just

1:11:24

here for the, we're just here for

1:11:27

the, we're just here for the, isn't

1:11:29

going to play out well for us

1:11:31

humans and we won't allow that to

1:11:33

be what happens but I suppose in

1:11:35

principle that could happen. I mean I

1:11:38

think it's a you know a thing

1:11:40

that I always find kind of kind

1:11:42

of amusing in the sense I never

1:11:44

really dug into this I don't really

1:11:46

know the facts as clearly as I

1:11:49

should but you know anthropologists tend to

1:11:51

say you know there are a few

1:11:53

cases in human history where there have

1:11:55

been some you know some tribes somewhere

1:11:57

where there were enough berries on the

1:12:00

bushes that they could, you know, feed

1:12:02

themselves and that it was... it was

1:12:04

an easy life type thing. And that,

1:12:06

so they weren't, you know, pushed by

1:12:08

necessity to do more. And then the

1:12:10

claim is, well, for a thousand years

1:12:13

or more, whatever it is, they were

1:12:15

just hanging out, you know, eating the

1:12:17

berries from the bushes and so on,

1:12:19

and not being pushed to do more,

1:12:21

not having sort of a struggle for

1:12:24

life. And then the question is, well,

1:12:26

what do people do in that situation?

1:12:28

And what I've heard said is, well,

1:12:30

then they do ritualisticistic things. Which is

1:12:32

of course a very strange thing to

1:12:35

say because in a sense viewed from

1:12:37

the outside Many things we do would

1:12:39

be thought could be thought of as

1:12:41

ritualistic like we don't understand from the

1:12:43

outside why anybody would do that But

1:12:46

from the inside we absolutely have a,

1:12:48

oh yeah, we're doing that because we

1:12:50

have a whole inner dialogue about why

1:12:52

we're doing those things. So, you know,

1:12:54

no doubt to the ritualistic tribes, so

1:12:57

to speak, the things they were doing,

1:12:59

you know, shaking sticks and, you know,

1:13:01

interacting with the rain gods or whatever

1:13:03

else they were doing or thought they

1:13:05

were doing, is, you know, that has

1:13:08

a complete meaning to them that to

1:13:10

the outside just looks like, oh, that's

1:13:12

ritualistic, we don't know what's going on.

1:13:14

you know, any kind of life of

1:13:16

the mind, I suppose, could look ritualistic

1:13:19

from the outside. And I kind of

1:13:21

think that, you know, we could imagine

1:13:23

a situation where us humans are doing

1:13:25

that. I kind of think that the

1:13:27

thing we will continue to be relevant

1:13:30

for is pick the path. There are

1:13:32

an infinite number of paths. We are

1:13:34

the defining elements of which path to

1:13:36

pick, unless we absolutely choose to advocate

1:13:38

on that and not do that. Oh,

1:13:41

weasel is saying the questions about

1:13:44

clouds and trees and why are

1:13:46

they those shapes? Yeah, interesting. Right.

1:13:48

Good question. I don't know. I'm

1:13:50

trying to remember if any of

1:13:52

my kids asked that question. I

1:13:55

should remember. I can remember many

1:13:57

interesting questions that they asked. which

1:13:59

were indeed, indeed, I mean, because

1:14:01

I've studied a bunch of sort

1:14:03

of foundational science, maybe I was

1:14:05

in a better position than most

1:14:08

to be able to answer lots

1:14:10

of kid questions. I guess that's

1:14:12

perhaps an inspiration for some of

1:14:14

these live streams, although you guys

1:14:16

tend to ask more sophisticated questions

1:14:19

than my kids when they were five

1:14:21

years old asked. And although my kids asked

1:14:23

some very interesting questions. I

1:14:25

remember one that stuck out for a

1:14:28

long time was... five-year-old, one of my

1:14:30

kids, was when there were dinosaurs, could the

1:14:32

Earth have had two moons? And that question

1:14:34

is really hard to answer, and I,

1:14:36

but for years, whenever I would

1:14:39

run into people who did celestial

1:14:41

mechanics, planetary dynamics, kinds of things,

1:14:43

I would ask that question.

1:14:45

And people would say, well, I'm

1:14:47

not sure. You know, we have

1:14:49

these simulations, I don't really know.

1:14:51

I think the answer now... is

1:14:53

pretty clearly no, the Earth couldn't

1:14:55

have had two moons six to

1:14:57

five million years ago, but it could

1:14:59

easily have had two moons

1:15:02

a billion years ago. We

1:15:04

wouldn't know that. And it's

1:15:06

quite possible that it did. But

1:15:09

that's kind of a, well, it's

1:15:11

sort of an obvious question.

1:15:13

I'd never thought of that question. Sancho

1:15:15

is commenting. At least at the time

1:15:18

when I went to a bunch of

1:15:20

them, commenting that Boston Dynamics shows a

1:15:22

lot of progress in the humanoid department.

1:15:24

I'm not sure. I live in the

1:15:26

Boston area and there's a building on

1:15:29

Route 128 in Boston that is now

1:15:31

the Boston Dynamics building. It's been other

1:15:33

companies in the past. And when I

1:15:35

drive by there, I'll sometimes kind of

1:15:37

try and look in the windows to

1:15:40

see whether there are any cool robots

1:15:42

there. They don't seem to be most

1:15:44

of the time. I'll probably run into there

1:15:46

at least. founder on some of them

1:15:48

that I'm going to soon. So maybe

1:15:51

I'll hear more about, I'll

1:15:53

probably see more actually, also

1:15:55

as an event that I'm going

1:15:57

to next week that will. have

1:16:00

probably a bunch of the latest

1:16:02

and greatest robotic things at it.

1:16:04

Always interesting. It's always the question,

1:16:06

you know, I was at a

1:16:08

robot company, oh, what was it,

1:16:10

sometime last year, and it's like

1:16:12

there's a humanoid robot and you

1:16:14

can shake its hand. And I'm

1:16:16

not quite trusting enough that I'm

1:16:19

prepared to let the robot put

1:16:21

its hand around my hand. I'm

1:16:23

more in the let me grab

1:16:25

your hand robot from the outside.

1:16:27

We'll see how that evolves. Let's

1:16:29

see, Mighty is commenting, anything to

1:16:31

say about the future of pie,

1:16:33

happy pie day. Happy pie day

1:16:35

to everybody. This is an American

1:16:37

dates, it's 314, the first digits

1:16:39

of pie. What's the future of

1:16:41

pie? Interesting question. I mean, we've

1:16:43

got a trillion digits computed so

1:16:46

far. Will we get more? Will

1:16:48

anybody ever find a pattern in

1:16:50

the digits of pie? A statistical

1:16:52

pattern, none has been found. Nobody's

1:16:54

ever proved that there isn't a

1:16:56

statistical pattern. I would say the

1:16:58

big stretch goal for pie is

1:17:00

to show pretty much anything about

1:17:02

the digits of pie. Right now

1:17:04

we have no idea. All we

1:17:06

know is that the digits of

1:17:08

pie don't repeat. Pi is not

1:17:10

a rational number, so the digits

1:17:13

of Pi don't repeat. But beyond

1:17:15

that, knowing whether in base 10,

1:17:17

for example, they're the same number

1:17:19

of ones and twos and threes

1:17:21

and fours and so on, nobody

1:17:23

has any idea. People have been

1:17:25

trying to figure that out for

1:17:27

the last 150 years, and really

1:17:29

nothing is known about that. That's

1:17:31

kind of one of these things

1:17:33

where one might say mathematics is

1:17:35

not really ready for that question

1:17:38

yet. you know as we build

1:17:40

out mathematics we kind of build

1:17:42

out these different paradigms these different

1:17:44

sort of in the space of

1:17:46

or possible theorems we're gradually building

1:17:48

out more and more that we've

1:17:50

kind of colonized we've reached maybe

1:17:52

three or four million theorems out

1:17:54

of the infinite number of possible

1:17:56

theorems of mathematics and the question

1:17:58

is have we yet colonized that

1:18:00

we yet visited that we explored

1:18:02

the part of mathematical space that

1:18:05

would lead us to answer a

1:18:07

question like, is pie so-called normal?

1:18:09

Does it have equal numbers of

1:18:11

every digit and every block of

1:18:13

digits and so on? So that's

1:18:15

an example of kind of a

1:18:17

coming attraction for pie. I don't

1:18:19

think that's close. I think it's

1:18:21

an interesting question whether, you know,

1:18:23

as we automate more mathematics, I've

1:18:25

been involved obviously in automating lots

1:18:27

of mathematics, I don't know. whether

1:18:29

sort of a question like that

1:18:32

will come, has more of a

1:18:34

chance to come over the horizon,

1:18:36

I don't know. As far as

1:18:38

other things about, well, it's, yeah,

1:18:40

I think that's the main thing

1:18:42

that I can see in sort

1:18:44

of the future of pie is,

1:18:46

you know, what pie day will

1:18:48

we know that there's something to

1:18:50

say generally about the digits of

1:18:52

pie? Let's see. There's a question

1:18:54

from Jay Chen, do you expect

1:18:57

a limb development to hit significant

1:18:59

diminishing returns within the next two

1:19:01

to three years? Well, I think

1:19:03

it already has. I mean, I

1:19:05

think that what the story of

1:19:07

machine learning tends to be a

1:19:09

story of sort of a new

1:19:11

domain gets cracked and then there's

1:19:13

a big jump at that point,

1:19:15

and then there's kind of incremental

1:19:17

progress after that. So, you know.

1:19:19

image recognition got cracked around 2011

1:19:21

to 2012. Things like speech to

1:19:24

text got cracked in the, when

1:19:26

was it that, something like, oh,

1:19:28

let's see, something around the late

1:19:30

2010s, you know, text generation, packed

1:19:32

with LLLans. Now these things get

1:19:34

gradually better and the harnesses around

1:19:36

the LLLans, the ways to use

1:19:38

these things, the ways to use

1:19:40

AIs, they get there gets to

1:19:42

be more and more understanding how

1:19:44

can you use this technology. I

1:19:46

mean it's just like when people

1:19:48

invented ways to do fast linear

1:19:51

algebra, fast matrix computations. It's like

1:19:53

well that's a thing and that's

1:19:55

kind of cool and then people

1:19:57

realize well we can use that

1:19:59

to do computer graphics and then

1:20:01

that started a whole direction and

1:20:03

sort of the core of what

1:20:05

was done there was is matrix

1:20:07

algebra. Same thing actually with Ella

1:20:09

Lambs it's lots of matrix algebra

1:20:11

going on inside the the GPUs

1:20:13

and so on that are driving

1:20:15

Ella Lambs. So You know, my

1:20:18

own feeling is that it's what

1:20:20

you see in pretty much all

1:20:22

areas, whether it's in science, technology,

1:20:24

whatever, there's a breakthrough, there's a

1:20:26

new methodology, there's a big jump,

1:20:28

then things level out. And you

1:20:30

know, in science it might take

1:20:32

100 years before you get to

1:20:34

the next kind of methodological breakthrough

1:20:36

and big jump up. I mean,

1:20:38

I like to think in physics,

1:20:40

for example, that our physics project

1:20:43

is finally... basically 100 years after

1:20:45

the last sort of big methodological

1:20:47

advances in physics, we finally got

1:20:49

another set of big methodological advances

1:20:51

that are opening up a bunch

1:20:53

of new new possibilities. But you

1:20:55

know, my guess is that what

1:20:57

will happen in the in the

1:20:59

LLLM world is that a bunch

1:21:01

of new modalities, whether it's video,

1:21:03

whether it's robotics, whether it's things

1:21:05

with, you know, chemical sensing, I

1:21:07

don't know, you know, there'll be

1:21:10

these modalities that kind of open

1:21:12

up where they were pretty much

1:21:14

Greenfield. There was nothing there before

1:21:16

and then they get they get

1:21:18

solved or they get substantially solved

1:21:20

and that's where the big advances

1:21:22

will come rather than you know

1:21:24

the the incremental kind of let's

1:21:26

make it a bit better if

1:21:28

we look at something like image

1:21:30

recognition you know what we have

1:21:32

today is better than what we

1:21:34

had in 2012 but not that

1:21:37

much. And what's important is we

1:21:39

understand much more how to take

1:21:41

the things that we could do

1:21:43

in 2012 and fit them into

1:21:45

a harness that really fits into

1:21:47

other kinds of things. So I

1:21:49

think that's that's kind of that's

1:21:51

that's the way I see that

1:21:53

developing and I see I need

1:21:55

to go back to my day

1:21:57

job in a minute here. Let's

1:21:59

see if there's maybe one more

1:22:02

quick question. commenting that automated theory

1:22:04

improving is interesting. I'm trying to

1:22:06

figure out how to make a

1:22:08

theory improver that demonstrably collapses the

1:22:10

wave function, like I was talking

1:22:12

about earlier, quantum LLMs. Yeah, I

1:22:14

mean, this question of how you

1:22:16

knit together sort of these kind

1:22:18

of computational methods like automated theory

1:22:20

improving with what's being done with

1:22:22

our limbs. Very interesting question. I

1:22:24

think I made a bit of

1:22:26

progress on that last year. Peace

1:22:29

I wrote about sort of why

1:22:31

does machine learning work and this

1:22:33

kind of discrete model of machine

1:22:35

learning that allows one sort of

1:22:37

lays much more bare the kind

1:22:39

of the essence of machine learning

1:22:41

I think gives one a much

1:22:43

better handle on what one would

1:22:45

need to do to kind of

1:22:47

thread together those lumps of machine

1:22:49

learning functionality. together with kind of

1:22:51

computational functionality to do things like

1:22:53

sort of integrate theorem proving with

1:22:56

LLLMs and that would be where

1:22:58

I would start to look for

1:23:00

trying to do that. I haven't

1:23:02

figured out how to do it

1:23:04

but that's it's on my to-do

1:23:06

list. I think it's such a

1:23:08

bit far away so it's it's

1:23:10

I encourage other people to to

1:23:12

look at it. All right well

1:23:14

we should wrap there I see

1:23:16

a bunch of other interesting questions

1:23:18

which I look forward to to

1:23:20

trying to address another time. But

1:23:23

it's always fun talking with you

1:23:25

guys and the questions you ask

1:23:27

are interesting and get me thinking

1:23:29

about things I'm not the worst

1:23:31

thinking about and I think I've

1:23:33

been, thank you for contributing to

1:23:35

sort of the future of science

1:23:37

and technology. I'll try to be

1:23:39

the executor of some of these

1:23:41

things that come up as ideas

1:23:43

here and perhaps some of you

1:23:45

can be as well. Anyway, thanks

1:23:48

for joining me and Until the

1:23:50

next time, bye for now. You've

1:23:52

been listening to the Stephen Wolfram

1:23:54

podcast. You can view the full

1:23:56

Q&A series on the Wolfram Research

1:23:58

YouTube Channel. For more information. on

1:24:00

Stephen's live streams,

1:24:02

and this podcast, visit

1:24:04

StevenWolfram .com.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features