Google's Secret AI Advantage (Why DeepMind Will Dominate)

Google's Secret AI Advantage (Why DeepMind Will Dominate)

Released Tuesday, 25th March 2025
Good episode? Give it some love!
Google's Secret AI Advantage (Why DeepMind Will Dominate)

Google's Secret AI Advantage (Why DeepMind Will Dominate)

Google's Secret AI Advantage (Why DeepMind Will Dominate)

Google's Secret AI Advantage (Why DeepMind Will Dominate)

Tuesday, 25th March 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:01

Hey everyone, I am excited. We have

0:03

a very very special guest. Logan Capatrick

0:05

from Google Deep Mind is here and

0:07

he's gonna break down the aspects of

0:09

Google Jim and AI that people aren't

0:12

using, that they should be. Gonna talk

0:14

about where the real value is, gonna

0:16

talk about their brand new image generation

0:18

API, their new reasoning models, deep research,

0:20

so much. If you were trying to

0:23

learn about Google AI, this is the

0:25

show for you. Let's get to today's

0:27

episode. We'll

0:31

be right back to today's show, but first

0:33

here's a quick word from Hub Spot.

0:35

Marketing in 2025 is wild. Savvy

0:37

customers spot fake messaging instantly. Privacy

0:40

changes make ad targeting a nightmare.

0:42

And everyone needs more content than

0:44

ever. That's why you have to

0:47

have Hub Spot's new marketing trends

0:49

report. It doesn't just show you

0:51

what's changing. It shows you exactly

0:53

how to deal with it. Everything's

0:56

backed by research, but focused on

0:58

marketing plays you can use tomorrow.

1:00

Ready to turn your

1:02

marketing hurdles into results.

1:05

Go to hubspot.com/marketing to

1:07

download it for free.

1:22

Logan really excited to have you on the podcast

1:24

big follow of yours on Twitter for a long

1:26

time or X for a long time Excited about

1:28

your move to Google having following that as well,

1:31

but let's maybe start not at the start for

1:33

you But maybe at the start for some of

1:35

us who started following along your journey with being

1:37

an early employed over at Open AI like one

1:39

of the most transformative companies there is One of

1:41

the things we want to start with is it

1:43

must have been a pretty wild time to be

1:45

part of that company What was one of those

1:47

exciting launches that you were there for? One of

1:50

the ones where you were like, wow, this is

1:52

just going to be much more impactful than anyone

1:54

really gives a credit for? Yeah, that's a

1:56

good question. I think the one that always

1:58

stands out for me and I'm Obviously in

2:00

hindsight we all know how impactful this launch

2:03

was. It was really simple and sort of

2:05

had all the fanfare as far as like

2:07

actually being a part of it on the

2:10

ground was honestly the GBT4 launch. When we

2:12

launched GBT4 if folks remember the live stream

2:14

that Greg Brockman did where he sort of

2:16

talked about the models and he sort of

2:19

did the infamous example of like drawing a

2:21

little picture of a website and showing the

2:23

model and then it writing the code like

2:25

you think now what. that model was able

2:28

to do two years ago versus today, it's

2:30

like we're in an entirely different world than

2:32

we were two years ago. But I think

2:35

that example, I worked with Greg a bunch

2:37

on that demo and was actually, I have

2:39

a great picture like sitting right in front

2:41

of him as part of that demo and

2:44

we were going through it. And I think

2:46

the reflection for me is just like on

2:48

how far we've gotten from that moment, both

2:50

from like an AI like actual capability standpoint,

2:53

but also like I think a lot of

2:55

the innovation has been the infrastructure to the

2:57

infrastructure to. bring AI to like actually be

3:00

useful. And I think even today I was

3:02

having a conversation last night with someone who

3:04

was talking about the raw capabilities of the

3:06

model versus like this sort of AI building

3:09

harness that has been created in the last

3:11

two years and how much that actually makes

3:13

a difference for like the raw capabilities that

3:15

you can get from these models. And I

3:18

think it still feels like we're so early

3:20

in that sense that like you can get

3:22

those like GBT4 level. And we were talking

3:25

off camera before about native image generation, which

3:27

I'm and like. having this like GPT-4 level

3:29

moment of people seeing, wow, this is like

3:31

an incredible experience like just out of the

3:34

box. So it's still so fun to see,

3:36

you know, you think all the juice has

3:38

been squeezed out and then actually you're right

3:41

around the corner from like something that's going

3:43

to change how people think about the world.

3:45

Yeah, there was a couple of quick pointers

3:47

about that specific video. Kipp and I covered

3:50

that video. I feel like that demo was

3:52

like the nearest we've ever had to like

3:54

the iPhone moment for it was like very

3:56

similar to like that iPhone moment, but for

3:59

AI where people were like, holy smokes, like

4:01

this is, I get it, right, like they

4:03

collect. world who watched that demo that one

4:06

succinct use case really kind of sparked creativity

4:08

across such a wide range of folks internally

4:10

were you always all kind of constantly surprised

4:12

by how big those launches were like just

4:15

how transformative they were or did you kind

4:17

of know this is going to rapidly change

4:19

the way that people think about the world

4:21

like when we launched GD4 everything is different?

4:24

Yeah I think GD4 actually was one of

4:26

those moments and there's sort of some somewhat

4:28

differing perspectives about chat GBT and how much

4:31

of a unknown success story that was going

4:33

to be. I think the GBT four folks

4:35

knew like the real reason that chat GBT

4:37

came out was so that open AI could

4:40

try to experiment with different experiences of bringing

4:42

capable models to the world knowing that GBT

4:44

four was coming. So like chat GBT was

4:46

sort of intended to be the early, you

4:49

know, get feedback from the world. Does this

4:51

chat about thing actually end up being useful

4:53

for models? Because at that point GPT4 had

4:56

already finished training and you know opening I

4:58

was trying to figure out how do we

5:00

productize this really really cool technology but I

5:02

think folks knew GPT4 was going to be

5:05

that useful like they'd been sitting on it

5:07

for a very long time like I think

5:09

the model finished training in summer of 2022

5:11

and sort of made its way out to

5:14

the world in March of 2023 so there

5:16

was a long while where like folks had

5:18

almost fully wrapped their head around this technology

5:21

and what I was capable of, sort of

5:23

experience of experience of getting to sit on

5:25

the technology for that long and actually having

5:27

it be differentiated with what was available externally

5:30

in the world and thinking about now, I

5:32

think about for us, it's like, you know,

5:34

model comes out of the oven from being

5:36

trained and it's like, it has all the

5:39

safety stuff baked in and like, let's get

5:41

it out to the world in 24 hours.

5:43

And I think about like, actually, there's a

5:46

real trade-off and I think I'm not sure

5:48

how relevant this is for people who aren't

5:50

releasing models to the world who aren't releasing

5:52

models to like. figure out the story and

5:55

like explore and build it and like we're

5:57

actually figuring out a lot of that stuff

5:59

like with the external world as we're all

6:01

using this model together publicly and maybe that's

6:04

actually the best thing for the world because

6:06

you don't want to be like sitting on

6:08

the alien technology for you know six months

6:11

before it actually makes its way out but

6:13

I think you only get that level of

6:15

like really cool demo that Greg was able

6:17

to do by being able to sit on

6:20

the technology and like really internalize and I

6:22

think Greg to his credit as part of

6:24

that demo like he drove that whole thing

6:26

and like he was able to put it

6:29

together because he had fully internalized like what

6:31

the model was actually capable of and like

6:33

I don't see a lot of that happening

6:36

with like today's era of launches which is

6:38

really interesting. Yeah I think the point you

6:40

bring up is really important is that like

6:42

2020-25 is like a year in which a

6:45

decade is going to happen right like the

6:47

pace is very aggressive and if you're watching

6:49

the show and you may be a distant

6:51

observer of AI What it generally is doing

6:54

for everybody, any company, is that the expectations

6:56

of pace and speed have just gone up.

6:58

It's not just the AI models, it's literally

7:01

every company. And so you have to know

7:03

what your stories are, kind of your core

7:05

principles of what you're building, so that you

7:07

can kind of continue to build the story

7:10

and the product in parallel, because it's rare

7:12

you're going to have the like, hey, I

7:14

know everything, I've got this smooth six months

7:17

launch period, gone are those days right now,

7:19

right, right, right, especially if you are out

7:21

there building stuff, because that six months somebody

7:23

might build something way better and the work

7:26

you have is just completely obsolete and that

7:28

is I think just what 2025 I will

7:30

probably remember the most is that just kind

7:32

of core change in speed and trajectory. Yeah

7:35

I wish for my own sanity that that

7:37

was not the case because it feels like

7:39

this is like the three-year sprint that never

7:42

stops so it's also at a very human

7:44

level feels more and more important than ever.

7:46

And this is what gets me so excited

7:48

about, you know, the human experience. This world

7:51

of AI and all the innovation, the pace

7:53

that's happening, like at the end of the

7:55

day, it just like further exacerbates how. important

7:57

it is for like to do all the

8:00

things that it means to be human and

8:02

to like have those experiences and I so

8:04

fundamentally believe that. Yeah, it gets me happy

8:07

on both ends of the spectrum because I

8:09

love to see all the coolness of stuff

8:11

coming out in AI but like I'm also

8:13

just so bullish on like the human experiences

8:16

that only humans can have and and create

8:18

and all that stuff even in a world

8:20

where AI is intelligent. So you changed, you

8:22

went to Google, Google's pace is also, I

8:25

think, expediated. Like they have some really killer

8:27

launches. I think maybe they're a little more

8:29

understated than Open AI or some of these

8:31

other companies because they have such a whole

8:34

host of other challenges to navigate when they

8:36

release AI. But what are maybe a couple

8:38

of subtle differences that you've seen between how

8:40

Open AI approach and AI in general and

8:43

what Google's approaches? Like what are some of

8:45

the subtle differences you've seen how those

8:47

two companies are trying to expediate AI

8:49

to the world? Yeah, I'll maybe take two

8:51

ends of the coin here. First from

8:53

like the core technology standpoint, but second

8:55

like how we show up in the

8:57

world and talk about our AI stuff,

8:59

which I think will be helpful for

9:01

folks who are thinking about this. On

9:03

the core technology side, one of the

9:06

things that Google really benefits from is

9:08

like the breadth of all the work

9:10

that's happening. So our team's now a

9:12

part of Google Deep Mind. On one

9:14

end of the spectrum, we have, you

9:16

know, alpha fold and protein folding and

9:19

Nobel Prize, you know, image generation models

9:21

and there's weather models and there's Gemini

9:23

and there's there's this huge breadth of

9:25

different stuff that's happening from an AI

9:27

perspective and Deep Mind is really the

9:30

only place in the world where that

9:32

depth is actually happening and you can look

9:34

at all the other labs and you can

9:36

see the models that they're creating and like

9:38

the models that are being created at other

9:40

labs is like one pillar of what Deep

9:42

Mind is doing and I think the really

9:45

exciting thing to me is like if you

9:47

buy into the idea that this like multi-modal

9:49

multi-capability enabled model is going to be the

9:51

thing that enables humans to do all the

9:53

things that we want to do like there's

9:55

only one place in which it's possible that

9:58

that's going to be created and it's inside

10:00

of DeepMind because of this breadth

10:02

of all the work that we're doing.

10:04

And we see this actually happening

10:06

in practice with the cross -pollination of

10:08

research from AlphaFold, again, to weather models,

10:10

to AlphaProof, which is our math

10:13

model, and how all that actually trickles

10:15

back into the mainline Gemini model

10:17

that is available to consumers and is

10:19

available to developers. So I'm super

10:21

excited about that. I'm excited for us

10:23

to lean more into that story

10:26

and lean more into that advantage that

10:28

DeepMind has on the other end

10:30

of the coin after the fundamental research

10:32

and product creation happens. I think

10:34

there's a huge difference in how we

10:36

approach going and telling the world

10:39

about the products and the models that

10:41

we're building. And I think a

10:43

lot of this is just grounded in

10:45

the positions that these different companies

10:47

are in. There's a

10:49

lot of factors at play.

10:51

I think people perceive Google as

10:53

a company very differently than

10:55

people perceive OpenAI as a company.

10:57

OpenAI also has a very

10:59

different product offering. Google has many,

11:02

many, many products across many,

11:04

many different domains. And there's just

11:06

a lot of downstream impacts.

11:08

Beautiful thing for OpenAI is they

11:10

have a clean slate in

11:12

many ways. So people either haven't

11:14

formed a prior about some

11:16

specific angle of what they're doing

11:18

or even something very tactically.

11:20

They have an open namespace. They

11:22

can call their products anything.

11:24

They can use whatever URL they

11:26

want, et cetera, et cetera,

11:28

because there's nothing conflicting with that.

11:30

They don't need to worry

11:32

about the crossover between these different

11:34

products. And I think about

11:36

this a lot because we get

11:38

a lot of feedback from

11:40

the external world that, hey, we

11:42

wish this thing was simpler.

11:44

We wish this naming schema was

11:46

a little bit easier to

11:48

follow. And a lot of this

11:50

is just the artifact of

11:52

the complexity that Google has because

11:54

of how large of a

11:56

company it is. And then in

11:58

turn, we have to find

12:00

ways to try to more authentically

12:03

lean into the things that

12:05

matter to us. And I think

12:07

this is one of the

12:09

biggest challenges for Google. Like, I'm

12:11

sure there's folks in your

12:13

audience who... have thought about this or experienced this at

12:15

different companies, like it's just hard to tell a really authentic

12:17

story as Google, not because there's not interesting authentic stories happening

12:19

inside of Google, it's just because it's the artifact of the

12:21

size of a company that you're a part of the size

12:23

of a company that you're a part of, and I think

12:25

as the size of a company that you're a part of,

12:27

and I think as the size of a company increases, like,

12:29

about how we can tell this authentic story because

12:31

I feel like we miss telling the magic

12:33

of why this technology is so important when

12:36

you sort of don't go the authentic route.

12:38

And I like feel these so deeply inside

12:40

of me these like really authentic interesting AI

12:42

stories that again that only Google can tell

12:45

and like but the only way to land

12:47

that message is in a really authentic way

12:49

and there's just like so many different angles

12:51

of this tension to reconcile and I just

12:54

don't think that open AI is an example

12:56

like they don't have to deal with this

12:58

because they're just not the size of a

13:00

company where like they have all those different

13:03

dimensions of tension right now. The one thing

13:05

I would probably add Logan as a consumer

13:07

of both products is that like one of

13:09

the advantages Google has is that like the

13:12

breadth of all of their tools and being

13:14

able to just like seamlessly integrate Gemini and

13:16

even if the story isn't there. When I

13:19

opened up Google Maps and I looked at

13:21

a place and Jim and I was now

13:23

on the place listings and I could just

13:25

ask Jim and I anything about that place,

13:28

I was like, holy cow, this is incredible.

13:30

I could just ask it, oh, I'm

13:32

going to this place in New York.

13:34

If I get there at 10 a.m. How

13:36

long am I going to have to wait

13:38

in line? Basic things like that that would

13:41

have been impossible or I

13:43

would have read like 30 credit threads

13:45

to find out the answer is now

13:47

like a 10. into those wow moments.

13:49

It's almost the story is getting told by

13:51

the user just discovering those features too. Yeah,

13:53

I think that's a great point. And I

13:55

think this speaks to like, again, one of

13:57

those advantages from the breadth perspective is like.

13:59

when Deep Mind builds Gemini. We're not

14:02

building it for a chat app. Like

14:04

if you think about like where Gemini

14:06

is actually being integrated, it's like across

14:08

some of the largest product suites that

14:10

touch the most users in the entire

14:12

world and like that has a very

14:14

different set of constraints that would potentially

14:16

be built and like Gemini is powering

14:18

search, Gemini is in YouTube and like

14:20

these are like billion billion billion user

14:22

products that like all of these very

14:24

nuanced characteristics matter a lot. And actually

14:26

we've seen like lots of really great

14:28

examples of you know the requirements for

14:30

a model to be really good for

14:32

search, actually leading to like something that's

14:34

really great from a developer perspective. We've

14:36

seen this with a few of the

14:38

last 1.5 series of models where, you

14:40

know, the search team needed something and

14:42

it ended up being like a really

14:44

great tradeoff of capabilities that developers also

14:46

ended up wanting those things and it's

14:48

cool to have those levels of sort

14:50

of internal engagement from these teams. One

14:52

of the things you mentioned Logan was

14:54

there's just like a plethora of different

14:56

AI tools available kept mentioning that we're

14:58

going to see a decade of progress

15:00

in a single year. And so one

15:02

of the things we wanted to do

15:04

in this show is to try to

15:06

distill it down into like what do

15:08

you think are great AI use cases

15:10

for our audience to take away from

15:12

the show and start to implement and

15:14

maybe we can specifically think about Gemini

15:16

and Google, but what are some of

15:18

the AI use cases today? Like that

15:20

you can actually just go and start

15:22

using Google Gemini for today that you

15:24

think are like widely underestimated or underused

15:26

by the average consumer of AI, the

15:28

person who maybe isn't in the details

15:30

and in the weeds day in day

15:32

out. This is a good question. I

15:34

think one of the challenging things about

15:36

this question is that it is a

15:39

ever evolving answer because literally the capability

15:41

flywheel is spinning as we speak, which

15:43

is awesome. I think today some of

15:45

the things that are getting me most

15:47

excited is in December, we launched deep

15:49

research. So we launched sort of the

15:51

world's first iteration of deep research, which

15:53

if folks haven't used it, is essentially

15:55

a research assistant. You can put in

15:57

whenever your query is and the model

15:59

will go off and search. in the

16:01

context of our deep research, visit, you

16:03

know. thousands potentially of different websites to

16:05

answer the question. I think that simple

16:07

product artifact of like showing you the

16:09

number of websites that the model browsed

16:11

through in order to get you the

16:13

answer is the thing that just makes

16:15

that product experience work for me. I'm

16:17

like in zero percent of any of

16:19

the research I've ever done in my

16:21

life, have I looked at more than

16:23

10 websites? So the fact that the

16:25

model went out and looked at a

16:27

thousand websites like just gives me like

16:29

a lot of confidence in the model.

16:31

I'm happy to up and we can

16:33

talk more about this, but I think

16:35

this is one of the biggest challenge

16:37

for AI products, which is like in

16:39

many cases AI products are like asking

16:41

the user to basically do all this

16:43

up front work in order to provide

16:45

them value. And I think deep research

16:47

is this great example of like the

16:49

model of the technology, just sort of

16:51

doing the heavy lifting for you and

16:53

you as a user get to just

16:55

like ask your silly question or your

16:57

serious question, and then the model goes

16:59

and finds it. So I love deep

17:01

research. line of reasoning models, which is

17:03

awesome. I agree with you, Logan, that

17:05

like far and away the number one

17:07

thing, if nobody's really using AI for

17:09

much than just some random questions, they

17:11

should use deep research. Deep research is

17:14

incredibly powerful. And the amazing thing about

17:16

Google Gemini, and you correct me if

17:18

I'm wrong, is that deep research is

17:20

a free feature. I don't think there's

17:22

any hard limits on deep research yet.

17:24

And really, the biggest thing there is

17:26

you then rolled out these reasoning models,

17:28

which makes it think through the ability

17:30

of what sources and what follow-up questions

17:32

to ask around those sources far, far

17:34

better than it was just a couple

17:36

of months ago. So you could enter

17:38

this drop-down and you would pick whatever

17:40

model you want to use. I don't

17:42

think people fully always grock. When do

17:44

I use this deep research? Like what's

17:46

the perfect way to delineate between flash,

17:48

which is also incredibly great an AI

17:50

model. Now I need the deep research,

17:52

plus you mentioned something really important, which

17:54

is now you've integrated reason and into

17:56

that. Maybe just explain what we... mean

17:58

to users that you've integrated reasoning to

18:00

this deep research? Yeah, I think that

18:02

basically the land is like your everyday

18:04

questions, you know, if it's a simple

18:06

question. just use 2.0 flash, it's going

18:08

to be very quick. It'll get you

18:10

an answer like almost instantly. If you

18:12

really do need something that is not

18:14

surface level, like if you're looking for

18:16

like, you know, who won the Cubs

18:18

game yesterday, you know, if you don't

18:20

need deep research for that, if you're

18:22

trying to understand like why the Cubs,

18:24

and I'm based in Chicago, and I'm

18:26

not even a Cubs fan either, but

18:28

I'm using the Cubs example, but if

18:30

you're trying to. understand why the Cubs,

18:32

you know, builds whatever the ivory wall

18:34

is around the back of the field

18:36

and like what the technique that they

18:38

used to build that and who the

18:40

people were who worked on it and

18:42

like what type of permitting they needed

18:44

in order to put that together, like

18:46

that level of depth in the question

18:48

that you have. There's no product that

18:51

can do that besides deep research. And

18:53

I think the reasoning models is the

18:55

sort of key enabler this and we

18:57

initially launched deep research back in December

18:59

with Jim and I 1.1.1.5 Pro. It

19:01

was really powerful, but a lot of

19:03

the techniques being used by 1.5 pro

19:05

were actually like trying to get it

19:07

to do what a reasoning model actually

19:09

does, which is be able to sort

19:11

of have essentially this like inner monologue

19:13

of, you know, thinking through different pieces

19:15

of a question, like actually reflecting back

19:17

on the initial answer that it's given

19:19

in like trying different versions of this.

19:21

And you can sort of think about

19:23

how we as humans think through this

19:25

process, by default AI models just kind

19:27

of. spit out an answer as quickly

19:29

as they can is basically the way

19:31

that models are trained today. And the

19:33

thinking models are trained to like don't

19:35

actually spit out the answer as quickly

19:37

as you can. iteratively go through this

19:39

process, try a bunch of different things,

19:41

make sure that you're sort of covering

19:43

the breadth and depth of what a

19:45

user might be asking for. And it

19:47

actually leads to some pretty substantively different

19:49

outcomes. So if folks have tried stuff

19:51

before with AI and you're like, ah,

19:53

the models are just dumb and don't

19:55

really have the ability to do these

19:57

things, try the new sort of generation

19:59

of reasoning models. I think there's a

20:01

lot of use cases that just were

20:03

not possible before that like all of

20:05

a sudden just work today. I am

20:07

very convinced that most humans do not

20:09

realize what like deep research especially is

20:11

capable of. For example, like I was

20:13

wanting to get rid of like a

20:15

tree in my yard and I was

20:17

like, I don't know what the permitting

20:19

process is. I don't know what you

20:21

would need to do. It did all

20:23

of that told me exactly what I

20:26

could and couldn't do, what the rules

20:28

are. And I literally just said. I

20:30

wanted to remove a tree and gave

20:32

it my address and it did everything

20:34

else, right? And it's like people just

20:36

wouldn't think that it could do things

20:38

like that. Like I had to estimate

20:40

an entire construction project. It's like, well,

20:42

all right, somebody's giving me this estimate.

20:44

What does deep research say? And it's

20:46

incredible how detailed it is and its

20:48

ability to go through complex documents and

20:50

frame things in simple ways is really,

20:52

really good now. Yeah, and I think

20:54

it's wild to just reflect on that.

20:56

This is like the V0 of the

20:58

product. Like truly, like, this is the

21:00

bare bones version of what deep research

21:02

can possibly be. And like today, it's

21:04

essentially just using search. And there's more,

21:06

and I actually think at the time

21:08

of this going out. because it's rolling

21:10

out today, the ability for deep research

21:12

to be combined with audio overviews from

21:14

Notebook LM is also rolling out. So

21:16

now you can sort of take that

21:18

deep research experience that you had and

21:20

then actually just click a button, turn

21:22

the entire thing and do an audio

21:24

overview and then you have a podcast.

21:26

Let's do that. And you're learning about

21:28

the permitting process in whatever city you

21:30

are. And you can ask it questions

21:32

and interrupt that podcast and be like,

21:34

well, I don't understand what you mean

21:36

here. Yeah, so I think the whole

21:38

experience and the way in which humans

21:40

sort of get this type of information

21:42

is changing, which is I think a

21:44

good thing. I'm curious if you had

21:46

this perspective on the permanent, I just

21:48

wouldn't do it to be honest or

21:50

like, or I would pay like, you

21:52

end up paying like this like market

21:54

inefficient price because you're like, well, I'm

21:56

not willing to. interesting to me so

21:58

like therefore I'm gonna get gouged by

22:00

somebody and they're gonna charge me twenty

22:03

thousand dollars to take a tree out

22:05

of my backyard or something crazy like

22:07

that. But it's also not even the

22:09

price right it's the time to get

22:11

to that same outcome would have taken

22:13

me weeks yeah and it took like

22:15

five minutes yeah right it's like I

22:17

can just accomplish so much more in

22:19

my life and I don't think any

22:21

human realizes like the rate of progress

22:23

you can now make. So

22:30

if you're listening to this podcast,

22:33

you are more than likely thinking

22:35

about business ideas and how to

22:37

market your business. Every entrepreneur has

22:40

a million ideas vying for attention

22:42

in their brain. And there's probably

22:44

that one idea you can't stop

22:46

thinking about. The one that came

22:49

to you at 2am and you

22:51

got up and you wrote it

22:53

down somewhere. Well, I'm here to

22:55

tell you that that idea could

22:58

be worth $5,000 right now. The

23:00

hustle and indie hackers are teeming

23:02

to find the next great founder

23:04

and get him or her started.

23:07

with $5,000. All you have to

23:09

do is record a 60-second video

23:11

pitch telling them your business idea

23:13

and how it makes money and

23:16

why it's awesome. One winner gets

23:18

$5,000 cash plus promotion to millions

23:20

of hustle readers. Make sure you

23:23

enter by April 4th at the

23:25

Hustle.co/Big Break. The winner will be

23:27

announced on April 11th on the

23:29

Hustle Daily Show podcast, the Hustle

23:32

newsletter, and on the Hustle's social

23:34

channels. Now get to pitch it.

23:37

It's a combination of there's things you would

23:39

never have done because you didn't have

23:41

the time or money and now you can

23:44

do and just how quickly you can do

23:46

them. And one of the ones I just

23:48

don't want to gloss over because you mentioned

23:50

something that is actually one of my favorite

23:53

workflows for deep research. We did an entire

23:55

video of it, which is deep research into

23:57

notebook and then turned it into audio, create

24:00

podcast and ask podcast questions. That's actually

24:02

how I've been learning about call options. which

24:04

this is the worst time to start doing

24:06

call options, but that's a whole other thing

24:09

that can I complain around. No financial advice,

24:11

but that's one of the things we've been

24:13

playing around with. But when you say that's

24:15

being integrated, do you mean I can just

24:18

go to Google Gemini, like you're integrated in

24:20

LM notebook, into Gemini, or is the

24:22

research still going to exist separately, or are

24:24

they going to be combined? How is that

24:27

integrated together? Yeah, I think for today's launch,

24:29

when you do a deep research query inside

24:31

of the Gemini app, you'll see an option

24:33

when the deep research process is done to

24:36

like convert the deep research into an audio

24:38

overview. And I actually think maybe, don't quote

24:40

me on this, I think maybe, don't

24:42

quote me on this, I think there's maybe

24:45

like a direct way on this. I think

24:47

there's maybe like a direct way on this,

24:49

I think maybe, don't quote me on this,

24:51

like, would go back in a notebook alum.

24:54

But it is cool to just be able

24:56

to have that single shot version of the

24:58

overview created for you. Yeah, maybe since

25:00

we're on notebook alum, why don't you give

25:03

the users a little bit of your ways

25:05

of using notebook alum? Because that's actually a

25:07

favorite tool of this show. Again, I don't

25:10

think enough people even know that tool or

25:12

are using it enough. Yeah, I think for

25:14

folks who haven't tried notebook LM, you can

25:16

think about notebook LM sort of as the

25:19

knowledge assistant that you might have, like

25:21

imagine or a learning assistant or a tutor

25:23

that you might have, or honestly, just like

25:25

a way to bring content to life is

25:28

another version of this. Josh Woodward, who leads

25:30

the Google Labs team, has this idea of

25:32

sort of like infinite repurposing of content and.

25:34

I think notebook alum is like a really

25:37

great example of this. You can take, imagine

25:39

you have, you know, something really boring

25:41

like a onboarding manual to set up, you

25:43

know, a vacuum cleaner. And for whatever reason,

25:46

you're one of the 10 people who actually

25:48

reads the onboarding manual and you have to

25:50

read it, but you're really bored and you

25:52

don't. actually want to go through the whole

25:55

thing. You could take a PDF version of

25:57

that, you know, a 150-page onboarding manual that

25:59

says all the different gizmos and gadgets

26:01

and things that your product does, stick it

26:04

in a notebook album, and really quickly generate

26:06

like a summary, a learning guide, a principally,

26:08

like one of the things that folks are

26:11

most excited about, a podcast level conversation with

26:13

like really sort of witty and smart sort

26:15

of nuance in the conversation about whatever the

26:17

content is that you're uploading and you're

26:19

uploading. has features like you can actually interject

26:22

mid-conversation and say things like, you know, hey,

26:24

this is actually really boring, spice things up

26:26

a little bit, or, you know, I don't

26:29

understand this point that you were just making,

26:31

can you sort of re-articulate what this is?

26:33

And one of the most common flows that

26:35

I actually see folks who I work with

26:38

and I work remotely so I don't

26:40

commute to the office, even though I probably

26:42

should go to the office every once in

26:44

a while, folks take a bunch of like

26:47

work documents, they put them in notebook alum,

26:49

create audio reviews, and they like listen to

26:51

them on the drive of the work. Yes,

26:53

exactly. I know Kip has a ton of

26:56

use cases here, but just two really quick

26:58

ones. So this one here is exactly

27:00

what you said. It's like a strategic knowledge

27:02

of this, and that's specific to one of

27:05

the cross-functional pods that I run here. And

27:07

one of the cool things I can do

27:09

is like add all the documentation. So I

27:11

add all like meat and transcripts, add all

27:14

the docs that have been created that month

27:16

that week. And then when I'm out walking

27:18

to dogs, I can actually either listen

27:20

to it or listen to it. way of

27:23

being able to talk to the podcast directly

27:25

and ask questions. And so it's a pretty

27:27

cool way where you can actually take knowledge

27:30

with you with work and actually conversate with

27:32

it. Like if you're out of the office,

27:34

if you're driving, if you're walking, I have

27:36

knowledge assistance trained on every single project

27:38

and I can always ask them questions. And

27:41

so it's like you have a project assistant

27:43

slash executive assistant for every project. And as

27:45

I said, I wasn't lying. Like all the,

27:48

this is like one of the example where

27:50

I had deep research. Teach me about call

27:52

options and then try to pick out like

27:54

five call options that are underpriced and why

27:57

believe so things are underpriced. priced, using

27:59

third party sources, and what I say is

28:01

I use the most trusted third party sources

28:03

to come up with that hypothesis. And then

28:06

I had a whole conversation with it in

28:08

this interactive mode when I loaded the podcast

28:10

and had a whole conversation that lasted like

28:12

15, 20 minutes, all around its prediction on

28:15

why meta was a good call option. And

28:17

I kept telling you that I was

28:19

actually, that was done four weeks ago, and

28:21

actually it was pretty right, because I think

28:24

Kipp has already done that trade. So, maybe,

28:26

pretty powerful. I don't keep you love this

28:28

tool, but that's pretty powerful stuff. Well,

28:30

hold on. One of the threads we're

28:32

going here, and I do think this

28:34

is true, is that the Google suite

28:36

of AI products that you all

28:38

have wrote out, I think are the best

28:41

to help humans learn. Because

28:43

I'm about to pull a deep

28:45

cut out of one of my favorite

28:47

Google tools that nobody talks about,

28:49

because we've talked about deep research, we've

28:51

talked about, about, notebook, you can basically

28:53

just decide you want to learn about

28:56

something like how do I make pasta

28:58

from scratch for example and what I

29:00

love about learn about it's like where

29:02

deep research is like a deep dive

29:04

on a topic to gain perspective and

29:07

then you can kind of ask follow-up

29:09

questions on notebook alum. This is

29:11

much more like almost like a structured

29:13

learning process and course and it creates

29:15

this amazing composable web experience over here

29:17

and then you can break down and

29:19

dive into different components and basically It

29:22

solves a lot of the empty box

29:24

problem of like, hey, I don't know

29:26

anything about this thing. I don't even

29:28

know what to ask. And it's prompting

29:30

me a bunch of different aspects about

29:32

making pasta. I happen to know how to make

29:34

pasta, but it's like, if you don't know how

29:36

to add pasta to boiling water because you've never

29:38

done it before, like this is very helpful, right?

29:40

And so I think if you look at Deep

29:42

Research, Notebook LM and Google Learn, like there

29:45

are three tools that are really master classes

29:47

that are really master classes in helping people

29:49

learn. Yeah, I love that. And I feel like

29:51

the magic, the real magic and the bow of

29:53

this is like, how can you bring all that

29:56

together, understand a user's intent, put

29:58

the right product. in front of

30:00

them. Yes. Yes. This is going to be

30:02

my point. This is the challenge. The sentence

30:05

that I just said sounds, you know, simple

30:07

to say, when you actually look at it,

30:09

like these are the engineering and product problems

30:11

of the decade. Like really, these are like

30:14

not a trivial thing to bring together that

30:16

level of technology, especially in a world where

30:18

you're sort of balancing this user context

30:20

problem, which is like if folks. talk

30:23

to AI models all the time, like

30:25

every time you talk to a new

30:27

AI model, it has no context of

30:29

who you are, it doesn't know what

30:31

you've done before, which is another thing

30:33

that sort of just landed in the

30:35

Gemini app, which is the ability to

30:37

personalize model responses with your Google search

30:40

history. And the model can really intelligently

30:42

say, like, here's what this user has

30:44

done before, here's what they're interested in,

30:46

how can we actually use that to

30:48

sort of prime the model to give

30:50

you the model to give you the right?

30:52

pulling in the right like for you capital

30:54

know like here's this learn about like experience

30:57

because Kip's been doing a bunch of stuff

30:59

over here and this other product versus you

31:01

know search for call options and bringing that

31:03

stuff so there's a whole like spectrum of

31:05

different use cases and I think having the

31:08

personalized context means that you can get the

31:10

right product surface or you can get the

31:12

right product experience in front of the right

31:14

user persona which I think is not something that

31:17

happens in today's software 2.0 product suite.

31:19

software is static, it's like predefined for

31:21

you in a lot of ways. Just

31:23

to touch on this, because I think

31:25

this is a really important point you're

31:28

making, and actually something even internally we

31:30

are dealing with a little bit, which

31:32

is what you're basically saying is instead

31:34

of at some point, go into this

31:36

kind of drop-down experience, you will have

31:38

one assistant that you talk to, and

31:40

then the background the assistant is basically

31:42

deciphering intent, and the assistant can pass

31:44

you to whatever model that thinks fulfills

31:46

that. then you all have to do.

31:48

We have a sales assistant that can

31:50

actually help sell through chat. We have

31:52

a support assistant that can help do

31:54

a bunch of support tickets. We need

31:57

to actually create an upsell assistant so

31:59

we can actually upgrade. customers to different

32:01

tiers and sell them on those different

32:03

tiers. And what we have is like

32:05

these individual assistants. And the way we

32:07

decide for intent is somewhat like wherever

32:09

you are in the go to market,

32:11

we say, well, this is probably the

32:13

right assistant. But even us, we need

32:15

multi-bought orchestration at some point where you

32:17

have one assistant, the assistant can say

32:19

you're trying to buy product for the

32:21

first time. You have a support question.

32:23

You are our existing customer and you

32:25

need to upgrade. And that. is really

32:27

complicated. And so I can't even imagine

32:29

how complicated it is for a Google

32:31

where the intent is like any single

32:33

possible thing in the world. Like it's

32:35

like, how do you ever decipher that

32:37

intent? But I think what you're saying

32:39

is even when I think about my

32:41

AI experience with Google, you all have

32:43

a ton of great products and I

32:45

would love to end with one of

32:47

the ones that I think is. maybe

32:49

transformational for this year, the image generation,

32:51

but they're disparate. We jumped into notebook

32:53

algorithm, kept jumped into the learning product,

32:55

we jumped into all these different AI

32:57

models you can choose through to the

32:59

drop-down. Is the plan for Google to

33:01

like pull them all together at some

33:03

point? So you just have like one

33:05

AI interface as a consumer. Yeah, this

33:07

goes back to one of the challenges

33:09

that Google has. It's a company that

33:11

has lots of product services. I do

33:13

think more and more of the Gemini

33:15

app is becoming this sort of unified

33:17

place to get a lot of these

33:19

experiences and sort of a externalization path

33:21

for Kipa showing learn about these experiences

33:23

and sort of a externalization path for

33:25

Kipa showing Learnabout, which is one of

33:27

the Google Labs experiments. I think more

33:29

and more of those experiences are finding

33:31

their way into the Gemini app. user

33:33

journeys. Like I think about the four

33:35

or five different products that I use

33:37

like all day every day and like

33:39

there's different journeys apart of those products

33:41

for different users and I think they'll

33:43

continue to be like different products. We

33:45

didn't talk about AI Studio but like

33:47

A.S. Studio is another one of these

33:49

where like that's actually how folks access

33:51

the native image generation capabilities today and

33:53

like I think in the future we're

33:55

still going to have A. I. Studio

33:57

because like the user persona that we

33:59

care about A. is a developer who's

34:01

sort of exploring these models and wants

34:03

to build something, which is very different

34:05

in a lot of cases than sort

34:07

of the main experience for users going

34:09

to the Gemini app who like are

34:11

trying to use the Gemini app really

34:13

as like an assistant on a daily

34:15

basis. It's like a daily active user

34:17

product versus AI studios really intended as

34:19

sort of the portal to the developer

34:21

world for folks interested in Gemini. Maybe

34:24

talk about. the latest image release. When

34:26

I think about AI and where we

34:28

might have been a little wrong, I

34:30

thought video would be further along than

34:32

it is today. And I mean, AI's

34:34

ability to go from text to video,

34:36

right? It's still pretty clunky in a

34:38

lot of cases. It's not like production

34:40

ready. And then the other one was

34:42

images, text to image. And I felt

34:44

like the first iteration of that were

34:46

all these great tools where you could

34:48

go to image, but for image to

34:50

be really useful for people, there had

34:52

to be great editing tools as part.

34:54

Google's latest release, Kip and I were

34:56

playing around with it last week, is

34:58

really awesome. And maybe you could just

35:00

give the context on what that model

35:02

is and why it's so good. Yeah,

35:04

no, 100%. So for folks who haven't

35:06

been following closely, we launched Gemini 2.0

35:08

back in December, and we showcased these

35:10

capabilities actually, and we rolled them out

35:12

to a small group of trusted testers

35:14

to get some initial feedback. And then

35:16

last week, we rolled out to every

35:18

developer, the ability to use Gemini's native

35:20

image generation. And I think the thing

35:22

that's actually capturing a lot of interest

35:24

is the native image editing capability. Because

35:26

the model is natively multimodal, you can

35:28

pass in an image. And you can

35:30

say. hey, update this image to, in

35:32

the example we're looking at on screen,

35:34

add a little chocolate drizzle to these

35:36

crescents, or add a strawberry drizzle to

35:38

these crescents, or I've seen a bunch

35:40

of really cool examples, which I did

35:42

not think about, of taking black and

35:44

white images and actually asking the model

35:46

to colorize those images. And you can

35:48

bring them back, you can take in

35:50

two images and you can sort of

35:52

fuse them together. You know, I saw

35:54

a funny example of like. like hot

35:56

dogs and stuff like that and merging.

35:58

together into this comedic image and lots

36:00

of different random things like this that

36:02

is the full spectrum of really useful

36:04

to really silly. But the thing that

36:06

I think is capturing folks' attention is

36:08

if you think about how you would

36:10

have had to do this workflow, pre-native

36:12

image generation and image editing, it's just

36:14

hard to do. Like the number of

36:16

people who can do that in whatever

36:18

the professional tool is, is pretty limited.

36:20

You can just do it with like

36:22

a very, very simple text prompt. So

36:24

like now the opportunity space of people

36:26

like creating dynamically edited images is now

36:28

essentially every human on earth is able

36:30

to do that, which is just. This

36:32

is the thing that continues to blow

36:34

my mind is you get a capability.

36:36

And this is what AI is enabling

36:38

across so many different domains. It's this

36:40

thing that only a few people could

36:42

do. And then overnight it ends up

36:44

being this thing that everyone can do.

36:46

And then overnight it ends up being

36:48

this thing that everyone can do. And

36:50

I think actually coding is like before

36:52

image generation is also having this like

36:54

parallel moment with vibe coding and everything

36:56

they want. pretty much get what they

36:58

were looking for. It's the same thing

37:01

with image editing today. Before you had

37:03

to be very good at using one

37:05

of these tools in order to do

37:07

this stuff, and now it's like every

37:09

human on earth can do this, and

37:11

it just changed overnight, which is just

37:13

such a weird experience to think about.

37:15

Yeah, vibe design. I like that. This

37:17

is the kind of dream, I think,

37:19

for AI in general, or the transformative

37:21

mission of AI in general, as it

37:23

unlocks. creativity somewhat overnight because it allows

37:25

people now to unleash their creativity and

37:27

they're no longer hindered by having to

37:29

learn the tools. And I know that

37:31

sounds like a lazy way to think

37:33

about creation, but I don't think it

37:35

is because I think the creation part

37:37

is the important part and the learning

37:39

the tools shouldn't hinder that part of

37:41

you, like the ability to create things.

37:43

Is this visual story? Could you just

37:45

end on what this product is? Because

37:47

we discovered this today. I'd love to

37:49

know like, what is the way we

37:51

should think about visual story, which again,

37:53

for people who were following along on

37:55

YouTube or even on RS. These are

37:57

all available in AI Studio and you

37:59

should really go in and play with

38:01

that product because there's a bunch of

38:03

great tools in there Yeah, AI Studio

38:05

is again. It's our surface for developers

38:07

intended to bring the models to life

38:09

in a way that ultimately wants to

38:11

get you to build with them. So

38:13

here we're trying to, this example that

38:15

we're looking at, trying to capture a

38:17

developer, sort of imagination of what products

38:19

they could go and build themselves, but

38:21

principally AI studio is not intended to

38:23

be like the daily assistant product. Like

38:25

it's a very thin surface on top

38:27

of the models. We make a bunch

38:29

of like very opinionated decisions to keep

38:31

the core AI studio experience the same

38:33

as the experience. you would get in

38:35

the API. So we don't have a

38:37

bunch of like fancy bells and whistles

38:39

that you yourself couldn't do as a

38:41

developer in AI Studio, which makes it

38:43

limiting as a product. Like if you

38:45

want, like people are always like, why

38:47

don't we have deep research in AI

38:49

Studio? I want that. It's like because

38:51

deep research is something you could build

38:53

sort of a similar deep research experience

38:55

using the API, but it's not available

38:57

to developers today. So we don't want

38:59

that experience in AI Studio. Yeah, so

39:01

if you're someone who wants to build

39:03

stuff, getting your Gemini API key, all

39:05

that stuff, testing out the latest capabilities

39:07

of the models, happens in AI Studio,

39:09

and our team sits in Google Deep

39:11

Mind right next to the model team

39:13

physically, so that oftentimes our product is

39:15

sort of the fast path to externalize

39:17

the latest Gemini models, which is a

39:19

lot of fun, and it's cool to

39:21

be on the frontier, and it's cool

39:23

to see the excitement with native image

39:25

generation. Very cool. kind of like the

39:27

way ChatGBT was for text. I feel

39:29

like this is similar for images. It's

39:31

really the first time I've seen the

39:33

ability to kind of get the image

39:35

crisp and concise to the way you

39:38

want it. This was awesome. Logan, we

39:40

really appreciate you coming on and giving

39:42

us a deep dive into what it's

39:44

been like to be part of this

39:46

kind of AI journey. You're working with

39:48

two of the most transformational companies there

39:50

are you or you have and also

39:52

just going deep into like how people

39:54

can start using the Google tools straight

39:56

after this episode. Yeah, this is the

39:58

time for having me. Hopefully the call

40:00

options go well and we'll... I'll be

40:02

starting somewhere together in a few months.

40:04

That's our infamous test to whether AI

40:06

is truly good or not. Is it

40:08

making us money with call options? I

40:10

love it. Thank you so much Logan.

40:12

This was awesome. Appreciate the time in.

40:14

Yeah, of course. Right

40:34

back to today's show, but first,

40:36

I want to tell you about

40:38

a podcast I love. It's called

40:40

Nudge. It's hosted by Phil Agnew,

40:43

brought to you by the Hubspot

40:45

podcast network. Ever notice that the

40:47

smallest changes can have a big

40:49

impact? On Nudge, you learn simple

40:51

evidence back tips to help you

40:54

kick bad habits, get a raise,

40:56

grow a business. Every bite-sized 20-minute

40:58

show comes packed with practical evidence

41:00

from admired entrepreneurs and behavioral and

41:02

you can apply. If you're a

41:04

marketer, you're going to love this

41:07

show because it discusses the psychology

41:09

behind stellar marketing. They just did

41:11

a great episode titled 99.9% of

41:13

ads are genuinely awful with Tom

41:15

Goodwin. You can learn why so

41:18

many ads today are ineffective and

41:20

what marketers are getting wrong. Listen

41:22

to the nudge wherever you get

41:24

your podcast.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features