Google Cloud Next is All About Agents [Shocker!]

Google Cloud Next is All About Agents [Shocker!]

Released Friday, 11th April 2025
Good episode? Give it some love!
Google Cloud Next is All About Agents [Shocker!]

Google Cloud Next is All About Agents [Shocker!]

Google Cloud Next is All About Agents [Shocker!]

Google Cloud Next is All About Agents [Shocker!]

Friday, 11th April 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

Today on the AI Daily Brief,

0:02

Entropic introduced

0:04

their power user tier. Before

0:06

that in the headlines, Google

0:09

Next is all about agents.

0:11

The AI Daily Brief is

0:14

a daily podcast and video

0:16

about the most important

0:18

news and discussions in

0:20

AI. To join the

0:23

conversation, follow the Discord link

0:25

in our show notes. First

0:27

up, as I've been mentioning a couple times,

0:29

for those of you who are looking for

0:32

an ad-free version of the A.I. Daily Brief,

0:34

you can now head on over to patron.com/A.I.

0:36

Daily Brief to find that. Second, next week

0:38

is spring break, so you will not not

0:40

have shows, but they will be a little

0:43

bit different than normal. Lastly, something I

0:45

want to gauge people's perspective on. The

0:47

A.I. Daily Brief community has been hugely

0:49

supportive of and important in the superintelligent

0:51

story. We're. However, I'm trying to gauge

0:53

interest. If this is something you think

0:55

we should explore, send me a note

0:58

at NLW at B super dot AI

1:00

with super in the title. Thanks in

1:02

advance for your perspective. And with that,

1:04

let's get into today's show. Welcome back

1:06

to the AI Daily Brief Headlines edition, all the

1:08

Daily AI News you need in around five minutes.

1:11

We kick off today with Antropic getting themselves

1:13

a power user tier. Now, you might

1:15

remember when Open Open AI introduced their

1:17

200 dollar a month tier back in

1:19

December, many believe that the demand just

1:21

wouldn't be there. However, while we don't know how

1:24

popular exactly that subscription is, it's generated enough

1:26

usage that Sam Altman said that the company

1:28

was actually losing money on the deal. Entropic

1:30

is calling their version of the premium subscription Claude

1:32

Max. The product allows users to pay to

1:34

get around Entropic's to pay to get around

1:36

Entropic's notoriously troublesome rate limits, gets priority responses

1:38

during heavier traffic periods, and early access to

1:41

new features thrown in as a sweetener. If

1:43

you've spent any time on AI Twitter, you'll

1:45

know that people have been constantly asking Entropic for the

1:47

ability to pay more for more for more for more

1:49

for more for more for more for more for more

1:51

service. There's actually two different levels of

1:53

Claude Max. For $100, users get five times

1:56

the rate limits of the $20 a month

1:58

pro tier. That math works out. But for

2:00

those users who really burn through tokens, they'll

2:02

want the $200 level that allows for 20

2:04

times the rate limit. What's not on

2:07

offer, unfortunately, for some is unlimited

2:09

usage. So is this about boosting revenue or

2:11

simply recognizing that power users are underserved?

2:13

As they say in the business, why

2:15

not both? Anthropics product lead Scott White

2:18

said that the company isn't ruling out

2:20

adding even more pricey subscriptions, saying we

2:22

could see a $500 a month level.

2:24

Ultimately, he said that the product roadmap

2:26

is guided by user feedback. And the

2:28

one loud, consistent piece of user feedback we've

2:30

seen over the past year is that power

2:32

users want to pay for more usage. Next

2:34

up, a little more on the terror fallout, Invidia

2:36

has secured a carve-out for their

2:39

China-based chips. Industry insiders had widely

2:41

expected the administration to clamp down on

2:43

export of H20 chips, which is the

2:45

downgraded GPUs that are designed to get

2:48

around export controls. NPR is reporting that

2:50

the additional restrictions won't go ahead after Jensen

2:52

Huang attended a dinner at Maralaga last week.

2:54

Sources said that restrictions had been in the

2:57

works for months and were ready to be

2:59

implemented as soon as this week. They said that the

3:01

president changed his mind after Huang promised new

3:03

data center investment in the US. Chris Miller

3:05

at Tufts University History Professor and semiconductor

3:07

expert commented, even though these chips are

3:10

specifically modified to produce their performance, thus

3:12

making them legal to sell to China,

3:14

they are better than many perhaps most

3:16

of China's homegrown chips. China still can't produce

3:18

the volume of chips it needs domestically, so

3:20

it is critically reliant on imports of invidious

3:23

chips. Then again, that view is a little up

3:25

in the air after recent reports of efficient training

3:27

runs on new Huawei chips. However, even if

3:29

the Chinese AI industry is trying to wean

3:31

themselves off of invidia, the market is still

3:33

critical for the dominant chipmaker. Thirteen percent of

3:36

invidious official demand comes from China, and that

3:38

figure could be much higher if you account

3:40

for evasion of export controls through Southeast Asia.

3:42

All in all, it adds up to us

3:44

continuing to not have a coherent picture of

3:46

the administration's strategy when it comes to chip

3:48

controls. Following inauguration, Trump pledged to wind

3:50

back many of the restrictions on the AI

3:53

industry. However, the enhanced export controls introduced in

3:55

the final weeks of the Biden presidency are

3:57

still in place. These regulations put limits...

3:59

on a huge portion of the world

4:02

including friendly countries like Israel and India.

4:04

Then again this does also seem like reinforcement

4:06

of the idea that there's always a deal

4:08

to be made when it comes to Trump.

4:10

Now over the last couple of days you

4:12

may have heard me wax poetic around what

4:14

I think the implications of some of this

4:16

tariff stuff are likely to be on venture

4:19

capital. Nominally I think that it's going to

4:21

be a harder fundraising environment not only for

4:23

startups but also for VCs themselves, but

4:25

countering that point appears to be Andrei

4:28

and Horowitz and Horowitz. Sources said

4:30

the firm is looking to capitalize

4:32

on high international demand for investments

4:34

in American companies. They added that international

4:36

LPs view the fund as a way to

4:38

more easily invest money in the USAI sector

4:41

without the restrictions. So it sounds like this

4:43

actually might be playing into and taking advantage

4:45

of some of the tariffs. Last year A16Z

4:47

raised 7.2 billion scattered across the

4:49

themes of American dynamism, apps, games,

4:51

infrastructure and growth. This fund then is

4:54

both significantly larger and more focused

4:56

than previous efforts. The gigantic size

4:58

brings up questions about whether venture capital

5:00

can scale up in this rarefied air.

5:02

Softbank is perhaps the obvious comparison. They

5:05

raised their $100 billion vision fund in

5:07

2017 with very mixed results. The second

5:09

vision fund raised in 2019 is a

5:11

relatively more modest but still massive 56

5:13

billion. The other comp that springs to mind

5:16

is Sequoio, who currently manage over 56

5:18

billion in assets overall. Still there hasn't

5:20

been a venture strategy as capital intensive as

5:22

AI in the past. In fact, part of

5:24

the reason that companies like Open AI had

5:26

to turn to big tech partners like Microsoft

5:28

is that there simply wasn't enough dry powder

5:30

in the venture capital coffers for them to

5:33

get what they needed. Reuter sources said that

5:35

a significant portion of the fund would

5:37

be set aside for follow-on investments in

5:39

companies already in A16Z's portfolio. And

5:42

with portcos like Mistral and Safe Super Intelligence

5:44

and data bricks, there is a lot of

5:46

money to be spent. There is certainly a lot

5:48

of capital need. Lastly today, a little

5:50

bit of institutional psychodrama. Open AI has countersued

5:52

Elon Musk asking the court to bring an

5:55

end to the billionaire's legal challenge. The court

5:57

filing called for Musk to be prohibited from

5:59

taking... quote, further unlawful and unfair action

6:01

and held responsible for the damage he

6:04

has already caused. It stated, Open AI

6:06

is resilient, but Musk's actions have taken a

6:08

toll. Should his campaign persist, greater harm is

6:10

threatened to Open AI's ability to govern in

6:13

service of its mission to the relationships that

6:15

are essential to furthering that mission and the

6:17

public interest. Musk's continued attacks

6:19

on Open AI culminating most recently in

6:21

a fake takeover bid designed to disrupt

6:23

Open AI's disrupt Open AI's future must

6:25

cease. Musk's attorney immediately fired back in

6:28

a press statement he said. Had Open AI's

6:30

board genuinely considered Musk's bid as they were

6:32

obligated to do, they would have seen how

6:34

serious it was. It's telling that having to

6:36

pay fair market value for Open AI's assets

6:38

allegedly interferes with their business plans. The case

6:40

is currently at a slow point as the parties

6:43

await a jury trial next spring. Musk's attempt

6:45

to seek an injunction to stop Open AI

6:47

from converting to a non-profit was rejected in

6:49

March. So technically there isn't anything

6:51

stopping California Attorney General Rob Bonsa

6:53

from making a decision on the

6:55

conversion, however complaints continue to roll

6:57

in and the active litigation

7:00

gives him a good excuse to delay.

7:02

Meanwhile, of course, Open AI has a

7:04

huge financial incentive to get this wrapped

7:06

up quickly. The company's latest fundraising round

7:08

featured $10 billion from Soft Bank that

7:10

is contingent on the conversion being completed

7:12

by the end of the year. That friends

7:14

is going to do it for today's Aye

7:17

Daily Brief Headlines Edition, next up the main

7:19

episode. In our fast-paced world, having

7:21

a solid AI plan can

7:23

make all the difference. Enabling

7:25

organizations to create new value,

7:28

grow, and stay ahead of

7:30

the competition is what it's

7:32

all about. KPMG is here to

7:34

help you create an AI strategy

7:37

that really works. Don't wait. Now's

7:39

the time to get ahead. Check

7:41

out real stories from KPMG of

7:44

how AI is driving success

7:46

with its clients at KPMG.US.

7:48

US slash AI. Today's episode is brought

7:50

to you by Vanta. Vanta is a trust management

7:52

platform that helps businesses automate

7:54

security and compliance, enabling them

7:57

to demonstrate strong security practices

7:59

and scale. In today's business landscape,

8:01

businesses can't just claim security. They have

8:03

to prove it. Achieving compliance with a

8:05

framework like SOC 2, ISO 2701, HIPAA,

8:07

GDPR, etc. is how businesses can demonstrate

8:10

strong security practices. And we see how

8:12

much this matters every time we connect

8:14

enterprises with agent services providers

8:16

at Superintelligent. Many of these

8:18

compliance frameworks are simply not

8:21

negotiable for enterprises. The problem

8:23

is that navigating security and compliance

8:25

is time-consuming and complicated. It can take

8:27

months of work and use up valuable

8:29

time and resources. Vanta makes it easy

8:31

and faster by automating compliance across 35

8:33

plus frameworks. It gets you audit ready

8:36

in weeks instead of months and saves

8:38

you up to 85% of associated

8:40

costs. In fact, a recent IDC white

8:42

paper found that Vanta customers achieve $535,000

8:44

per year in benefits, and the platform

8:46

pays for itself in just three months.

8:48

The proof is in the numbers. More

8:51

than 10,000 global companies trust Vanta,

8:53

including Atlian, Cora, Cora, and more. For

8:55

a limited time, listeners get

8:57

$1,000 off at vanta.com/

8:59

NLW. That's va-n-t-a.com/NLW for $1,000 off.

9:02

Today's episode is brought to you

9:04

by Super Intelligent and more specifically

9:06

Super's Agent Readiness Audits. If you've

9:09

been listening for a while, you

9:11

have probably heard me talk about

9:13

this. But basically the idea of

9:15

the agent readiness audit is that

9:18

this is a system that we've

9:20

created to help you benchmark and

9:22

map opportunities in your organizations where

9:25

agents could specifically help you solve

9:27

your problems, create new opportunities in

9:29

a way that again is completely

9:31

customized to you. When you do one

9:33

of these audits, what you're going to

9:35

do is a voice-based agent interview where

9:37

we work with some number of your

9:39

leadership and employees to map what's going

9:41

on inside the organization and to figure

9:44

out where you are in your agent

9:46

journey. That's going to produce an agent

9:48

readiness score that comes with a

9:50

deep set of explanations, strength, weaknesses,

9:52

key findings, and of course, a

9:54

set of very specific recommendations that then

9:56

we have the ability to help you

9:58

go find the right part. to actually

10:00

fulfill. So if you were looking for

10:02

a way to jumpstart your agent strategy,

10:04

send us an email at agent at

10:06

b-super dot AI, and let's get you

10:08

plugged into the agentic era. Welcome back

10:10

to the AI Daily Brief. In a

10:12

massive shock to literally no one paying

10:15

any attention to this space, Google Next

10:17

is all about agents. Yes, friends, this

10:19

week we are headed into the next

10:21

round of big tech conference season. And

10:23

unsurprisingly, Google's Cloud Next conference featured a

10:25

huge lineup of AI announcements designed to

10:27

all in some, I think, make the

10:29

technology feel and be more useful. The

10:31

company's annual cloud conference was held in

10:33

Las Vegas this week and was absolutely

10:35

squarely focused on taking the next step

10:37

in the AI race. We've got agentic

10:39

infrastructure, new models, a new AI chip,

10:42

and much, much more. A pair of

10:44

announcements about agentic infrastructure could end up

10:46

having the most impact of all. Back

10:48

a couple of weeks ago on March

10:50

30th, in the wake of Open AI

10:52

announcing that they were going to support

10:54

MCP or the Model Context Protocol, Google

10:56

CEO Sundar Pichai tweeted to MCP or

10:58

not to MCP, that's the question. Let

11:00

me know in the comments. One point

11:02

eight million views and a thousand comments

11:04

later, we got the answer and it

11:06

was a yes. Deep Mine CEO Demiss

11:08

Asavis tweeted, MCP is a good protocol

11:11

and it's rapidly and it's rapidly becoming

11:13

and SDK. look forward to developing it

11:15

further with the MCP team and others

11:17

in the industry. And to reiterate, this

11:19

means that all three of the leading

11:21

US labs are now supporting MCP. Now

11:23

for those of you who don't know

11:25

what the heck I'm talking about, we

11:27

covered MCP in depth a couple weeks

11:29

back, but in short, if you're just

11:31

trying to understand the implications, it means

11:33

a significant boost for interoperability and compatibility

11:35

across the rapidly developing agentic infrastructure layer.

11:38

The second big agentic announcement was the

11:40

unveiling of Google's agent development kit and

11:42

a new interoperability standard called agent to

11:44

agent. As the name suggests... the standard

11:46

is seeking to harmonize the way agents

11:48

communicate with each other rather than how

11:50

they interact with tools. Rouse, who are

11:52

up in any VP of Google's cloud

11:54

business application platform, insisted that A2A is

11:56

not competing with MCP, which of course

11:58

is more about tool use. They said,

12:00

we see MCP and A2A as complementary

12:02

capabilities. The way we are looking at

12:05

agent to agent is at a higher

12:07

layer of abstraction to enable applications in

12:09

agents to talk to each other. So

12:11

think of it as a layered stack

12:13

where MCP operates for tools and data.

12:15

Words, words, words, but basically the point

12:17

is that they're trying explicitly to not

12:19

compete with MCP here, but instead be

12:21

a standard for something different. H2A has

12:23

50 companies on board to support the

12:25

open standard, including Salesforce, service now, and

12:27

workday. Sir Apanini said that Google isn't

12:29

necessarily looking to compete with other consortiums

12:32

working on their own solutions, saying, we

12:34

will look at how to align with

12:36

all of the protocols. There will always

12:38

be some protocol with a good idea,

12:40

and we want to figure out how

12:42

to compete with a good idea, and

12:44

a good idea, and a good ideas,

12:46

Standard ways for agents to coordinate could

12:48

reduce the amount of complexity, for example,

12:50

in multi-agent systems, could mean that agents

12:52

talk to their counterparts and other companies,

12:54

making them more capable of getting work

12:56

done without getting humans involved. So what

12:59

is it all amount to? Well, MIT

13:01

PhD Tobin-South says, my take on Ado

13:03

away from Google and friends is that

13:05

they're trying to create a communications hierarchy

13:07

with MCP as tool use and A-to-A

13:09

as coordination and communication. Hubspot founder and

13:11

now agent.a creator Darmes Shah writes, shockingly

13:13

this doesn't change everything. It's very very

13:15

early, but here are my initial thoughts.

13:17

I'm a big believer in multi-agent networks

13:19

in agent-to-agent communication. It's good that there's

13:21

now an open standard out there for

13:23

it. They cover some very key needs.

13:26

Capability discovery, agents being able to send

13:28

messages to each other, being able to

13:30

work on tasks that are long-lived, to

13:32

a-sink, weaving human UX into the agentic

13:34

flow, etc, etc. This is not a

13:36

replacement for MCP. In fact, during their

13:38

announcement post, Google included a helpful diagram

13:40

that illustrates how EDA and MCP fit

13:42

together. Still, this feels a bit... heavy

13:44

to me. It's trying to do a

13:46

lot. In a way that's good because

13:48

you get a bunch of capabilities out

13:50

of the box like Asing tasks and

13:53

user experience negotiation, but the trade-off is

13:55

that heavier protocols are harder to implement

13:57

and as such you don't get the

13:59

quick adoption you see with lighter weight

14:01

things, so I don't anticipate MCP-style adoption.

14:03

Reading between the lines, this feels like

14:05

they're solving for a lot of mega

14:07

enterprises and big consulting firms looking to

14:09

build multi-agent systems inside the corporation. It's

14:11

less about connecting agents' will be interesting

14:13

to see actual usable implementations of this

14:15

outside the Fortune 1000 companies. Overall, this

14:17

is good news though, moves us further

14:20

down the multi-agent systems road. Other random

14:22

little agentic notes, Gemini Code Assist is

14:24

getting an agentic upgrade. This is their

14:26

cursor competitor nominally and it can now

14:28

deploy agents to complete complex programming tasks

14:30

across multiple steps. This kind of agentic

14:32

feature has been a game changer on

14:34

other platforms with programmers using it to

14:36

automate repetitive tasks like code migration. Google

14:38

has also released a security agent as

14:40

part of their new unified security platform.

14:42

The goal is to have an agent

14:44

on the beat that can recognize and

14:46

remediate threats before they become major problems.

14:49

The Chief Information Security Officer at Charles

14:51

Schwab, Bashar Abbasido said, Google is transforming

14:53

security operations and enabling our vision to

14:55

stay proactive in responding to cyber threats.

14:57

The platform has empowered our team to

14:59

focus on strategic initiatives and high-value work.

15:01

We didn't get a massive new model,

15:03

but what we did get was Gemini

15:05

2.2. Like its predecessor, this is a

15:07

smaller model designed to deliver efficient performance

15:09

with extremely low latency. The big change

15:11

is a lot more customization. Google wrote,

15:13

you can do the speed, accuracy, and

15:16

cost balancing for your specific needs. This

15:18

flexibility is key to optimizing flash performance

15:20

and high volume cost sensitive applications. Gemini

15:22

2.5 flash is a reasoning model at

15:24

launch and will likely end up being

15:26

the cheapest on the market. The model

15:28

is designed to adjust the depth of

15:30

reasoning based on the complexity of the

15:32

prompt, which is similar it seems to

15:34

the approach that Open AI is pursuing.

15:36

Google is aiming to provide a model

15:38

that hits the sweet spot between performance

15:40

and cost writing. It's the ideal engine

15:43

for responsive virtual assistance and real-time... summerization

15:45

tools where efficiency at scale is key.

15:47

Google also announced that they plan to

15:49

bring Gemini models to on-premise deployments starting

15:51

in Q3. On the chip side, Google

15:53

has announced the seventh generation of their

15:55

tensor processing unit, which they're calling Ironwood.

15:57

Now, unlike GPUs, are specifically designed for

15:59

AI computing tasks. GPUs, meanwhile, are more

16:01

generalized across all mathematical functions. In fact,

16:03

in the early days, it was a

16:05

slight coincidence that the chip architecture that

16:07

powers 3D gaming was also highly performed

16:10

for AI. The core bet with developing

16:12

TPUs instead of GPUs was that specially

16:14

designed architecture would be more efficient. So

16:16

far that theory hasn't totally played out

16:18

with invidious GPU architectures still at the

16:20

top of the pack. However, Google is

16:22

hoping that Ironwood might finally validate that

16:24

thesis. The company claims that their new

16:26

processor can deliver 24 times the computing

16:28

power of the world's fastest supercomputer when

16:30

deployed at scale. Previous generations of the

16:32

hardware were designed for inference. I'm in

16:34

Vadat, Google's vice president of machine learning

16:37

said, Ironwood is built to support this

16:39

next phase of generative AI and its

16:41

tremendous computational and communication requirements. This is

16:43

what we call the age of inference,

16:45

where AI agents will proactively retrieve and

16:47

generate data to collaboratively deliver insights and

16:49

answers, not just data. Now there are

16:51

a bunch of numbers and specs that

16:53

come with this thing, but rather than

16:55

try to explain an exa-flop, the promise

16:57

here is that Ironwood is around twice

16:59

as in invidious age 100. Still the

17:01

biggest difference is actually scale. A maximum

17:04

size pod of Blackwell B200 chips is

17:06

576 chips before they require outside networking,

17:08

while Ironwood claims to be capable of

17:10

being deployed in a 9, 216 chip

17:12

pod. This is also a massive jump

17:14

from Google's previous generation of TPUs, which

17:16

were called Trillium. The company has achieved

17:18

a fourfold increase in computing power compared

17:20

to the 2024 model. Ironwood is also

17:22

more efficient, delivering twice the performance per

17:24

watt compared to Trillium. And this TPU

17:26

and the focus on efficient inference definitely

17:28

suggests that Google is scaling up to

17:31

service compute-hungry reasoning models and the... agents

17:33

by them. Now there is

17:35

more going is more

17:37

going on In a well.

17:39

of In a big

17:41

vote of confidence

17:43

for Google's Silicon, Ilia announced

17:45

that his startup

17:47

will use Google will use

17:49

Google Samsung announced that

17:51

Gemini would be

17:53

added to their be added

17:55

to their new Google's Google's

17:58

platform now features a

18:00

music generation model.

18:02

Ultimately, while there wasn't

18:04

one big huge

18:06

thing, at least not

18:08

a hitch you

18:10

over the head, over the

18:12

head. a pretty remarkable

18:14

shift from from black

18:16

glue on pizza on

18:18

all of this in

18:20

just about a

18:22

year. a year. AI for success writes,

18:25

Google Deep Mind is destined

18:27

to win the win

18:29

the AGI why, they

18:31

have the data advantage,

18:33

they own the they

18:35

own distribution channel, have all the

18:37

best models right now, they have

18:39

everything and they dominate all four

18:41

key areas, and they foundation models, key accelerator,

18:43

hardware. I'm not sure how

18:46

this all shakes out, but if you're

18:48

Google, you gotta be happy that that is

18:50

the narrative among many these days. Google, That's

18:52

gonna do it for today's AI Daily Brief. the

18:54

Appreciate you listening or watching as always That's

18:56

until next time, peace. AI Daily Brief. Thank

19:02

you.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features