Reid riffs on antitrust, AI coworkers, and Chinese manufacturing

Reid riffs on antitrust, AI coworkers, and Chinese manufacturing

Released Wednesday, 30th April 2025
Good episode? Give it some love!
Reid riffs on antitrust, AI coworkers, and Chinese manufacturing

Reid riffs on antitrust, AI coworkers, and Chinese manufacturing

Reid riffs on antitrust, AI coworkers, and Chinese manufacturing

Reid riffs on antitrust, AI coworkers, and Chinese manufacturing

Wednesday, 30th April 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:02

I'm Reid Hoffman. And I'm

0:04

Aria Finger. We want to

0:06

know what happens if, in

0:08

the future, everything breaks humanity's

0:11

way. With support from Stripe,

0:13

we typically ask our guests

0:15

for their outlook on the

0:17

best possible future. But now,

0:19

every other week, I get

0:21

to ask Reid for his

0:23

take. This is possible. So

0:28

read, there have been a lot

0:30

of headlines lately on antitrust

0:32

and a federal judge ruled

0:34

that Google had illegally monopolized

0:36

specific online advertising markets. The

0:38

DOJ is seeking the divestiture

0:40

of Google ads manager and

0:42

other assets to restore competition.

0:45

No surprise, Google is fighting

0:47

back. So happy to hear your

0:49

thoughts on this specific case, but

0:51

also more broadly, how will recent

0:54

rulings... reshape the digital landscape, or

0:56

influence the future of innovation and

0:58

competition in tech. So not surprising. Well,

1:00

actually, maybe surprising the anorro people

1:03

since I'm on the Microsoft board

1:05

and, you know, obviously have to

1:07

have a certain care and how

1:09

I talk about antitrust competitors with

1:11

Microsoft, etc., even though my actual... thing

1:13

is I'm a venture capitalist. I'm on

1:15

the side of scale tech. I'm on

1:18

the side of building new things. And

1:20

so that's where my interest in society,

1:22

my interest in economics, my interest in

1:24

how do we create a better world,

1:26

you know, kind of most align with

1:28

scale tech. So startups going to scale.

1:30

And so blocking out, the possibility

1:32

of scale is one of the

1:34

places where I think antitrust legislation

1:36

can be very good and important.

1:38

Now, that being said on the

1:40

Google specific case, there's a couple

1:43

of notes. So one, I think

1:45

probably the most robust note

1:47

is, you know, where high

1:49

economics are being used

1:51

to buy exclusive channels,

1:53

that's probably a pretty

1:55

good sign that something

1:57

is being done to law.

2:00

in a monopoly or build a monopoly,

2:02

the challenge in a lot of these

2:04

monopoly cases is what do you take

2:06

as the size of the market? For

2:09

example, do you take the size of

2:11

the market as the general search market,

2:13

in which case you go, okay, Google

2:15

is massively dominant, or do you take

2:18

it digital advertising? You include meta, you

2:20

include a whole bunch of other things.

2:22

And so, you know, part of the reason

2:24

why there is smart people arguing on

2:27

both sides as you get to a

2:29

artifact. of what is the comparable market

2:31

to determine this, I tend to look

2:34

at it as, well where is it

2:36

starting, where is it stopping potential

2:38

scale-up competition? And by the

2:40

way, it's not all scale-up

2:42

competition because you say, hey,

2:44

I would like to start

2:46

a new desktop search company

2:48

using the same techniques that

2:51

Google did to build it. That's

2:53

not clear that that's a benefit

2:55

to society to try to squash

2:58

Google enough. to allow random startups

3:00

or any large tech company to

3:02

come in and compete. Now

3:04

the last comment is one that

3:06

I think is maybe the most

3:08

unpopular, but I think it's important

3:10

to track, which is, you know,

3:12

as we move to a more

3:14

of a multipolar world, and, you

3:17

know, the kind of classic thing

3:19

is, for example, a US tech

3:21

industry, a Chinese tech industry, you

3:23

know, Tik, etc. And we say,

3:25

well, only the US is going to be...

3:27

doing monopolist remedies and

3:30

the others aren't, you have

3:32

to track this within national

3:34

competition. And so part of

3:36

the question is, is this a,

3:38

is this part of a global resorting

3:41

for competition? And that's

3:43

actually, in fact, extremely important.

3:45

Because while we definitely want, you

3:47

know, the next generation of companies

3:49

coming out of, you know, Silicon

3:51

Valley, where I'm... here I am

3:53

right now, you know, kind of

3:55

talking about this, but on the

3:57

other hand, of course, if you

3:59

say... well, the scale ones that have

4:01

the scale benefit, we're limiting ours, but

4:03

we're not limiting China's, we're

4:05

not limiting other prospective ones,

4:07

that could be damaging, not

4:10

just to obviously American industry

4:12

and American prosperity and American

4:14

society, but also, of course, damaging in

4:16

terms of the balancing of the world. So

4:18

you have to also pay attention to that,

4:20

and I think one of the things that

4:22

is too often not included in the considerations

4:25

in these cases. Actually, I'll say one other

4:27

thing, which is... There's always some politics

4:29

in this, even though the Google case

4:31

went across different administrations. Yeah, started

4:33

under Trump. Exactly, started under Trump,

4:35

continued under Biden, returned under Trump.

4:37

Yep. But there's always some political

4:39

considerations that are not necessarily red

4:41

versus blue. There's also the, do

4:43

I look like I have a

4:45

win because I was fighting like

4:48

I'm the anti-monopoly division, etc. So

4:50

you always have to pay some attention to

4:52

this. And so you tend to go after the

4:54

targets that have more of a populist or press

4:56

dent. I mean the most obvious one for me,

4:58

like if I was forced ranking all the ones

5:00

that I would consider, would actually in fact

5:03

be the Apple App Store, right, because

5:05

you're like, okay, this is hugely locked

5:07

in, you're not allowed to have other

5:09

App Store, you're not allowed, like it's

5:11

very highly controlled, right. And there's some

5:14

arguments around. hey we got to maintain enough

5:16

security and so forth but by the way that's

5:18

also part of how all of these things

5:20

are. But they're also taking a cut

5:22

of every transaction and it makes using

5:24

Amazon mobile apps sort of unusable because

5:26

you have to go off-app to buy

5:28

certain things and so it's actually not

5:30

good for consumers. Yeah well and also

5:33

like for example one tell that I

5:35

started with Google is say hey you're

5:37

spending massive amounts of money to lock in

5:39

a you know effectively exclusive position well

5:41

that's actually a tell but another tell

5:43

is you're charging 30% right off the

5:46

top to everybody and making a whole

5:48

bunch of money from that and it's

5:50

like okay well that's another tell so

5:53

those are kind of tells where you

5:55

say well those are things that that

5:57

you should examine carefully look at this

6:00

and consider doing remedies on. So

6:02

I guess the overall thing is

6:04

I think it's good to do

6:06

these things to enable scale competition,

6:08

but we don't want to lose

6:10

the kind of gems we have

6:12

as an American society. Like for

6:14

example, probably most Americans don't realize

6:16

that these companies, Google, Apple, etc.

6:18

get over half the revenue extremely.

6:21

Outside the US. They're one of

6:23

our, you know, massive trade benefit

6:25

companies. And you're like, okay. that's

6:27

important to us as a society.

6:29

Doesn't mean we shouldn't do antitrust

6:31

things, but it means that we

6:33

don't want to, that it's not

6:35

just the kind of the, you

6:37

know, hit hard with hammer as

6:40

a approach. It's, it's, it's, be

6:42

careful about maintaining the, the kind

6:44

of vigorous American tech industry strength

6:46

to our prosperity. Right, probably not

6:48

the best. And to your point,

6:50

like geopolitics does matter. And so

6:52

we can take that into account

6:54

when we're doing these things, but

6:56

there are certain tells for the

6:59

position. And so, you know, the

7:01

Chinese market continues to embrace AI

7:03

as an accelerator, and Chinese tech

7:05

companies are growing their global footprints.

7:07

We've seen that BYD, which is

7:09

China's leading electric vehicle manufacturer, is

7:11

rapidly expanding its footprint in Europe,

7:13

especially as Tesla really tumbles. And

7:15

so, especially with this increasing terror

7:17

regime. there are going to be

7:20

some real problems with US auto

7:22

manufacturers selling to Europe, which could

7:24

mean that China gets a greater

7:26

foothold there. What do you think

7:28

that most people are still getting

7:30

wrong in how they think about

7:32

competition with China? Well, there's a

7:34

couple things in terms of competition.

7:36

So for example, you know, already

7:39

in person what you said, one

7:41

of the real damaging to American

7:43

prosperity, American quality of life, both

7:45

in purchasing of things and in

7:47

jobs and everything else is Actually,

7:49

in fact, trade partners matter. And

7:51

so it's part of the reason

7:53

why there's been different trading blocks

7:55

and you and an after and

7:58

those trade blocks actually matter because

8:00

being part of them gives the

8:02

people in them. advantages and edges

8:04

against the people who are not.

8:06

And so when you say, hey,

8:08

I'm just going to go apply

8:10

terrorist everyone, and you know, and

8:12

the absurd thing is, I'm going

8:14

to apply terrorist to islands that

8:17

have no people on them and

8:19

penguins only, but you know, I

8:21

mean, that's just the incompetence part

8:23

of the of the whole clown

8:25

show. But when you start doing

8:27

that, you're going, okay, I'm going

8:29

to kind of declare a trade

8:31

war on everybody, whereas what you'd

8:33

want to be doing is saying,

8:36

And so by saying, hey, we're

8:38

going to declare trade aggression with

8:40

Canada, trade aggression with Europe, the

8:42

natural thing for Canada and Europe

8:44

to do is say, great, go

8:46

trade with China, thank you very

8:48

much. And your so-called compete with

8:50

China policy is literally a gift

8:52

to China. And by the way,

8:54

the BYD product is very good.

8:57

So I think this is something

8:59

that is highly harming of American

9:01

society that starts from the general

9:03

prosperity of our society to our

9:05

functioning of our industries to the

9:07

prices and engagement of consumers in

9:09

this. And I don't just mean

9:11

consumers in wealthy cities, I mean

9:13

across the entire country. And so

9:16

this is the thing that is

9:18

kind of... call it most obviously

9:20

wrong about thinking of competition as

9:22

China. Another interesting thing you said

9:24

was that the BYD product is

9:26

very good. And I think some

9:28

people's conception of the world is

9:30

still sort of China 10 to

9:32

15 years ago, where it's like,

9:35

oh, we're going to flood the

9:37

market with cheap, knock off Chinese

9:39

goods. Well, no, they're doing advanced

9:41

manufacturing. This isn't just, you know,

9:43

low level t-shirts. Yeah. We have

9:45

to think about the prosperity of

9:47

our society. Well, one of the

9:49

best in a vector, massively important

9:51

scale vector, one of the best

9:54

manufacturing capabilities, societies, cultures in the

9:56

world. Like when I go to

9:58

Shenzhen, or have gone to Shenzhen,

10:00

is the only place I've gotten

10:02

the experience where someone coming to

10:04

Silicon Valley must feel like, holy

10:06

shit, like I am seeing part

10:08

of the future in terms of

10:10

speed and how people are operating

10:13

and it's more manufacturing there. And

10:15

I think that the thing that

10:17

we don't realize is even though

10:19

they have this advantage, they're going

10:21

full-on AI robotics manufacturing. And by

10:23

the way, that's what we should

10:25

be doing too. Those will be

10:27

the new manufacturing jobs of the

10:29

future. you know, will be the

10:31

ones working in robotic factories. That's

10:34

I think really, really key. The

10:36

Chinese know that even though they

10:38

have an edge with all their

10:40

human labor right now, their building,

10:42

like BYD specifically is building, is

10:44

intensely roboticizing its thing. There's not

10:46

only going to have a high

10:48

quality product, it's going to be

10:50

able to produce it at half

10:53

the cost of, you know, any

10:55

other competitor. Let the reduction of

10:57

cost is not because of... we're

10:59

going to claim that it's unfair

11:01

competitive practices, but it's actually in

11:03

fact, well, we're just smarter about

11:05

how we build it. On this

11:07

podcast, we like to focus on

11:09

what's possible with AI, because we

11:12

know it's the key to the

11:14

next era of growth, a truth

11:16

well understood by Stripe, makers of

11:18

Stripe billing, the go-to monetization solution

11:20

for AI companies. Stripe knows that

11:22

when launching a new product, your

11:24

revenue model can be just as

11:26

important as the product itself. In

11:28

fact, every single one of the

11:31

Forbes top 50 AI companies that

11:33

has a product on the market

11:35

today uses Stripe to monetize it.

11:37

See what Stripe can do for

11:39

your business at stripe. Switching

11:44

gears a little bit to what I'll

11:47

call the cost of being polite to

11:49

AI. So Open AI CEO Sam Maltman

11:51

recently admitted that the addition of words

11:54

and phrases like please and thank you

11:56

in users interactions with chat TV. at

11:58

a real cost, he tweeted that these

12:00

pleasantries contribute to tens of millions of

12:03

dollars in electricity costs for open AI

12:05

each year. And so there's one thing

12:07

about the dollar costs, but a lot

12:10

of people are really concerned about the

12:12

environmental implications in particular. Altman said that

12:14

expenditures like this, tens of millions of

12:17

dollars, well spent. So question for you,

12:19

do you think it matters, like philosophically,

12:21

how polite we are when we communicate

12:23

with AI? Well, yes, but maybe not

12:26

for the reasons that people might reflexively

12:28

think, which is more about, like, when

12:30

we're interacting with AI, it also involves

12:33

us. It also involves how we behave

12:35

with ourselves, with ourselves, with other people,

12:37

not just with devices. It's actually one

12:40

of the things I was always worried

12:42

about, how the initial Alexa home applications

12:44

were creating bad training for children, or

12:46

even adults who are not paying attention.

12:49

you know, now already do this, is

12:51

like what? That's not the way we

12:53

should be interacting with each other, it's

12:56

not the way we should be thinking

12:58

about it, we should be kind of

13:00

generally more civil, you know, politeness is

13:02

actually in fact I think a good

13:05

thing and that's worth it, let alone

13:07

the question around, well what outputs do

13:09

we get? Because by the way, you

13:12

know, people who are deep studies of

13:14

the prompting, you know, part of, you

13:16

know, why... release the earlier book in

13:19

prompt-to, people who are deep studies this,

13:21

notice they actually get different responses from

13:23

please and thank you and so forth,

13:25

because in part, this is generalized from

13:28

a trillion plus words of human communication,

13:30

and you're prompting when you say please

13:32

or not, you're telling it a little

13:35

bit about what kind of interaction you

13:37

want from it, you're having from it,

13:39

etc. And so... it's actually a useful

13:42

part of the prompt too. Now I

13:44

think part of what Sam's talking about

13:46

here is, if you go, wow, it

13:48

was such a great conversation, thank you

13:51

so much, and I really appreciated it,

13:53

and you're not actually building something else.

13:55

then that's good for you as per

13:58

my earlier comments. Yeah. And I think

14:00

it's a good pattern to be in,

14:02

but on the other hand, you're not

14:05

getting anything out of it. And the

14:07

electricity is being spent. It's like leaving

14:09

the light on for an hour. Totally.

14:11

Maybe you're gonna walk into the room.

14:14

And you know, so you could decide

14:16

to be a little bit less cautious

14:18

there, but I would err on the

14:21

side of politeness. I feel like AI

14:23

might be the new waiter test. It's

14:25

like it used to be when you

14:27

went on a date with someone and

14:30

they were rude to the waiter. That

14:32

was like the ultimate red flag. And

14:34

so the new red flag will be

14:37

like, ah, I really like them. They

14:39

were so rude to their AI. I

14:41

just, I couldn't get past it. So

14:44

maybe that will be in modern dating.

14:46

We'll see. But you mentioned something how.

14:48

Saying please and thank you actually couldn't

14:50

be a good form of prompting. You're

14:53

going to get something better. So sort

14:55

of analogous to that. Do you think

14:57

that open AI, anthropic, pie, etc. Should

15:00

they be training people on how to

15:02

prompt better themselves? It's like a big

15:04

conversation, of course. How do you get

15:07

the best prompts? But should the frontier

15:09

models be doing that for their users?

15:11

I mean, it's always helpful to do

15:13

it. I do think learning how to

15:16

prompt well is really key. I do

15:18

think it's really critical for people to

15:20

be learning how to do this prompting

15:23

in good ways because part of the

15:25

whole AI amplifying humanity, the amplification intelligence,

15:27

is the theory that we can by

15:30

us bringing something fun and interesting and

15:32

unique and you know, kind of adapting

15:34

to using these tools at the table

15:36

that actually in fact we are much

15:39

stronger together than being replaced in the

15:41

work. And this is an area of

15:43

active debate about, you know, within the

15:46

both general work community but also the

15:48

tech community as well, where over time

15:50

are that line of transformation versus replacement

15:52

B. And, you know, I don't know.

15:55

No one really knows like the claim

15:57

that you know that that for sure,

15:59

other than there will be some replacement,

16:02

is foolish. Because we know that we

16:04

have some replacement, like, for example, customer

16:06

service jobs with, you know, Sierra and

16:09

others doing this. But, well, like, for

16:11

example, a hot debate, software engineers, will

16:13

software engineers be amplified, which I think

16:15

is actually, in fact, myself more likely

16:18

the case, or would they be replaced,

16:20

because we're definitely getting... you know, kind

16:22

of higher quality with the kind of

16:25

chain of thought models and all the

16:27

rest to things that could lead to

16:29

better coding. Everyone's working and coding assistance.

16:32

And I think what we want is

16:34

we want the maximum probability chance of

16:36

all the jobs that people want to

16:38

do or have any affinity for doing

16:41

are transformation jobs, not replacement jobs. Now

16:43

part of that is how we're building

16:45

the technology. I'm not actually an advocate

16:48

of limiting the power and scope of

16:50

the technology because you go... Yeah, fine.

16:52

You know, make the cars really slow,

16:55

so humans can outrun them. It's like,

16:57

no, that doesn't really work as a

16:59

strategy. But the nudges to say, hey,

17:01

A and B are both performing systems,

17:04

but A allows a much better partnering

17:06

with human capability to get a much

17:08

better output. That's good for the individual,

17:11

good for society, etc. And that gets

17:13

all the way back to your question.

17:15

Well, people've got to be learning. to

17:17

use the devices better, if you're going

17:20

to be learning to prompt better. And

17:22

it's one of the reasons why I

17:24

love Ethan Mulloch's work. It's part of

17:27

the reason why, you know, I myself,

17:29

like, you know, when I was on,

17:31

Dak Shepard's podcast, you know, armchair expert,

17:34

I was like, okay, let me pull

17:36

out my phone and let me show

17:38

you how to prompt the phone to

17:40

start doing this in ways that are

17:43

useful to you, because that's actually, in

17:45

fact part of how part of how

17:47

we all shaped the future, What would

17:50

you say though, just last week, the

17:52

news from Dario, the CEO of Antropic,

17:54

was that he said that they'd be

17:57

rolling out AI... coworkers as soon as

17:59

2026. You think that's exaggeration? You think

18:01

they're going to be doing parts of

18:03

roles? Or do you think that's real?

18:06

That we're going to have AI coworkers

18:08

coming as soon as, you know, a

18:10

year from now? Well, it depends. Co-worker

18:13

can mean a lot of different things.

18:15

So in a sense, with a fuzzy

18:17

definition of co-worker, I think it can

18:20

be a co-pilot, yeah. Yes, exactly. Like

18:22

we have co-pilas today, right. Right. Right.

18:24

and not a surprise, co-pilots will continue

18:26

to improve. And what I think he

18:29

also means is, unlike the, hey, I

18:31

am helping guide each step, I might

18:33

be able to just like chain of

18:36

thought thinking, you know, with the one

18:38

models and others, is I send it

18:40

out on a task like a set

18:42

of work, and it comes back with

18:45

all the work done, right? Like it

18:47

kind of, it guided through it, changes

18:49

plans, some, etc, etc, etc. And that's

18:52

obviously what we see being developed being

18:54

developed already. You say, well, that's a

18:56

co-worker. That's more robust than the current

18:59

co-pilots. But it's also not a co-worker

19:01

in that the, you know, hey Ari,

19:03

I've got this really cool project. Hail,

19:05

I'll talk to you in two or

19:08

three weeks about it. You know, and

19:10

you're off assembling resources doing all this

19:12

stuff. So it's like, this is a

19:15

whole continuum. So will we be advancing

19:17

the continuum? Guarantee. Is it? a capable

19:19

independent agent in the way that a

19:22

human being is a capable independent agent.

19:24

It has context awareness, can change its

19:26

goal sets, can remake plans and triage

19:28

based on new data, can defend itself.

19:31

It's like, well, no, I are asking

19:33

to work on this book thing, and

19:35

you came back with this really interesting

19:38

art project. Like, why is that? Is

19:40

like, well, no, no, but this is

19:42

the reason I only we were talking

19:45

about, like what we were trying to

19:47

accomplish. This suddenly turned into the really

19:49

interesting thing and I changed my mind.

19:51

I didn't just listen to you. I

19:54

like based on my own goals and

19:56

what I know about's going on, right?

19:58

You have some agency. Yes, and so

20:01

the question is, is, were we heading

20:03

towards that? And the answer is more,

20:05

but how much more? And I tend

20:07

to think that these is kind of

20:09

two theories of kind of what is

20:12

the next, call it five, ten, twenty

20:14

years of agents look like, and 20

20:16

is like, you know, I'm possibly long

20:18

in these things. Most people are trying

20:20

to talk two years. But it's like,

20:22

well, is it a progressing set of

20:25

savants where it does amazing,

20:27

amazing things? But part of the reason

20:29

why you stay close to it is

20:31

because occasionally it fucks up in like stunning

20:34

ways that literally, like if a human

20:36

did it, you'd be like, what were

20:38

you thinking? Like what happened? I mean,

20:40

as you often say, like predicting the

20:42

future, even a year or two out

20:44

right now with AI is pretty impossible.

20:46

Which by the way, just

20:48

for everyone else. a lot

20:50

of people's natural responses to

20:53

go, oh shit, you can't

20:55

predict the year out, that's

20:57

terrifying. Hence, all the superagency,

20:59

you know, doomer, glomer, but

21:02

actually in fact, it's super

21:04

interesting and we can help

21:06

shape it. That's what's great,

21:09

like, you know, navigate

21:11

risk concern, but it's

21:13

exciting as well. Absolutely.

21:15

It's produced by Wonder Media

21:18

Network. It's hosted by R.

21:20

Effinger and me, Reed Hoffman.

21:22

Our showrunner is Sean Young.

21:24

Possible is produced by Katie

21:26

Sanders, Edie Allard, Sarah S.

21:29

Lied, Vanessa Handy, Alia Yates,

21:31

Palomo Moreno Jimenez, and Malia

21:34

Agudello. Jenny Kaplan is our

21:36

executive producer and editor. Special

21:38

thanks to Suria Yala Manchili,

21:41

Saeda Sepia Sepieva, Thanasi Delos,

21:43

Ian Alice, Greg Vigbyato, Park

21:46

Patil, and Ben Relis.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features