Episode Transcript
Transcripts are displayed as originally observed. Some content, including advertisements may have changed.
Use Ctrl + F to search
0:00
Hey upstream listeners, today we're
0:02
re-releasing a conversation I had
0:05
last year with economist Tyler
0:07
Cowan when he released his book, Goat,
0:09
who is the greatest economist
0:11
of all time and why does it
0:14
matter? We discussed the book, talent
0:16
spotting, stagnation, fertility, AI, and more.
0:18
Please enjoy. Tyler,
0:34
the latest book is The Greatest Economist
0:36
of All Time, The Goat, welcome to
0:38
the podcast. Excited to chat again. Thank
0:40
you. Good to be here, Eric. Before we
0:42
get into it, what is your algorithm
0:44
for determining what to write? When you look
0:47
at the next few decades of your writing
0:49
career, I think you're over 30 books at
0:51
this point or something like that. How do
0:53
you think about what you just have a
0:55
big list of books you want to write?
0:57
Is it much more, you know, what at
0:59
a time? How do you think about that?
1:01
It's not rational. It is much fewer
1:03
than 30, though it's a fair number.
1:05
But this book has a specific story
1:07
that I was trapped at home during
1:10
the pandemic, couldn't travel, and
1:12
my libraries were closed. So
1:14
I thought, what's the thing I can work
1:16
on so I don't go crazy? And
1:18
to write about the great classics of
1:21
economics, which mostly I own or they're
1:23
online, you don't have to go anywhere.
1:25
That was the one project I could
1:27
do. So that's how it came about.
1:29
And I wasn't planning it at all, just
1:31
one day I woke up and well, I
1:33
want to do something. Yeah. Why does it
1:36
matter? That's part of the subtitle of your
1:38
book. Like even zooming out, when you think
1:40
about the impact that macro micro
1:42
micro has made, you know, we were
1:44
in a clubhouse a few years ago,
1:47
I was reviewing that conversation with Mark
1:49
Andreessen and we were talking about how.
1:51
you know, macro certainly hasn't been predictive
1:54
and you said it will likely never
1:56
be predictive. And so what is the
1:58
importance or what is is a
2:00
contribution to society that
2:02
understanding macro or contributes,
2:04
or macro, the field
2:07
within academic contributes? Well,
2:09
my book in particular, I take to be
2:11
a sequel to my talent book
2:13
with Daniel Gross. So here's case
2:15
studies of the very top economic
2:17
talents. What made them great? What
2:19
were the weaknesses? How do we
2:21
think about that? How can we
2:23
learn to appreciate them? So that's not
2:25
a gain from macro per se, but
2:27
you need to understand. macro to
2:30
evaluate these individuals. Now, if you're
2:32
looking at the global stage and
2:34
you want to figure out how
2:36
different economies will do, macro
2:39
has some uses. So for instance, we
2:41
were just both in Argentina at
2:43
the same time, and if someone
2:45
asks me the question, will Malay
2:47
succeed? I'm quite sure I can't
2:49
predict that. But I do have a
2:51
sense what variables it depends on, and
2:53
the notion that he needs to
2:56
raise about $30 billion to dollarize.
2:58
I understand how that works
3:00
because I understand macro. So there
3:02
is value to macro because Malay
3:04
understands macro. So far, he's done
3:06
some good things. He may or
3:08
may not continue on a good track.
3:10
So it is practical value, but
3:12
not for forecasting. Yeah. Just to
3:15
ask you to make a forecast, you did
3:17
just write a great post in Argentina.
3:19
If you had to guess what Malay's
3:21
legacy will be, what do you think?
3:24
What would you predict? My total seat
3:26
of the pants judgment is that
3:28
his chance of succeeding, even if
3:30
he does everything right, is somewhat below
3:32
50%. So his legacy is most
3:34
likely to be, here's the guy who came
3:37
in really knowing what needed to be
3:39
done, and he still failed because
3:41
the rest of the system wouldn't
3:43
cooperate. That said, I'm not a
3:45
total pessimist, and he has a good
3:47
chance of succeeding. If the other parties don't
3:49
just wake up and decide to pull
3:52
the plug on him, and send
3:54
him in their own country
3:56
to its doom, I think what
3:58
he's doing will work. Is there
4:00
an example of a president, the
4:02
different president of the last few
4:05
decades, who's been able to come
4:07
in and dismantle massive parts of
4:09
the administrative state or your sort
4:12
of bureaucratic machine? Well,
4:14
you mean only president's prime
4:16
ministers? Prime Ministers, yeah. Well,
4:18
New Zealand was a terribly status
4:20
country and they had a series
4:23
of governments, both labor party and
4:25
national party, that made it much
4:27
freer. I wouldn't say they dismantled.
4:29
the administrative state, and in
4:31
some ways they enabled it
4:33
to grow. But overall they moved
4:35
to much lighter and cleaner
4:37
regulation and better transparency, and the
4:40
country avoided bankruptcy because
4:42
of that. So that's one example.
4:44
Now the collapse of communism is
4:46
a more complicated story, but
4:48
there's definitely a kind of
4:50
regulation. They really did get rid of. Yeah.
4:53
Over the last few decades, has
4:55
the US become a more statist
4:57
or more libertarian? economy, or a
4:59
freer or more controlled? Well, the
5:01
economy is more controlled, obviously.
5:03
The American electorate has
5:05
shifted to the left. We redistribute
5:07
more income. We regulate much, much
5:10
more. The different layers of
5:12
regulation from federalism, they bite
5:14
more. But we're freer on many personal
5:16
issues, so we have gay marriage. And
5:19
if you do something with marijuana,
5:21
your chance of going to jail, I
5:23
guess it depends on the state, but
5:25
it's really very low. It might even
5:28
be legal or decriminalized. So a mix,
5:30
but in economics, not way more status.
5:32
I don't think that's disputable. Is
5:34
that true also in academia? Has
5:37
academic economists moved more to the
5:39
left as well? Because the academics
5:41
in general, I have a colleague
5:43
Daniel Klein who works on measuring
5:46
this, and I don't have his
5:48
numbers on the tip of my tongue, but there
5:50
are just very few right leaning
5:53
academics left, and a lot of them
5:55
are quite old. or not that
5:57
active, if you look at the
5:59
flow. It's a pretty grim picture.
6:01
And even to be centrist makes
6:04
you a right-winger in today's academic
6:06
world. In economics too, because libertarians
6:08
like to celebrate and say, hey,
6:11
I know we haven't won any
6:13
sort of formal political battles, but
6:15
they like to say that they've
6:17
won the war of ideas in
6:20
popularizing free market, ideas among
6:22
smart people. Is that cope? Or
6:24
meaning, is that untrue? Economics is
6:26
the sayenest field, but last I looked.
6:29
It's something like five-sixth
6:31
of economists with party
6:33
affiliations or Democrats. They tend
6:36
to be more conservative
6:38
or market-leaning Democrats. But on
6:40
a lot of issues, trade and
6:43
protection is a minimum wage,
6:45
yes, economists have shifted to
6:47
the left. Yeah. So you outlined here
6:49
in the book, I believe six economists
6:51
that you focus on. Why don't you
6:53
just as a survey of people who
6:56
haven't yet yet died into it? Give
6:58
a brief survey of who are the
7:00
economists that you you know attribute to
7:02
a tie of the greatest of all
7:04
time and what were their main sort
7:07
of contributions to take on your thought. Sure,
7:09
but let me first say the book is
7:11
online and it's free. So if you want
7:13
to read it you can just Google to
7:15
it Google Tyler Cowan goat, GOAT, GOAT.
7:17
You have a great GPT. Yeah. It comes
7:19
with an AI GPT four linked bot. which
7:22
will answer questions about the book,
7:24
you can just read it on your Kindle
7:26
or printed out. But if you just
7:29
want to ask GPT, summarize Chapter 3.
7:31
I don't want to read about Keynes,
7:33
it'll do it. But anyway, the figures
7:35
I cover are Milton Friedman, John
7:38
Maynard Keynes, Friedrich Hayek, John
7:40
Stewart Mill, Thomas Robert Malthus,
7:42
and of course, Adam Smith.
7:44
They're the contenders. Who are
7:47
the audible mentions? Paul Samuelson.
7:49
Kenneth Arrow, who are both
7:51
uncles of Larry Summers, by
7:53
the way, Joseph Shumpertur, Alfred
7:56
Marshall, Gary Becker, who
7:58
are like very strong. you know,
8:00
top 10 or top 12 figures, but I
8:02
don't think they're really in the running
8:05
to be number one. And where
8:07
do you think your list, I
8:09
saw you did an episode with
8:11
the market monotrists, I forget, which
8:13
macromusings, where does your list differ
8:15
from your peers in terms of,
8:17
you know, where do you think
8:19
they're more likely to feel differently?
8:22
It depends who you take as
8:24
my peers, but if you took kind
8:26
of median economists at top
8:28
30 schools, I don't think
8:30
they would deviate far from my
8:32
list. They probably wouldn't put malthus
8:35
on it, or maybe they wouldn't put mill
8:37
on it, and they might have Samuelson an
8:39
arrow a bit higher. But I don't think
8:41
I'm far from a consensus in
8:43
whom I'm considering. Now Scott Sumner,
8:45
speaking of market moniturists, he
8:47
wrote a post on the greatest
8:50
monetary economists, and he said they
8:52
are David Hume, Milton Friedman,
8:54
Friedman, and Irving, and Irving, and
8:56
Irving, with Scott on that. Those are
8:58
the greatest monetary
9:00
economists. What's Hume's contribution
9:03
there? He was first, for one
9:05
thing. He understood why mercantilism
9:07
was wrong. He outlined the
9:10
quantity theory of money and
9:12
what causes inflation. And he
9:14
had a very good understanding
9:16
of how inflation moves through
9:18
an economic system. All in
9:20
clear, beautiful prose. I think
9:22
it was published in 1752, but
9:24
written sometime before then. And Hume
9:27
was one of the smartest
9:29
people ever, maybe the smartest,
9:31
and in economics too. He's amazing.
9:33
And my sense is that 100 years
9:35
from now, if we're talking about the
9:37
best philosophers of all time, that most
9:40
of that list will still be the
9:42
same from today, that we look back
9:44
for, let me know if you think
9:46
that's incorrect. I'm curious, yeah, I'm curious in
9:49
economics, if we look out 100 years
9:51
from now, how many... Are we going
9:53
to have new economists that come up?
9:55
Or is it also going to be
9:57
sort of a sort of, you know,
10:00
old field like philosophy. I don't
10:02
think it will change much. I
10:04
think Malthus is the candidate who
10:06
might be revalued. So if we
10:08
do collapse due to environmental problems,
10:10
which is not my prediction, not
10:13
my view, but it's possible, right,
10:15
then I think Malthus will really
10:17
rise in people's estimations. And
10:19
if we don't collapse, he'll
10:21
probably fall somewhat more. The others
10:24
will keep their posts of honor,
10:26
so to speak. Yeah. Well on
10:28
the Malthus question. That feels to
10:30
have resurfaced in light of AI
10:33
safety, resurfaced in different ways, this
10:35
idea of how much growth can
10:37
we have. The guy wrote the
10:40
book's scale, I can't remember, Jeffrey
10:42
West, he quoted in the book,
10:45
maybe quoted someone else, this quote
10:47
that said, the only people that
10:49
think exponential growth is possible are
10:52
idiots and economists. And I remember
10:54
asking you that a few years
10:56
ago, and you said, kind of glibly
10:59
I think something like you you're half
11:01
joking I think like I'll take 700
11:03
years of great growth you know at
11:05
some point maybe you know we'll be
11:08
unable to handle it. What are your
11:10
thoughts on sort of like runaway growth
11:12
and how should we think about you
11:14
know our potential ability to
11:17
handle it or not handle it?
11:19
Well economic growth brings dematerialization
11:21
right so we can do
11:23
more using less water using
11:26
less land using less material
11:28
resources. So there's not some
11:30
inexorable logic where the material pressure
11:33
on the earth just causes it all
11:35
to explode or run out of
11:37
everything. On that I'm quite optimistic,
11:39
as are most economists.
11:41
My main worry, which I've had for a
11:43
long time, is simply that as
11:46
energy becomes cheaper, not intelligence,
11:48
but energy, it will at
11:50
some distant point be too easy
11:52
for someone to build by something
11:54
like a nuclear weapon. and just set
11:56
it off, and that may lead to the collapse
11:58
of order in the world. Say you could
12:00
set off something like a
12:02
nuclear weapon for $50,000. What would
12:04
the world look like? I don't know.
12:07
I just don't think that's stable
12:09
forever. And that's my biggest
12:12
fear for the future. Our mutual
12:14
friend Nathan LeBens, he equates
12:16
or makes analogies between AI
12:18
and nuclear. He says we
12:20
wouldn't open source nuclear codes
12:22
or sort of, you know,
12:25
bio weapons. AI is going to get
12:27
that good, why would we think
12:29
about open sourcing something that is
12:32
so potentially dangerous? Are
12:34
you sympathetic to that instinct? I
12:36
would reframe it. I would say some
12:38
of the cats are already out of
12:41
the bag, and we have to decide,
12:43
do we have to deal with our
12:45
problems through Chinese AI or American
12:48
AI? And on that, my choice is clear.
12:50
So I think he just has a
12:52
quite wrong starting point for the
12:55
question. I don't think it will necessarily
12:57
be easy. I'm not worried about
12:59
the Ilea's scenario where it just
13:01
kills us all. I just think any
13:03
technology that's super dynamic is
13:05
also disruptive, and yet
13:07
we have all these institutions geared
13:10
around stasis, including our politics.
13:12
And again, it's an open question
13:14
how well that's going to go. But
13:16
I would still rather confront
13:18
it with more intelligence than
13:20
less intelligence. That relates to
13:23
another topic that we discussed
13:25
the other week, which was
13:27
fertility and underpopulation, the population,
13:30
sort of demographic collapse.
13:32
My AI friends in San Francisco
13:34
will say, don't worry about that,
13:37
because the losses that you're forecasting
13:39
as a result of a lower
13:41
population will be offset by the
13:43
gains in productivity that AI will
13:46
bring. And in fact, If we have
13:48
too many people, we won't have enough things
13:50
for them. Although that you could make that
13:52
claim doesn't need to be part of
13:54
that argument for it to make sense. Do
13:57
you do you sympathize with that? The
13:59
implication is that's. we shouldn't worry
14:01
about population as much
14:03
because AI will take care of
14:05
it. Elon is right. I mean,
14:08
I'm optimistic about AI on the whole,
14:10
but you need people around to enjoy
14:12
what the AI will do for us.
14:14
So if South Korea has 17
14:16
people, an awesome AI, and those
14:19
17 only average half a kid,
14:21
they're done. Doesn't matter how
14:23
good the AI is. Now my hope, which
14:25
I'm, again, I'm not sure. I hold
14:27
this as a prediction, but if
14:30
AI raises productivity enough, we
14:32
can subsidize people having kids
14:35
to the tune of like three
14:37
or four hundred thousand dollars
14:39
per kid. And a subsidy that
14:41
large, I do think, would matter.
14:43
Maybe something like robots
14:45
or certain devices would make
14:47
rearing a child easier also. But
14:49
you just can't have a
14:52
shrinking population for too long.
14:54
We'll continue our interview in a moment
14:56
after a word from our sponsors. Even if
14:58
you think it's a bit overhyped, AI
15:00
is suddenly everywhere, from self-driving
15:03
cars to molecular medicine to business
15:05
efficiency. If it's not in your
15:07
industry yet, it's coming fast. But
15:09
AI needs a lot of speed
15:11
and computing power. So how do
15:13
you compete without costs spiraling out
15:15
of control? Time to upgrade to
15:17
the next generation of the cloud.
15:19
Oracle cloud infrastructure or OCI. OSI
15:21
is a blazing fast and secure
15:23
platform for your infrastructure, database, application
15:25
development, plus all your AI and
15:27
machine learning workloads. OSI cost 50%
15:29
less for compute and 80% less
15:31
for networking. So you're saving a
15:33
pile of money. Thousands of businesses
15:36
have already upgraded to OSI, including
15:38
Botafone, Thompson Reuters, and Suno AI.
15:40
Right now, Oracle is offering to
15:42
cut your current cloud bill in
15:44
half if you move to OSI
15:47
for new US customers with minimum
15:49
financial commitment. Offer ends March 31.
15:51
March 31. See if your company
15:53
qualifies for this special offer at
15:55
Oracle.com/Turpentine. That's Oracle.com/ Turpentine. Most
16:03
of you probably know how much the
16:05
turpentine team loves notion. The sponsor of
16:07
today's episode. We use it every day
16:09
for editorial planning, transcript markups, project managing,
16:11
knowledge sharing, and it's where I drop
16:13
random 3 AM ideas that turn into
16:15
episodes. Now it has the power of
16:17
AI, built right inside of notion and
16:19
woven across our entire workspace. It's been
16:21
a total game changer for our productivity
16:23
and focus. We just used it the
16:25
other day to architect the content
16:27
schedule for next month across our
16:30
network. Notion combines your notes, docs,
16:32
and projects into one space that's
16:34
simple and beautifully designed. The fully
16:36
integrated Notion AI helps you work
16:38
faster, write better, and think bigger.
16:40
Doing tasks that normally take you
16:42
hours in just seconds. Try Notion
16:45
for free when you go to
16:47
Notion.com/Upstream pod. That's all lower case
16:49
letters. Notion.com/Upstream pod. To try the
16:51
powerful, easy to use Notion AI
16:53
today. And when you use our
16:55
link, you're supporting our Show. Upstream
16:57
pod. We
17:01
talked about also how there have been
17:03
no policies that have seemed to make
17:06
a difference. The one idea I had
17:08
since then is what if there was
17:10
an equivalent of a DUI, but for
17:12
people who've had kids, i.e. in
17:14
order to progress in your career
17:16
or rise up within a company,
17:18
you have to have two kids.
17:21
Or kind of your preferential access
17:23
if you have that. Do you
17:25
think that could? Because we don't
17:27
just want to encourage anybody to
17:29
have kids who want to encourage
17:31
that. you know, most driving, you
17:33
know, most intelligent, most impressive people
17:35
as well. I would think that's hard
17:37
to enforce. There's going to be
17:40
plenty of exceptions. And a
17:42
simple cash subsidy appeals to my
17:44
background as an economist. But
17:46
whether it's a mandate or a subsidy, you
17:48
have to pay for it in some manner.
17:50
Yeah. So I think there's a race to find
17:52
a way to pay for it. And that's
17:55
one reason why just turning our back on
17:57
AI is not an option. Because it's
17:59
by far. the best chance we have of
18:01
being able to afford a lot
18:03
of other protections we're going to
18:06
need. Asteroid defense, whatever other
18:08
problems you think are there, just
18:10
having a more productive Walmart is not
18:13
going to pay for it. Yeah,
18:15
it is interesting. If we would have
18:17
said, you know, 20 years ago, 30
18:19
years ago, if we would have said,
18:21
hey, our goal is to have more
18:24
Latinos. Yeah, as a half Latino, I'll
18:26
use my example in represented in Big
18:28
Tech or something, we could have said,
18:30
hey, we should just, you know, give
18:33
Big Tech companies a subsidy. Or we
18:35
could have said, hey, let's put all
18:37
sort of cultural pressures on them in
18:39
order to do so. And it feels
18:42
like the cultural pressures, I guess, plus
18:44
some legal, you know, sort of civil
18:46
rights laws as well, sort of
18:48
emboldening those culture pressures. It feels
18:50
like those worked. Feel free to
18:52
let me know if you don't
18:54
agree with that characterization. But I
18:56
wonder if cultural pressures are underrated
18:58
as a force to encourage people
19:00
to do things. Well, cultural pressures
19:02
to have more kids. I do
19:05
see them coming back. That if someone
19:07
is somewhat well off, to me,
19:09
anecdotally, it seems much higher status
19:11
to have four kids now than it
19:13
did 10 years ago. That may not
19:16
be enough people to change matters.
19:18
But keep in mind, there's this
19:20
other evolutionary element of the
19:22
system that cultures that encourage
19:25
childbearing, Amish is the extreme
19:27
example, they will become
19:29
more important. as other people don't
19:32
have kids. If you think there's
19:34
something about East Asian culture
19:36
where people have especially low fertility,
19:38
well that will become less influential.
19:41
So that might work in our
19:43
favor, just a numbers game, through
19:45
social evolution. Again, I'm not
19:47
sure that's strong enough to overturn
19:49
it because it seems almost every
19:52
culture is headed for below 2.0.
19:54
And you actually need a bit higher
19:56
than 2.0 to maintain a population.
19:58
Except Israel. What can we
20:00
learn from, what can other countries learn
20:03
from, from Israel's situation? Some
20:05
of it is religion, but
20:07
even secular Israelis and Tel
20:09
Aviv, I believe, are above two. It
20:11
may be the sense of being threatened and
20:14
beleaguered, that if we don't have
20:16
children, like there won't be an Israel
20:18
at some point, an acute awareness of
20:20
numbers. So I think every Israeli
20:22
knows how many Jews were killed in
20:25
the Holocaust, and so if how many Jews
20:27
were back then, how many there are now.
20:29
It's too high a price to pay
20:31
for an awareness of numbers, but
20:33
that may be where it comes
20:36
from. Yeah. Are you with Robin
20:38
Hansen that we are likely
20:40
to have a dark age
20:42
as a result of global
20:44
population collapse? Or how bad
20:46
is this going to get? No,
20:48
I think AI has quite a good
20:50
chance of limiting or for stalling
20:53
or preventing that dark
20:55
age that the big gains will
20:57
arrive in time. But there's still
20:59
the stark age if there are
21:01
not many people around to enjoy
21:03
it. So in that sense, Robert
21:05
is broadly correct, but narrowly, I
21:08
would say incorrect. He thinks
21:10
when you have shrinking population,
21:12
just tech progress stops. That's
21:14
an interesting hypothesis, but
21:17
I don't think it's obviously
21:19
true, because you have more
21:21
brains through AI working for you. Why
21:23
don't you think there are more people
21:25
sort of... interested in
21:27
this or obsessed with this? We
21:29
had some hypotheses in our conversation,
21:32
but given sort of the scope
21:34
of this problem and the sort
21:36
of ambition of what it would
21:38
take to fix it, wouldn't it
21:40
make sense that we would have more
21:43
brilliant people focused on it? Why
21:45
don't you think that is? Well,
21:47
an awful lot of potential grandparents are
21:49
focused on it, right? And that's
21:51
a big chunk of America. and
21:54
a lot of smart people are
21:56
focused on it but they're afraid
21:58
to speak up in public. society
22:00
it's a little difficult to talk
22:02
about because it sounds at least
22:05
like you're suggesting women
22:07
should deprioritize career and
22:09
prioritize having three or four
22:11
children and that's just not a
22:14
popular thing to say or even hint
22:16
at. So the discourse is stunted
22:18
but I think Elon has really
22:20
shifted the Overton window as have
22:22
a number of other people and
22:24
there's so many potential or
22:27
current grandparents out there would
22:29
be parents. It's not going to
22:31
go away. So by focusing, I don't
22:34
just mean have have more than two
22:36
kids. I mean, identify the
22:38
bottlenecks, or whether they be reproductive
22:40
technology, whether they be child care,
22:42
whether they be housing, whatever it's
22:45
none of those. I know all
22:47
the imbies will make things cheaper.
22:49
Yeah. You know, I'm all for yimbi
22:52
and making things cheaper. But South
22:54
Korea is a super yimbi countries.
22:56
And there is 0.7. Japan is
22:58
a MB country. They're at 1.3.
23:00
I think the main bottleneck
23:03
is that to most people kids
23:05
aren't that much fun. And that's
23:07
a really tough bottleneck to
23:10
overcome. And women often are
23:12
the ones who have to put in
23:14
more work and who have a greater
23:16
say in the matter. So yeah, the
23:18
opportunity cost of women is too
23:21
high. So what you're saying is
23:23
the economics. If you paid people
23:25
$300, $300,000. lower the opportunity, you
23:27
know, that would increase the gains.
23:29
I also wonder if it would increase
23:31
the status somehow, or if there's a
23:33
way to increase the status where it
23:35
almost feels like you're not making a
23:37
tradeoff between career and mother, because mother
23:40
is a career that is both remunerative
23:42
and respected. That's right. We would
23:44
do that, but you know, there are
23:46
negative historical associations. So the Nazis
23:48
made a big point of paying people to
23:50
have kids and raising its status, and
23:52
that was part of something really
23:55
extremely unpleasant. And you might say,
23:57
well, there's some intrinsic reason
23:59
why. having the government pay for
24:02
kids and raise the status
24:04
of kids will lead to other bad
24:06
views. I don't think that's the
24:08
case, but again I don't dismiss
24:10
that argument. It is possibly
24:12
true and that maybe in
24:14
a lot of ways you don't
24:17
want to reverse the feminization of
24:19
society because it is more peaceful
24:21
and less fascistic in some
24:23
ways. So I don't dismiss that
24:26
argument. So like I've never made
24:28
it my personal mission. to go around
24:30
saying, oh, we need to like pay
24:32
400K for a kid. Because it's
24:35
very hard to foresee all
24:37
the levers and pulleys and
24:39
what you're really doing. But
24:41
I'll just say, you know, sort of
24:43
predictably, the societies that
24:46
do something like that will
24:48
in numbers outper the
24:50
societies that don't. Yeah. A
24:52
female friend who was talking to
24:54
about this, one thing that she
24:56
said, she loves the idea of
24:58
having kids. but she's more motivated
25:00
in her career. One of the
25:02
reasons is because she feels like
25:04
she can really differentiate in her
25:07
career. She can become the best
25:09
in the world at something and
25:11
really prove herself and gain sort
25:13
of respect from that. Where she
25:15
feels as a mother, she would
25:17
be kind of like everyone else.
25:19
It doesn't feel that there's distinct
25:21
ways to differentiate or become exceptional
25:23
in public societal recognized ways. So
25:25
then I said, well, if you have
25:27
20 kids, that would be exception. But
25:30
I didn't actually say that. But look,
25:32
as a mother or father, you are
25:34
truly exceptional to your own children, right?
25:36
There's no doubt about that. The question
25:39
is, how much is that worth to you? And
25:41
that gets into these deep cultural
25:44
issues about choice of audience.
25:46
And it does seem there's
25:48
something about the contemporary world that
25:50
gives you such a portfolio range of
25:52
choice of audience. And kids are still
25:54
there, but they've always been
25:57
there, and they're just another
25:59
possible audience. And they're
26:01
not necessarily winning out. Yeah.
26:03
If you were ahead of X-risk,
26:05
and you were given a
26:07
massive budget, where would sort
26:09
of population collapse be on
26:11
your priority list? Or what would
26:14
be, how would you kind of
26:16
pie chart your priority list if
26:18
you were ahead of existential risk
26:20
planning? Well, I don't think
26:22
it's a problem you solve well
26:25
by spending lots of money. So
26:27
I would in general. as I would for
26:29
a lot of problems and X-risk
26:31
problems, cultivate a lot of talent.
26:34
You don't know what your X-risk is
26:36
going to be, typically. Everyone thinks
26:38
they know. And I have my own
26:40
hypothesis about very cheap, very
26:42
powerful weapons. But if you don't know
26:45
in advance, you want to make these
26:47
very general investments. And in fact,
26:49
the work I do is geared
26:51
toward that end. So in a sense, I
26:53
do had a kind of X-risk fund. It
26:55
doesn't think of it. Yeah,
26:58
emergent ventures, it doesn't focus
27:00
on X-risk, but under the
27:03
surface it believes that a lot
27:05
of general talent being
27:07
mobilized is what's going to
27:09
do the trick. Yeah. So I
27:12
understand sort of fast grants
27:14
and how that ties to
27:16
it. Say more on emergent
27:18
ventures in terms of. your algorithm
27:20
for picking people and how you
27:22
think it ties to it? Because
27:24
my read from the outside in
27:26
is that you pick a number
27:28
of people are almost kind
27:31
of like mini Tyler's in
27:33
that they're very strong generalists,
27:35
you know, exceptionally curious, very
27:38
creative, very knowledgeable. Is
27:40
that too crude or how do
27:42
you think about your your algorithm
27:44
for what you're trying to do
27:46
with merchant ventures? Well, many
27:48
more of the winners are
27:50
specialists than generalists, but I think
27:52
given who and what you are, the
27:55
generalists are the ones you hear about
27:57
and read about. Just today, I made
27:59
an... to a 15-year-old young woman in
28:01
Canada who wants to study how
28:04
to make batteries last longer for
28:06
a variety of reasons, which are
28:08
mostly obvious. And she's extraordinarily impressive
28:10
whether she ends up working on
28:12
batteries. I wouldn't say I don't
28:14
care, but it's not an investment
28:16
in batteries. It's an investment in
28:19
her. She's an extraordinary writer, thinker,
28:21
full of energy. engages with a
28:23
lot of other very smart young
28:25
people. It just seemed obvious to
28:27
me we should make an award.
28:29
I wouldn't call that an algorithm.
28:31
This is some number of people
28:34
where it seems obvious to me
28:36
we should make them an award.
28:38
We now have over 500 winners,
28:40
noting that Shrudi and Rashid Griffith,
28:42
you know, do India and Africa
28:44
respectively. I do the rest of
28:46
the world. Yeah. So I see
28:49
how the battery technology present, but...
28:51
When you pick someone who's working
28:53
on some sort of liberal arts
28:55
thing, like how do you justify
28:57
that? Or like, what is the
28:59
scope of emergent ventures? If you
29:01
saw someone who was doing something
29:03
interesting in basketball analytics, would that
29:06
be, you personally found it interesting,
29:08
but would that be outside of
29:10
the scope of who wins? How
29:12
do you think about that? I
29:14
don't think it's outside of the
29:16
scope. I would be interested in
29:18
whether they work to draw broader
29:21
lessons from basketball analytics. I could
29:23
imagine supporting that. There's one woman
29:25
we supported, her original emphasis, which
29:27
she still does by the way,
29:29
and it's very good, but to
29:31
compose contemporary music and Azerbaijani traditions,
29:33
which is like beyond niche. Her
29:36
music is excellent. She's actually ended
29:38
up becoming an important figure in
29:40
the medicine science movement, super smart,
29:42
great work ethic. and will do
29:44
great things in both areas. Now
29:46
I had not at all foreseen,
29:48
she would end up working on
29:51
medicines in a significant way. but
29:53
I mean there you go so
29:55
I mean you know how venture
29:57
capital works you take a lot
29:59
of chances yet you've got to
30:01
look for things other people are
30:03
not looking for or what's the
30:05
point you can always find people
30:08
on the rise and fund them
30:10
and both that you are responsible
30:12
for it but that's a waste
30:14
of time like I'm not paid
30:16
to do this so I really
30:18
want to pick the people who
30:20
will have a marginal impact right
30:23
yeah the venture capital works is
30:25
Exactly described, but the result is
30:27
clear, the scoreboard is clear. They
30:29
either have a return or they
30:31
don't, whereas for what you're doing,
30:33
the return is a bit more
30:35
subjective, right? We need more non-legibility
30:38
in modern society is one of
30:40
my strongly held beliefs. And emergent
30:42
ventures is designed not to be
30:44
fully legible. Like does the person
30:46
believe passionately in what they're doing?
30:48
Like it probably can't be a
30:50
hair salon in Dayton, Ohio, but
30:53
an awful lot of different things
30:55
are eligible. Why do we need
30:57
more knowledgeability? What does that bring?
30:59
Well, so many different institutions are
31:01
being bureaucratized, right? So there's committees,
31:03
there's boards, there's letters of recommendation,
31:05
there's credentials, all mattering more and
31:07
more, too much. And someone who
31:10
wants either like a small grant
31:12
or something that's hard to explain
31:14
or justified to a committee. They're
31:16
the cases where maybe, I think,
31:18
my marginal return is highest. So
31:20
I'm attracted to those quite often.
31:22
The 15-year-old young woman, there's no
31:25
way she can just apply to
31:27
a foundation and sort of get
31:29
through the process for what is,
31:31
you know, a small grant and
31:33
part of it is the networking
31:35
opportunity and inspiration and chance to
31:37
consult with different people. And that's
31:40
what we're trying to provide. Yeah,
31:42
speaking of bureaucratization and you know
31:44
and ministrification, let's talk about Harvard
31:46
is what's happening at Harvard, a
31:48
sort of watershed moment for the
31:50
institution? Or is it kind of
31:52
new person, same, same, it feels
31:55
like Bill Ackman is conducting an
31:57
activist campaign to take over Harvard
31:59
in a similar way that Elon
32:01
maybe did to take over Twitter?
32:03
It turned out easier to take
32:05
over Twitter than then start a
32:07
new one, and it's really hard
32:09
to start a new Harvard. No
32:12
one has been able to do
32:14
anything like it in the last
32:16
few. you know, a few decades
32:18
maybe 100 years. Yeah, more than
32:20
a few decades. I think Harvard
32:22
will shift to the center and
32:24
change marginally and the next president
32:27
will be quite a safe choice,
32:29
probably too safe a choice, and
32:31
the worst outrages will slowly be
32:33
purged or downplayed, but Harvard will
32:35
not fundamentally change. And it's never
32:37
been fair or meritocratic. Now today
32:39
it's unfair and non-meritocratic. in quite
32:42
new and different ways, which many
32:44
people hate. And I'm not myself
32:46
fond of them. But don't be
32:48
too shocked. There's no idealized Harvard
32:50
in the background that we've strayed
32:52
from. So it was not designed
32:54
to be fair and meritocratic, and
32:57
it will continue to be Harvard.
32:59
So the Harvard of 2023 is
33:01
not that significantly different from the
33:03
Harvard of different decades in terms
33:05
of this idea of like pursuit
33:07
of truth people there are doing
33:09
real work that matters Well many
33:11
departments are better in that regard
33:14
than they used to be Maybe
33:16
not all of the humanities, but
33:18
a lot of the sciences are
33:20
much stronger and there's more funding
33:22
and stronger checks and better networking
33:24
So significant parts of Harvard have
33:26
never been better than they are
33:29
today It's important to realize that.
33:31
But within the administration, there's so
33:33
much rot and bad ideology and
33:35
lack of accountability and clueless people
33:37
living in bubbles, maybe even in
33:39
some cases having personality disorders. That
33:41
it's not very functional at some
33:44
of the macro levels. Yeah. Our
33:46
friend Mark Andreessen is famously much
33:48
more negative on the state of
33:50
current higher education. You know, we're
33:52
using Harvard as kind of a
33:54
placeholder for the elite education. But
33:56
even in the even in the
33:59
hard sciences, he thinks a lot
34:01
of it has been either politically
34:03
corrupted or or more just like
34:05
non reproducible or just. you know,
34:07
majority of what people are producing
34:09
doesn't, isn't real, doesn't matter. That's
34:11
always been the case. I would
34:13
stress that a lot of the
34:16
tech excitements that markets excited about
34:18
are in fact downstream of universities.
34:20
Even still, or like, I know,
34:22
maybe more. Got it. So when
34:24
you, when we look at the
34:26
future of higher education, let's say
34:28
we're having this conversation 40 years
34:31
from now, do we think that
34:33
higher education is going to look
34:35
pretty similar to how it does
34:37
today in the sense that there
34:39
will be Harvard there will be
34:41
Stanford there will be the major
34:43
players that exist will be the
34:46
ones that already exist and the
34:48
experience is not going to be
34:50
that different or how do we
34:52
think the experience will be quite
34:54
different I would still bet on
34:56
those to be the top schools
34:58
I think AI will change higher
35:01
education a great deal whether higher
35:03
ed likes it or not just
35:05
as the internet has changed it
35:07
a great deal So I'm not
35:09
predicting stasis. And I do think
35:11
at least half of these top
35:13
schools have truly terribly macro-level dysfunctional
35:15
administration in a way where something
35:18
does have to give. So the
35:20
critics are not wrong when they
35:22
talk about rot. But if they
35:24
think just like the quality of
35:26
what's coming out of MIT Stanford
35:28
Harvard is junk, it's just simply
35:30
not true. Again, you might think
35:33
it's true in some number of
35:35
fields. the less scientific ones, but
35:37
it's just flat out not true.
35:39
It's not true in economics. Smart
35:41
people still want to go to
35:43
those places. It's not true for
35:45
biomedicine. All the vaccines and other
35:48
innovations we're getting, largely downstream of
35:50
universities. We'll continue our interview in
35:52
a moment after a word from
35:54
our sponsors. Every time I hop
35:56
on the 101 freeway in SF,
35:58
I look up at the various
36:00
tech billboards and think, damn, turpentine
36:02
should definitely be up there. And
36:05
I bet I'm not alone. In
36:07
a world of endless scrolling and
36:09
ad blockers, sometimes the most powerful
36:11
way to reach people is in
36:13
the real world. That's where Adquic
36:15
comes in. The easiest way to
36:17
book out-of-home ads, like billboards, vehicle
36:20
wraps, and airport displays, the same
36:22
way you would order an Uber.
36:24
Started by Instacard Alums, Adquic was
36:26
born from their own personal struggles,
36:28
to book and measure ad success.
36:30
Adquix audience targeting, campaign management, and
36:32
analytics, bringing the precision and efficiency
36:35
of digital to the real world.
36:37
Ready to get your brand the
36:39
attention it deserves? Visit adquick.com today
36:41
to start reaching your customers in
36:43
the real world. Again, that's AD,
36:45
QUICK.com. Hey everyone, Eric here. In
36:47
this environment, founders need to become
36:50
profitable faster and do more with
36:52
smaller teams, especially when it comes
36:54
to engineering. That's why Sean Lanahan
36:56
started squad. A specialized global talent
36:58
firm for top engineers that will
37:00
seamlessly integrate with your org. Squad
37:02
offers rigorously vetted top 1% talent
37:04
that will actually work hard for
37:07
you every day. Their engineers work
37:09
in your time zone, follow your
37:11
processes, and use your tools. Squad
37:13
has front-end engineers, excelling in typescript,
37:15
react, and NextGS, ready to onboard
37:17
to your team today. For back-end,
37:19
Squad engineers are experts at NojS,
37:22
Python, Python, Java, and a range
37:24
of other languages and frameworks. While
37:26
it may cost more than the
37:28
freelancer on Upwork billing you for
37:30
40 hours, but working only two,
37:32
Squad offers premium quality at a
37:34
fraction of the typical cost, without
37:37
the headache of assessing for skills
37:39
and culture fit. Squad takes care
37:41
of sourcing, legal compliance. and local
37:43
HR for global talent. Increase your
37:45
velocity without amping up burn. Visit
37:47
choose squad.com and mention turpentine to
37:49
skip the wait list. Would you
37:52
agree that Elon taking over Twitter
37:54
has had a massive culture shift?
37:56
Absolutely, and positive. I think he
37:58
just had the stones to say,
38:00
I'm willing to lose, what, 30
38:02
billion? I don't know the number,
38:04
30 billion. Whatever hell end up
38:06
losing. What he paid 44 billion,
38:09
it'll be worth something. willing to
38:11
lose 30 billion to steer this
38:13
beast in one direction because he
38:15
understood how much is downstream of
38:17
Twitter and it was brilliant and
38:19
for all the critics he's basically
38:21
pulling it off of course he's
38:24
going to lose the money though
38:26
yeah is is that possible could
38:28
something similar happened to Harvard well
38:30
you can't buy Harvard so someone
38:32
could in some sneaky way try
38:34
to take over the board of
38:36
directors I don't know what their
38:39
bylaws are like Typically that's very
38:41
very hard and I would be
38:43
very surprised if that were a
38:45
realistic option Have you been following
38:47
Bill Ackman's Twitter? He's of course
38:49
He's been extraordinarily effective, but I
38:51
think he will find it much
38:54
harder to get broader reform than
38:56
to get a few scalps Yeah,
38:58
I'm very glad he did what
39:00
he did my kudos to him
39:02
But at some point you just
39:04
have systems filled with people of
39:06
particular points of view, and they
39:08
have tenure, I just don't see
39:11
how that budgets. There's this narrative
39:13
some people have where they basically
39:15
say that the US is so
39:17
much of downstream, you mentioned of
39:19
Twitter, but also, but between academia,
39:21
you know, between Harvard, the New
39:23
York Times, and government, there's what
39:26
some people call the cathedral, this
39:28
gated institutional complex, the sort of
39:30
like web of relationships and kind
39:32
of self reinforcing. kind of power
39:34
generators. I'm using pretty like vague
39:36
terms, but you You kind of
39:38
get what I'm getting at. Sure,
39:41
of course. You agree with that
39:43
sort of narrative? Like, what would
39:45
be the, if someone were able
39:47
to take over Harvard, would it
39:49
have a similar culture shift to
39:51
what Elon has done for Twitter?
39:53
I don't think so, because you're
39:56
not changing the research that's done,
39:58
what's taught in the classroom. So
40:00
I think the people who talk
40:02
about the cathedral, in general, they
40:04
underrate the quality of the establishment.
40:06
They underrate the degree of dissent
40:08
within the establishment, and they underrate
40:10
the degree to which the establishment
40:13
can reform itself. And to think
40:15
that the outsiders will nudge the
40:17
establishment in a better direction, and
40:19
some of them will become part
40:21
of it, and will muddle through,
40:23
and a lot of things will
40:25
continue, is to me by far
40:28
the best bet. Yeah. It's interesting,
40:30
right now as we're communicating, Richard
40:32
Anania and Bronze Age pervert. are
40:34
getting in a feud on the
40:36
internet about sort of this. And
40:38
if people don't know who those
40:40
are, God bless you, these are,
40:43
you know, sort of Twitter, public
40:45
intellectuals on the right. Richmond is
40:47
a part of a merchant and,
40:49
you know, an emergent winner and
40:51
his fascinating thinker and broad age
40:53
pervert is a, you know, Yale
40:55
PhD philosopher turned a non, you
40:58
know, book author and Twitter personality,
41:00
and they are getting in an
41:02
argument over how to explain to
41:04
the public why there are differences
41:06
in outcomes among people's. And the
41:08
reason we're having that conversation is
41:10
because DEA is kind of being
41:12
questioned in a way that it
41:15
hasn't been questioned for a long
41:17
time because it was happening at
41:19
Harvard. the logic that some people
41:21
say is Amy Wax is proposing
41:23
this view and others is it
41:25
takes a theory to beat a
41:27
theory. And the reason why D.
41:30
I was able to or affirmative
41:32
action was able to be implemented
41:34
so effectively is because people look
41:36
at disparities and they say, hey,
41:38
this must be explained by racism
41:40
or prejudice or systemic discrimination or
41:42
whatever. And if you don't have
41:45
a response to that then you
41:47
have to give into that framing.
41:49
Thomas Sol, of course, wrote books
41:51
trying to explain how culture matters
41:53
and history matters and and explain
41:55
how all groups are different, whether
41:57
it's the French or Russians or
42:00
Italians, like every group is going
42:02
to have different outcomes. And so
42:04
Charles Murray, of course, has been
42:06
on this kick for a recently
42:08
and also a long time, if
42:10
there's a question as to how
42:12
honest society should be about things
42:14
that are maybe ugly or not
42:17
flattering or could lead people to
42:19
take the wrong interpretations of them
42:21
when we think about why outcomes
42:23
are different outcomes are different. How
42:25
do you react to this debate?
42:27
Either on the meta-conversation of how
42:29
honest should we be as a
42:32
society or how should we think
42:34
about inequalities as a society between
42:36
people? Should we think that, hey,
42:38
that's just a natural outcome? Should
42:40
we try to mitigate it within
42:42
certain constraints? Where do you net
42:44
out? There's a lot of different
42:47
questions packed into there. I think
42:49
a lot of that current debate,
42:51
I haven't seen all of it,
42:53
but it seems to me, Karen
42:55
are productive. It's people talking to
42:57
themselves, going in circles. And what
42:59
people really should be working on
43:02
is innovations and whatever space that
43:04
will help groups that have been
43:06
at a disadvantage do better. So
43:08
I think we should all admit
43:10
racism has had a huge impact
43:12
in the American past. It is
43:14
by no means gone. There are
43:16
some markers that it's coming back,
43:19
at least in significant corners, even
43:21
maybe if not at the median.
43:23
And that's a bad thing. And
43:25
yeah, we need to be. more
43:27
you know more positive looking more
43:29
forward looking and more positive some
43:31
that probably puts me closest to
43:34
Seoul. Some people say that Bronze
43:36
Age pervert and we don't have
43:38
to talk about him specifically but
43:40
just has popularized this idea of
43:42
kind of like a you know
43:44
I think Christianity as a blog
43:46
post around sort of a mythos
43:49
for the non-Christian right as as
43:51
sort of a nichean liberalism or
43:53
something, but like emphasis on the
43:55
niche and or sort of revitalizing,
43:57
reviving niche for a kind of
43:59
online, memetic, you know, Twitter, Twitter
44:01
era. It seems like, you know,
44:04
we talked about how earlier the
44:06
culture is, it's pretty, pretty feminized,
44:08
pretty, you know, I mean, it's
44:10
a pretty Christian. Do you also
44:12
see sort of this rise of,
44:14
of kind of a new niche
44:16
inism? And what do you, what
44:18
do you make, make sense of
44:21
it? So there's a corner of
44:23
the right wing that has made
44:25
it their thing as a kind
44:27
of active rebellion I'm not Bronze
44:29
Age pervert understands Nietzsche. I'm not
44:31
sure how many of the others
44:33
do It's not going anywhere So
44:36
I mean to look at it
44:38
very concretely, you know to take
44:40
the world of Latin America, which
44:42
you and I are Both reasonably
44:44
familiar with what really changes people's
44:46
people's people's people's people's people's people's
44:48
people's people's people's people's people's people's
44:51
people's people's people's people's people's people's
44:53
people's people's people's people's people's people's
44:55
people's people's people's people's people's people's
44:57
So now Panama and Dominican Republic
44:59
and Chile are doing pretty well.
45:01
And there's different things you can
45:03
do to encourage them to do
45:05
better or have more cases like
45:08
that. I hope Colombia joins that
45:10
group. They have the potential to,
45:12
but they haven't yet. I hope
45:14
Argentina does. And if those things
45:16
can go better, you're going to
45:18
get much better outcomes. And not
45:20
like, don't worry too much about
45:23
Nietzsche or whatever. That's where the
45:25
action is. So you're less excited
45:27
about any sort of post- liberalism
45:29
or the people? who talk about
45:31
how liberalism has atomized us or
45:33
continued to lead to sort of
45:35
a lot of these cultural problems
45:38
because the economic growth suffered? Migrants
45:40
want to live in the liberal
45:42
countries, right? The liberal countries are
45:44
at risk of losing some of
45:46
that liberalism. We should all fight
45:48
to defend it and make it
45:50
better to be clear. Liberalism, by
45:53
its nature, never claimed to be
45:55
perfect or close to perfect. It
45:57
ought to focus on self-improvement in
45:59
a way that a lot of
46:01
these intellectual discourses do not. Do
46:03
you think that Patrick Dine kind
46:05
of, I can't remember, the book
46:07
is called Why did liberalism fail
46:10
or are you familiar with that?
46:12
Yeah. Do you think he's kind
46:14
of misguided? Yes, sure. He's misguided.
46:16
Where does he want to move?
46:18
I've never met him that I
46:20
can recall, but I would just
46:22
ask him. Patrick, where do you
46:25
want to go? And we do
46:27
have 50 states. I know there's
46:29
common elements. But there's a lot
46:31
of countries in the world, and
46:33
I know where I want to
46:35
be. And if I had to
46:37
leave my own country, I'd be
46:40
looking at places like Canada, Australia,
46:42
Denmark, UK, Ireland, right? Nothing that
46:44
would surprise you. Yeah. And speaking
46:46
of, or kind of adjacent, I
46:48
remember we had a conversation a
46:50
long time ago, offline, and you
46:52
mentioned how you were a, you
46:55
thought that the rolls sort of.
46:57
veil of ignorance was highly overrated
46:59
or not a great view. Can
47:01
you can you explain that? Well,
47:03
Rawls asks, you know, if you
47:05
were stripped of a lot of
47:07
features of your identity and you
47:09
would be randomly born into a
47:12
particular place, what kind of institutional
47:14
rules would you want and what
47:16
kinds of decision procedures would you
47:18
favor for evaluating outcomes? And I
47:20
think what he came up with
47:22
in a principle of liberty, parade
47:24
of improvements, difference principle. Some I
47:27
like, some I don't like, but
47:29
there are all things he stuffed
47:31
into the choice. to begin with.
47:33
He's not really proving or generating
47:35
anything. It's you put a bunch
47:37
of things in and you get
47:39
a bunch of things out and
47:42
it's just his, excuse me, his
47:44
personal recipe. So there's no derivation.
47:46
I don't think there's a case
47:48
based on reflective equilibrium. There's not
47:50
enough history. I would start with
47:52
what are the countries people actually
47:54
want to live in and concretely
47:57
how can we improve them? And
47:59
roles for me is far too
48:01
much of a foundationalist and not
48:03
enough of a historically minded thinker,
48:05
at least in theory of justice.
48:07
Now there's other roles books and
48:09
articles that are fairly different and
48:11
that's a more complicated story. But
48:14
that would be my response to
48:16
the roles that people mean when
48:18
they usually talk about roles. I'm
48:20
curious to learn more about how
48:22
you think about sort of career
48:24
or how you spend your time
48:26
in a whatever reason. conversations over
48:29
the years, I asked if you
48:31
would want to start a venture
48:33
fund, giving you so much great
48:35
access and relationships in the industry,
48:37
and especially because of what, you
48:39
know, what emergent is doing, but
48:41
also other relationships. And you mentioned,
48:44
you kind of said, hey, why
48:46
would I do that? I kind
48:48
of have the life and the
48:50
relationships and the respect and the
48:52
ability to do things that I,
48:54
that I have what I want
48:56
right now. I don't need to
48:59
do that. And it feels like
49:01
you've been able to build up
49:03
this influence. in a, or those
49:05
assets in a non, or in
49:07
a kind of a circuitous way,
49:09
in ways that other people haven't,
49:11
or couldn't emulate. How do you
49:13
explain that? Or how would you
49:16
edit my characterization? And what you
49:18
say sounds true to me, and
49:20
I think some of it you
49:22
repeated from what I said to
49:24
you. So it's no surprise I
49:26
like it. How did I do
49:28
it? I guess that means I'm
49:31
really skilled in some ways. That's
49:33
the immodest answer. But I think
49:35
people who act modest are actually
49:37
not modest, and one of my
49:39
New Year's resolutions is to not
49:41
try to be modest unless I
49:43
really mean it. So I think
49:46
that's something where I've had a
49:48
lot of luck and other people
49:50
have treated me very well, but
49:52
I've been really good at it.
49:54
Is the algorithm of how you
49:56
spend your time, what are you
49:58
really good at and what do
50:01
you enjoy and where you think
50:03
you can make the biggest impact?
50:05
I guess I'm in, I asked
50:07
today, I'm undergoing a career transition
50:09
and thinking about how to best
50:11
utilize my talents and interests, what
50:13
questions do you think I should
50:15
be asking? Well, what's worked for
50:18
me, I'm not saying it works
50:20
for everyone or for you, but
50:22
congratulations, by the way, is to
50:24
always pursue compounding returns in learning.
50:26
And I think that will work
50:28
for a lot of people, but
50:30
I'm in academia, that's a special
50:33
setting, there's a thing called tenure
50:35
you can get, and you can
50:37
be granted long-time horizons if you're
50:39
willing to take short-run reputational lumps.
50:41
A lot of other kinds of
50:43
projects have budget constraints that don't
50:45
allow that. And I'm not sure
50:48
where you're going to fall on
50:50
that spectrum. But that would be
50:52
my starting point. Compounding returns through
50:54
learning. You wrote a book going
50:56
full circle here. Your book is
50:58
on the legacy of economists. You,
51:00
of course, are also an economist
51:03
and a broader public intellectual and
51:05
have your own legacy. Do you
51:07
think that the legacy is the...
51:09
both the depth and the breadth
51:11
that you've been able to go
51:13
into not just economics but also
51:15
you know multiple fields and in
51:17
public and kind of in the
51:20
community built with marginal revolution or
51:22
do you do you say no
51:24
I actually wanted to be these
51:26
core ideas around the great stagnation
51:28
or complacent class or or or
51:30
what else? I don't think I
51:32
have much of a legacy I'm
51:35
not even sure what it means.
51:37
I think the main thing I've
51:39
done is show people a certain
51:41
kind of intellectual life as possible.
51:43
and that it can be fun,
51:45
and that it can have some
51:47
influence. And I think people will
51:50
still think about that for 10
51:52
or 20 years after I die,
51:54
and then they won't think or
51:56
talk about it anymore, and my
51:58
legacy will be gone. I'm
52:01
still very happy I can reach
52:03
and influence a lot of people
52:05
now and for some time and
52:07
that's very meaningful to me. But
52:09
I see so many people who
52:11
don't have legacies and I think
52:14
I'll be another one of them.
52:16
Yeah. As you think about what
52:18
to spend time on, does part
52:20
of you think, hey, I should
52:22
go deeper in this and create
52:24
some sort of fundamental, you know,
52:26
insight that or contribution or I
52:29
should go broader or I should
52:31
just... pursue what interests me or
52:33
how bottoms up or stops down
52:35
is your sort of intellectual, you
52:37
know, research or curiosity or how
52:39
you spend time? I don't have
52:42
fundamental insights to offer. It's pretty
52:44
simple. I wish I did. If
52:46
I did, I would do it,
52:48
but I don't. So just like
52:50
I can be immodest on saying
52:52
I set this life up very
52:54
well. Maybe that makes me more
52:57
credible when I say I'm not
52:59
foregoing or even weighing some option.
53:01
to say these like new extra
53:03
profound things. It's like I'm empty
53:05
in a way. Like I take
53:07
stuff in and something comes out
53:10
of it. But you're not going
53:12
to get any more other than
53:14
like the flow will continue for
53:16
as long as I can. And
53:18
so when you say the precise
53:20
skill is this kind of unique
53:23
style of intellectual sort of exploration
53:25
and sharing. live life as an
53:27
infor, as I call it, but
53:29
also do well in doing things
53:31
practically and show people the two
53:33
can be integrated. I think that's
53:35
a big part of it. So
53:38
take on a lot of projects,
53:40
they will succeed to varying degrees,
53:42
but I've been happy with how
53:44
most of my projects have gone,
53:46
and they all have a practical
53:48
side. And that's part of the
53:51
vision of how to live life
53:53
in a particular way. If you
53:55
could clone yourself and we'll get
53:57
your towards closing here, what's another
53:59
project? intellectual or practical that you
54:01
would want to pick up or
54:04
put differently what's a for aspiring
54:06
Tyler aspiring emergent venture you know
54:08
candidate who's listening to this episode
54:10
what's a project that you want
54:12
them to pursue that isn't being
54:14
worked on to the degree that
54:16
it should be that would be
54:19
important or valuable? Well I don't
54:21
want them to listen to me
54:23
or for that matter to you
54:25
or any other particular person so
54:27
just that they can take some
54:29
inspiration from the fact that one
54:32
can be a thing or have
54:34
a role that's not like any
54:36
other previous role and that that's
54:38
really possible is I hope something
54:40
I communicate to a small number
54:42
of people and I don't want
54:44
them to like be my role
54:47
or pursue my priorities. So that's
54:49
another way to think about like
54:51
legacy for a short window of
54:53
time. you can show people that
54:55
non-mold models of life are possible
54:57
and rewarding. Yeah, that makes sense.
55:00
It's my belief that that University
55:02
of Austin is not going to
55:04
compete with Harvard or Stanford, although
55:06
I really respect the effort and
55:08
agreeing them the best, but that
55:10
if the TEO fellowship scaled up
55:13
and tried to compete with Harvard
55:15
and Stanford, I think they could
55:17
because the credential is so significant
55:19
based on the talent level who's
55:21
been inside of it. And I
55:23
believe that. Will you build with
55:25
the emergent ventures too? It has
55:28
an alumni so impressive on a
55:30
quantity or on a capital percentage
55:32
basis that it could also compete
55:34
if you want it to scale
55:36
up, but I suspect you don't
55:38
want to. It's not want to.
55:41
I don't think, you know, we
55:43
can. So we have myself, Shrudi
55:45
and Rashid as evaluators. They're both
55:47
great. If we had a mix
55:49
of more money, more evaluators, we
55:51
could scale a bit. But it's
55:53
the non-scaling, it's like saying, oh,
55:56
I... I love the music of
55:58
the Beatles. Those guys really need
56:00
to scale. Well, yes and no,
56:02
right? There's a lot of covers
56:04
of Beatles songs. It's a kind
56:06
of scaling. But at the end
56:09
of the day, you know, you
56:11
do your thing. I would add,
56:13
I am optimistic about University of
56:15
Austin. I'm on their advisory board
56:17
to be clear, but I don't
56:19
have a personal stake in it,
56:22
but I think they will get
56:24
very good applications. I think so
56:26
too. I think if what I
56:28
was saying with that is I
56:30
think someone who's picking between Harvard
56:32
or Stanford will likely have and
56:34
UATX doesn't yet have better career
56:37
opportunities if they go to UATX,
56:39
but if they're a TFLO they
56:41
probably do, at least in technologies,
56:43
or startup specifically. Yeah, sure, that's
56:45
true. But I think they'll get
56:47
people who are equally talented as
56:50
Harvard and Stanford students, who A,
56:52
A, want something different, B, were
56:54
unfairly rejected. or just rejected because
56:56
not enough students are taken. And
56:58
they just want a change of
57:00
pace and they're like anti-establishment in
57:02
some way and they'll go on
57:05
and do great things. And they
57:07
don't need that many people to
57:09
fill the ranks. It's like it's
57:11
one school. Like can you have
57:13
80 University of Austin's is a
57:15
much tougher question. I'm quite sure
57:18
you can have one, two, three.
57:20
When you get to 10, I'm
57:22
not so sure anymore. Yeah. Well,
57:24
we're over time, we'll wrap on
57:26
that. The book is, who is
57:28
the greatest economist of all time
57:31
and why does it matter? And
57:33
people can read it on Go
57:35
at Greatest Economist of All Time.a.
57:37
Tyler, always a pleasure. Thanks so
57:39
much for coming on the podcast.
57:41
Thank you very much, Eric. And
57:43
the simpler website is just econ
57:46
goat.a.i. Or Google Tyler Cowan Goat.
57:48
Thanks Eric. Great to chat. See
57:50
you next time. Upstream
57:55
with Eric Tormberg is a show from
57:57
turpentine, the podcast network behind moments. of
57:59
Zen and Cognitive Revolution. If
58:02
you like the episode, the please leave
58:04
a review in review in the Apple Store.
Podchaser is the ultimate destination for podcast data, search, and discovery. Learn More