EP 295 John Robb on How a Networked Organization Blitzed D.C.

EP 295 John Robb on How a Networked Organization Blitzed D.C.

Released Monday, 31st March 2025
Good episode? Give it some love!
EP 295 John Robb on How a Networked Organization Blitzed D.C.

EP 295 John Robb on How a Networked Organization Blitzed D.C.

EP 295 John Robb on How a Networked Organization Blitzed D.C.

EP 295 John Robb on How a Networked Organization Blitzed D.C.

Monday, 31st March 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

Howdy, this is Jim Rut and

0:02

this is the Jim Rut Show.

0:04

Listeners have asked us to

0:06

provide pointers to some of

0:08

the resources we talk about

0:11

on the show. We now

0:13

have links to books

0:15

and articles referenced in

0:17

recent podcasts that are

0:19

available on our website.

0:21

We also offer full

0:23

transcripts. Go to Jim

0:25

Rut Show.com. That's Jim

0:28

Rut Show.com. Today's

0:32

guest is John Rob. John is

0:34

a leading thinker in military, political,

0:36

and geopolitical strategy. He's a

0:38

former Air Force pilot and

0:41

special ops guy and one

0:43

of the first analysts for

0:45

that new fangled internet thing

0:47

back at Forrester research. He's

0:49

an entrepreneur, a writer, and

0:52

a thinker. Today he has

0:54

a great substak that I

0:56

subscribe to. John Rob at

0:58

substak.com. Welcome! Thanks for having me

1:01

back. Yeah, John's been on too

1:03

many times the county's one of

1:05

our go-to guys and we're thinking

1:07

about strategy and interesting ways of

1:09

looking at the world. He most

1:11

recently was on an EP 254.

1:13

We talked about one of his

1:15

sub-stack essays. What went wrong with

1:17

America? Well worth reading. Today we're

1:19

going to talk about his substack

1:21

essay, Blitzing DC, where we're going

1:23

to talk about the idea of

1:25

network organizations taking control. So John,

1:27

I'm going to jump around a

1:29

little bit, change the orders, a tad

1:31

in your essay. Why don't we start

1:34

with the early roots of network warfare

1:36

from the Iraq insurgency? Okay, if I

1:38

could back up even a little bit

1:40

more. I mean, what we're seeing is the

1:43

emergence of a way of

1:45

making decisions with networks. and

1:47

doing that within an organizational

1:49

structure. And that's being driven

1:51

by networking, changing the way we

1:54

think and the way we organize

1:56

society. It's kind of McLuhan-esque, you

1:58

know, the medium. is the message it's

2:01

rewiring us, just like the printing press

2:03

rewired us and changed society as a

2:05

result. You know, constitutional government, everything else,

2:08

those printed documents that we use as

2:10

the basis for organizing society. So what

2:12

I saw in Iraq, you know, being

2:15

special ops and having my experience as

2:17

internet analyst is that the way the

2:19

insurgency was operating was different than the

2:21

analysis that was coming out of DC,

2:24

coming out of the DOD, coming out

2:26

of the NSA and CIA, and it

2:28

was operating differently, and it was making

2:31

it very hard to run a counterinsurgency.

2:33

And some of the core dynamics was,

2:35

you know, some of the differences for

2:38

instance were instead of like one large

2:40

insurgent organization, maybe two that mirrored the

2:42

nation state, you know, with the political

2:45

arm and a military arm, etc. That

2:47

you could roll up, you know, a

2:49

big pyramid organization that you could roll

2:51

up by taking the... people at the

2:54

bottom flipping them and going up to

2:56

the top. We saw 70 different groups.

2:58

I mean, each of these groups had

3:01

a motivation for fighting. Some were jihadi,

3:03

all sorts of different flavors of jihadi.

3:05

Some were prosodom, some were anti-sidam, nationalists.

3:08

They were criminal organizations, tribal organizations. Each

3:10

group was paper thin. So you could

3:12

roll them up very quickly, but somehow

3:15

they were coordinating to take on the

3:17

US military and keep it at bay.

3:19

So I started piecing together how it

3:22

was working and that was my first

3:24

insight into the networked organization as an

3:26

emergent organizational structure. That makes sense. It

3:28

was quite interesting. Here we are the

3:31

most powerful country on earth with some

3:33

powerful allies and we couldn't really cleanly

3:35

defeat, you know, a bunch of ragtag

3:38

dudes. It was, as you say, it

3:40

was a swarm. You'd smack this one.

3:42

Another one would pop up and the

3:45

survivors from this would join with that

3:47

one. And they could coordinate even though

3:49

they were enemies, right? And of course,

3:52

this is a famous Middle Eastern saying,

3:54

the enemy of my enemy is my

3:56

friend, right? Exactly. At least operation. And

3:58

so the next time we saw this

4:01

really pop up or more or less

4:03

simultaneously with the Arab Spring and Occupy.

4:05

Well the organizational type that I identified

4:08

in Iraq and we saw it also

4:10

in Nigeria was a kind of open

4:12

source dynamic is that there wasn't a

4:15

hierarchy that these groups were all united

4:17

based on a single what I call

4:19

plausible promise was the name from open

4:22

source software, a single unifying goal. It's

4:24

usually a very general, very simple goal

4:26

that everyone can agree with, you know,

4:29

70, 80% of the people, given country

4:31

or whatever, could agree with that one

4:33

goal. And they all worked together to

4:35

try to realize that goal. And each

4:38

had a different reason why they were

4:40

trying to achieve it and different ideas,

4:42

what that actually meant when they did

4:45

achieve it, what to do afterwards. That

4:47

allowed them to coordinate and they used

4:49

a kind of very sloppy kind of

4:52

throw it against the wall, see what

4:54

works. If it works, it's reported by

4:56

the press and other things and it

4:59

starts propagating very quickly. Everyone just copies

5:01

the success. Innovated the US. US had

5:03

a $3 billion counter IED program and

5:06

these guys were running circles around it.

5:08

They'd have fixes within two weeks. I

5:10

think 30 times faster in terms of

5:12

their innovation rate than the IRA. So

5:15

this was like, wow. And then I

5:17

wrote up a New York Times op-ed

5:19

on this and saying, you know, here's

5:22

how you defeat it is you hit

5:24

it in different directions and by using

5:26

the Shia militias as a kind of

5:29

anvil. And that worked, that kind of

5:31

died down after the Iraq war kind

5:33

of subsided. And then we saw it

5:36

again in the Arab Spring, the same

5:38

open source dynamic in this case in,

5:40

you know, the best example that was

5:42

getting rid of Numbaric in Egypt and

5:45

that unifying goal united a huge protest

5:47

moment. and it was done largely online

5:49

and they kicked it off and no

5:52

matter what they did they couldn't unravel

5:54

it because there were so many people

5:56

that wanted that similar goal even though

5:59

that every single person that protest probably

6:01

had a different idea as to what

6:03

removing Mubarak did. The leadership, there wasn't

6:06

really any leadership, and anytime they had

6:08

an influential figure come in and try

6:10

to take control of it and say,

6:13

okay, we should be talking about constitutional

6:15

reform or this is what we have

6:17

to do, they were kind of pushed

6:19

aside. No one really wanted to talk

6:22

about all that stuff. Really, they wanted

6:24

to follow anybody who was leading them

6:26

towards removing Mubarak. We saw a little

6:29

bit of that in the US with

6:31

the Occupy movement and the Tea Party,

6:33

like similar open source dynamic and protest

6:36

structure. So we find with all of

6:38

these is that, you know, they can

6:40

create massive crowds, they can swarm the

6:43

streets, and that it just is all

6:45

the typical ways that you would use

6:47

to unravel a protest, like negotiating with

6:49

them, giving them a couple concessions, didn't

6:52

work because you couldn't give enough concessions

6:54

to everybody. to unwind it. You might

6:56

take out a teeny piece of it,

6:59

but you wouldn't be able to take

7:01

out the entire group. Yeah, let me

7:03

respond a little bit on Arab Spring

7:06

and Occupy in particular. What are the

7:08

things you emphasized is empathy triggers. You

7:10

know, famously the Arab Spring was triggered

7:13

when a merchant who had been treated

7:15

unfairly by the police, I think, burned

7:17

himself to death in Tunisia and then

7:20

at one of these high... emotional valence

7:22

things, occupying it was less a significant

7:24

event, but the trauma of the first

7:26

really devastating financial crisis in the lifetimes

7:29

of most of the people that were

7:31

participating with the Great Recession, which is

7:33

now already receded into the rearview mirror,

7:36

but it was pretty scary at the

7:38

time. There was two weeks there where

7:40

it looked like the whole house of

7:43

cards was coming down. But this is

7:45

an interesting thing, is that both of

7:47

those basically failed. You know, the Arab

7:50

Spring, yeah, they threw out Mubarak, but

7:52

they got CECI, an even worse dictator.

7:54

And I did throw at Gaddafi in

7:56

Libya, but now they got chaos, right?

7:59

They might, you know, hard to say

8:01

if they wouldn't have been better off

8:03

with Gaddafi. Occupied, fizzled out, basically. One

8:06

thing to the memetic stage was the

8:08

99 versus the 1. So that was

8:10

their one victory. But it didn't really,

8:13

it didn't last. So, you know, this

8:15

shows the weakness in some sense of

8:17

the networks of network warfare, at least

8:20

in terms of long-term staying power against

8:22

an entrenched power. Yeah, I mean, it

8:24

swarms have had a spotty record of

8:27

success, you know, these warm protests. empathy

8:29

triggers played a big part. I mean,

8:31

the opportunity triggers come out of, you

8:33

know, I dug into empathy and tried

8:36

to figure out exactly what it was,

8:38

and I found out it isn't sympathy,

8:40

it isn't what we think it is.

8:43

It's really just a mental modeling of

8:45

a victim. And it's a high information

8:47

flow event. When you have empathy for

8:50

somebody being victimized in front of you,

8:52

like George Floyd or the man burning

8:54

in that, you know, street peddler that

8:57

was immoleated himself in Tunisia. You feel

8:59

like you're then and you mentally modeled

9:01

them and their enemies are your enemies

9:03

for that moment onward and that other

9:06

people who do the same thing are

9:08

united together. It forms a kind of

9:10

a loose tribalism that serves as the

9:13

basis for swarm developing. You know, empathy

9:15

works particularly well online is that we

9:17

don't have a lot of the filters

9:20

that we would have in real life

9:22

for empathy. I mean, in order to

9:24

function you can't be empathetic to everything

9:27

that's going on with everybody else because

9:29

you're you'll be totally incapacitated. Imagine a

9:31

doctor or nurse trying to get through

9:34

their own day if they had to

9:36

empathize with every single patient to the

9:38

extreme levels that would be necessary to

9:40

do it. So we have filters in

9:43

real life but online we see it

9:45

like it hits us viscerally when we

9:47

see a video or a picture or

9:50

something else coming in on our feed

9:52

and we feel it at a deep

9:54

level and it connects us and we

9:57

we can't filter it out or it's

9:59

hard. So yeah, empathy triggers is one

10:01

of the mechanisms that can, you know,

10:04

create a swarm. We saw, you know,

10:06

saw that at work with the swarm

10:08

in response to the invasion of Ukraine,

10:11

and that kicked off. In that regard,

10:13

you know, escalated the conflict, you know,

10:15

from a regional war to a new

10:17

Cold War. And so now swarms are

10:20

starting to get like dangerous because we

10:22

weren't in control of that. That was

10:24

just a kind of a group mind,

10:27

you know, reframing the conflict as a

10:29

new struggle against an emergent Hitler or,

10:31

you know, a threat to the world

10:34

and we had to go back to

10:36

Cold War thinking. We disconnected them all

10:38

at once and isolated and pushed them

10:41

towards China. So. At a strategic level

10:43

for the United States, that was a

10:45

terrible thing, but we didn't really want

10:47

Russia with China together, but the swarm

10:50

pushed it. We'll talk about Ukraine later,

10:52

because I think we have quite a

10:54

different point of view on Ukraine, but

10:57

that's all right. So let's move through

10:59

the historic timeline here. The next big

11:01

example that you give, and this is

11:04

where it becomes, I'm called hybrid swarm,

11:06

maybe. You don't use that term, but

11:08

I'm going to throw it out there.

11:11

And that's the Trump 2016 campaign. Our

11:13

mutual friend Jordan Hall did a really

11:15

interesting deep dive on the red religion

11:18

versus the blue church. Tell us a

11:20

little bit about your take on the

11:22

2026 campaign and how that was a

11:24

new variant on network insurgency. Yeah, Jordan

11:27

took a, yeah, it's classic because he

11:29

did a reformation kind of framework for

11:31

this where you have the wild Protestants,

11:34

you know, fiery Protestants, each with their

11:36

own Bible, each with their own little

11:38

variations on the church versus the kind

11:41

of universal Catholic church, which was sparked

11:43

by, you know, the printing press. Martin

11:45

Luther did. spark it off because he

11:48

posted his thesis is because it was

11:50

printed afterwards and spread out, right? So

11:52

we have the same thing with this

11:54

is that, you know, the networking is

11:57

that what we saw with the Trump

11:59

campaign is that we took this kind

12:01

of open source dynamic and applied it

12:04

to politics. And there were a lot

12:06

of people who were not happy with

12:08

the way things were going and how

12:11

things were this globalization of the United

12:13

States. was a playing out and this

12:15

kind of globalized perspective was not yielding

12:18

the kind of prosperity, but they saw

12:20

a lot of fraud. They saw it

12:22

a lot of corruption. They didn't feel

12:25

like they're making progress and all sorts

12:27

of other reasons. And that Trump represented

12:29

a kind of possible promise, a unifying

12:31

goal. He was a grenade, a political

12:34

grenade that could stick in the White

12:36

House to disrupt stuff. And in that

12:38

regard, he really wasn't a candidate in

12:41

the traditional sense. He was an idea,

12:43

bomb disruption. insurgency formed around them to

12:45

get them into office. And they pushed

12:48

them and pushed them into innovative, constantly

12:50

adopted a form of political maneuver. It

12:52

was maneuver-based operations where they were constantly

12:55

shifting topics. Trump's natural style, our natural

12:57

style, kind of fed into that. He'd

12:59

switch topics constantly, be disruptive, make big

13:01

bold claims, in one direction or another,

13:04

or faint, went in one direction another.

13:06

drawing the opposition off, he confused her

13:08

psychology, disrupted the establishment, and it was

13:11

successful. Got him into office. But he

13:13

didn't have any staying power. So as

13:15

soon as he got into office, the

13:18

objective was complete, the insurgency kind of

13:20

melted away, and he was there alone,

13:22

just, you know, throwing punches, and it

13:25

was crushing him, and he wasn't really

13:27

able to do as much disruption as

13:29

he possibly could. But they, you know,

13:32

achieved what they wanted. Yeah, it is

13:34

interesting that the Trump administration really achieved

13:36

very little. The only really substantive thing

13:38

it did for the perspective, at least

13:41

some of its backers, was the big

13:43

tax cut. And of course, that was

13:45

not very populist. You know, what was

13:48

the most favored group in the tax

13:50

cut? Real estate developers. There's a real

13:52

shock for you. And, you know, it

13:55

wasn't horrible for average Joe's, but it

13:57

was certainly waited towards the fat cats.

13:59

And so, you know, you know, not

14:02

populist in the slight. Yeah, no, its

14:04

policies really didn't matter. I mean, there

14:06

were a lot of, you know, things

14:08

like, remember the Access Hollywood video and

14:11

stuff like that. that other stuff. I

14:13

mean, everybody came to be able to

14:15

like would say, oh, he's gone, he's

14:18

dead, he's like, by all conventional measure,

14:20

he's gone. He's finished. For about the

14:22

20th time, right? Yeah, but this one

14:25

was like, had to be it. I

14:27

go, this has to be it, he's

14:29

done, right? And I was like, no,

14:32

he's not a candidate. He's a weapon,

14:34

he's a grenade. I mean, they don't

14:36

judge him based on this, you know,

14:39

you know, they want him, to be

14:41

disruptive, to be disruptive, to be disruptive,

14:43

like that, like that, like that, like

14:45

that, like that, like that in there.

14:48

Let's say they. Now this is interesting

14:50

because there is no they. It's not

14:52

like there's like four guys sitting in

14:55

a room saying we're going to stir

14:57

this shit up do you think or

14:59

is this a true populist network vibration

15:02

that forms these this particular network? Yeah

15:04

I mean it was open source there

15:06

was lots of different people with lots

15:09

of different motivations for joining it. And

15:11

you can see it on read it,

15:13

you can see it on other places

15:16

I did that whole, you know, the

15:18

most popular group that was promoting Trump

15:20

on read it. They did an interview

15:22

with me the week after they did

15:25

Trump. because I was the Donald was

15:27

that what it was yeah I was

15:29

doing a statistical analysis on the Donald

15:32

in real time and I calculate what

15:34

was the name of their their little

15:36

god that they produced I would calculate

15:39

the the ratio of phrases versus their

15:41

little god and to show the rise

15:43

and fall memes in the Donald it

15:46

was very interesting it was an idea

15:48

factory it was it was a it

15:50

was a clearinghouse for beams and other

15:52

ideas and it was completely outside of

15:55

the control of the campaign the fact

15:57

actually Actually, this is an interesting thing.

15:59

This is the insight I had from

16:02

Jordan's insight and some research I did.

16:04

It wasn't quite the meme factory, it

16:06

was the beam amplifier. The memes were

16:09

coming out of 4-10. And then they

16:11

would get stuck into the Donald. Some

16:13

would get upregulated and improved. And then

16:16

those would catch the eye of somebody

16:18

in the Trump administration or Fox News,

16:20

and then they'd get broadcast out, and

16:23

then they'd get recycled into a new

16:25

generation. So it was like, exactly. It

16:27

was really a wild emergent system for

16:29

metic propagation. That didn't really get recreated

16:32

in 2020. or 2024, it seemed like

16:34

it was a one-time thing, this unintentional

16:36

emergent meme cyclone that got going there

16:39

in 2016. Yeah, I mean it had

16:41

to do with the fact that we

16:43

were moving towards, you know, Packatized media.

16:46

And the Packatized media broke everything down

16:48

all, you know, little videos, little pictures,

16:50

little takes on stuff, you know, little

16:53

posts that encapsulated various parts of a

16:55

description of an event. that packetized flow

16:57

is massive. I mean, it's torrential. What

16:59

they were doing, using this mechanism, was

17:02

that they were easing together into patterns,

17:04

patterns of understanding, sense making, and those

17:06

patterns are basically those memes and then

17:09

and other things, and they put them

17:11

into a, you know, up regulation system,

17:13

they were, you know, they were voted

17:16

up or voted down or, you know,

17:18

modified, and it was kind of a

17:20

little more formalized than what we see

17:23

in the open social media. Then it

17:25

ended up at the top where they

17:27

were being used by Fox and Trump

17:30

and others as weapons. And it was

17:32

an amazing thing to see from the,

17:34

you know, the ground, you know, seeing

17:36

it from, you know, my interviews with

17:39

them. I mean, I wasn't fully, you

17:41

know, pro-pro, everything Trump does as a

17:43

god. And then my interview on the

17:46

top, but I talked about the dynamics

17:48

and they were exactly, you know, following

17:50

this pattern. Yeah, I was watching it

17:53

with great interest as a student of

17:55

network dynamics, though also, but in my

17:57

particular case, a foe of Trump. But

18:00

nonetheless, I had to admire the ad

18:02

hoc creation of this engine, and I

18:04

thought it fairly likely that it would

18:06

work, and it did. So Trump gets

18:09

in, and Red has made a move,

18:11

right? learned how to or has discovered

18:13

accidentally or something network warfare was okay

18:16

at it, you know, unprecedented actually in

18:18

American politics, but now Blue strikes back

18:20

with their counter network. Let's tell us

18:23

about that. The Red Insurgency rolled over

18:25

the Republican Party, rolled over the Democratic

18:27

Party, got him into office. The blue

18:30

network started to form and in response

18:32

to Trump, and it took on a

18:34

different dynamic, a different way of actually

18:37

organizing it based itself on a kind

18:39

of moral standard, a common moral standard.

18:41

It was a collective kind of sense-making,

18:43

very different than the kind of chaotic

18:46

maneuver-based. sensemaking we saw in the Red

18:48

Network. And the Blue Network grounded a

18:50

lot of their moral standards-based approach on

18:53

being against something evil, like against racism,

18:55

against sexism, against colonialism, against, you know,

18:57

something that everyone agrees on is evil.

19:00

And then they started to use that...

19:02

moralism, that kind of moral standard as

19:04

a weapon against everything that Trump did

19:07

and everything that the red network did.

19:09

They started convincing very quickly that, you

19:11

know, the people running the social networks

19:13

that these, you know, people in the

19:16

red insurgency were evil. They're doing something

19:18

morally reprehensible, were a threat, an existential

19:20

threat to everybody. And the social networks

19:23

started banning them, started censoring, started squeezing

19:25

them to the side. That was really

19:27

the big play in 2020, is that

19:30

Biden didn't even run a campaign. By

19:32

all measures, he just sat there, did

19:34

nothing. What did happen is this network

19:37

squeezed the red network down to almost

19:39

nothing, censoring it, and I was worried

19:41

that we were headed towards an era

19:44

of universal censorship and control, particularly as

19:46

AI was jending up to make it

19:48

even easier to do. which I called

19:50

the long night, you know, one orthodox

19:53

way of looking at the world that

19:55

would prevent anyone from having any new

19:57

ideas or challenging the the Orthodox. And

20:00

that control of the social networks, if

20:02

you remember looking back at 2016, Trump

20:04

was everywhere on Facebook and Twitter and

20:07

everywhere else and there was, but in

20:09

2020. He was gone. You know, there

20:11

was barely any kind of flow and

20:14

people were banned and censored and that

20:16

led to the victory of the Blue

20:18

Network. But interestingly, let's talk a little

20:20

bit about the governance of the Blue

20:23

Network and also at about the same

20:25

time just before the emergence of the

20:27

Victory of the Blue Network, the very

20:30

curious George Floyd thing. You know, I've

20:32

researched the George Floyd situation and there

20:34

actually is zero excess... Killings by police

20:37

of black men. It's exactly proportionate to

20:39

the crime rate of black men, as

20:41

it is for white men, as it

20:44

is for Hispanic men, as it is

20:46

for Asian men. It's exact. Two, one

20:48

percent. Roland Fryer's much more detailed work

20:51

with the Houston police also says that

20:53

per encounter with police, if anything, less.

20:55

black men are killed by police than

20:57

you'd estimate based on the number of

21:00

encounters. There's a completely fallacious argument, but

21:02

talk about the empathy. It was a

21:04

horrific act and a horrific video of

21:07

George Floyd being suffocated by a road

21:09

cop and it somehow... resonated with people

21:11

to get them to believe something that

21:14

was just manifestly not true and yet

21:16

it swept the country killed 20 people

21:18

billions of dollars of damage very very

21:21

very interesting and curious movement you know

21:23

based on false premises basically correct yeah

21:25

and the numbers are very very small

21:28

you know we're talking like 20 is

21:30

a delta even if you don't correct

21:32

it for a crime rate or socioeconomic

21:34

status poor neighborhoods tend to have more

21:37

crime. Yeah, no, there wasn't a factual

21:39

basis, but there was a kind of

21:41

a sense with the kind of emerging

21:44

or growing lack better placid. They were

21:46

tired of it, you know, being treated

21:48

like, you know, criminals and getting stopped

21:51

and pulled over. That might have played

21:53

a factor, but frankly, it was the

21:55

empathy trigger that kicked it off, and

21:58

it was a way to oppose. Trump

22:00

as well. And the Blue Network framed

22:02

it that way and it zoomed. We

22:04

had a swarm in the streets,

22:07

billions of damage, a dozen people

22:09

killed, 2,000 cops in the hospital,

22:11

and they didn't even aggregate

22:13

the stats on this. I mean

22:16

there was a resistance within the

22:18

establishment to kind of hide everything

22:20

and bury it. It wasn't any

22:22

coverage of the kind of violence

22:25

in the protests by the major

22:27

media. And they saw as a necessary...

22:29

You know, they, they, you know, mostly peaceful,

22:31

that kind of language. And

22:34

that kind of excess was overlooked

22:36

once with this, and then, you

22:38

know, the problem with a kind

22:40

of a moral framework that's a,

22:42

if you don't agree with this,

22:44

you're evil, your enemy, your

22:47

existential threat, is that it

22:49

can lead to excesses. And

22:51

the Blue Network was incapable

22:53

of policing the excesses of

22:56

the excesses of the framework. particularly

22:59

once they won and they started applying

23:01

it. So, you know, they were able to pull

23:03

in the overreach of that. So just like

23:05

a tiny example, more, you know, kind

23:08

of an example of it is that, you

23:10

know, the blue networks approach to trans

23:12

and going all the way down to

23:14

sports and kids and other things

23:16

and not being able to police

23:18

that and saying, okay, this is

23:20

a mix. You know, what you're

23:22

talking about nine-year-olds. You know, it's,

23:24

of course, generated a countera counter

23:26

reaction. Yeah, that was quite interesting. I

23:29

was actually part of that. The, uh,

23:31

helped form the MIT Free Speech Alliance,

23:33

which is part of the alumni free

23:36

speech alliance quite early on. I'm trying

23:38

to remember quite when that was, I

23:40

think it was, uh, 2021, when the,

23:42

when. We just thought this had gone

23:45

too far. I mean, certainly there is

23:47

still injustice in America, which needs to

23:49

be remediated. Black citizens still don't get

23:52

a fully fair deal in our society.

23:54

It's a hell of a lot better

23:56

than it was when I grew up,

23:58

and I still remember. seeing a

24:00

white's only water fountain when

24:03

I was a kid, which

24:05

was pretty disturbing. But there's

24:07

still a little bit left,

24:09

but it's nothing like the

24:11

screaming that you got. And

24:13

then to insist on censorship,

24:16

and in particular, Dorian Abbott

24:18

came to MIT, was invited

24:20

to a endowed speech on

24:22

exoplanet atmospheres, very esoteric topic,

24:24

but he had written with

24:27

another scholar up. public popular

24:29

piece in Newsweek, basically criticizing

24:31

DEAI, and suggesting replacing DEA

24:33

with MFE, merit, fairness, and

24:35

equality. It was a very

24:37

level-headed piece, but he had

24:39

violated the taboo of the

24:41

time that you can't criticize

24:43

DEA. And so a bunch

24:45

of people, MIT graduate students,

24:48

swarmed on Twitter. We can't

24:50

have this guy give me

24:52

this speech at MIT. And

24:54

unfortunately, the head of the

24:56

department that had invited him

24:58

with the support we believe

25:00

of the administration succumbed and disinvited

25:02

him. And we all thought that

25:04

was gone too far. You can't

25:06

criticize something. So we launched the

25:08

MIT Free Speech Alliance, and it

25:11

quickly got lots of members and

25:13

got lots of influence. And there's

25:15

now hundreds of these free speech

25:17

alliance all under the American. alumni

25:19

free speech alliance umbrella and I

25:21

felt that we were right at

25:23

the tide turning and I told

25:25

people that at the time. It feels

25:28

to me that they have overreached and

25:30

now the tide is starting to flow

25:32

the other way. Yeah, no, it wasn't

25:34

just, you know, universities. I

25:36

mean, you know, corporations got pulled

25:38

into it. I mean, you saw

25:41

what happened after January 6th. I

25:43

mean, they came together to disconnect

25:45

Trump from social media in

25:47

all at once. and you

25:49

know, kind of a corporate

25:51

takeover of all media and

25:53

that anyone who was promoting

25:56

ideas that were counter

25:58

to this kind of. blew a

26:00

narrative structure were being disconnected and marginalized

26:03

and pushed out and the resistance grew.

26:05

I mean, you know, because it is

26:07

excessive. I mean, to go after everything

26:09

and everybody and in the stifle free

26:12

speech at that kind of level, you

26:14

know, edging towards what I thought was

26:16

a long night. And I can tell

26:19

you personally listeners to the show know

26:21

I've been involved with the game B

26:23

effort for a long time and we

26:26

had a we still have a very

26:28

popular group on Facebook and the day

26:30

after Biden was inaugurated the Facebook algorithm

26:33

killed the three admins of our Facebook

26:35

group including me with death penalty unappealable

26:37

the lesions of our account. And for

26:39

no reason at all. If you go

26:42

to the Game B group, Game B

26:44

on Facebook, you'll find it's a very

26:46

benign, good citizen, nice kind of thing,

26:49

and the people are very well behaved.

26:51

There's absolutely no reason for it to

26:53

do so. But they had obviously turned

26:56

the knob on their algorithm up the

26:58

day that Biden got inaugurated. and they

27:00

whacked us and fortunately we had we

27:03

had some noisy friends including Joe Rogan

27:05

who put up quite a loud yell.

27:07

I think we got seven million views

27:10

of our announcement of this. thing and

27:12

we knew some people who knew people

27:14

and Facebook and 12 hours later they

27:16

reversed it. But for a quite, you

27:19

know, yes, maybe we're a bunch of,

27:21

you know, Blue Sky, utopian, quasi-utopian, hippie

27:23

thinkers, but we're certainly not anything bad,

27:26

but they were so anti-heterodoxy of any

27:28

sort at that point that that convinced

27:30

me that this now had to be

27:33

fought full on. Yeah, no, the only

27:35

way you could, I mean, you don't

27:37

know, it was one of these situations

27:40

where people would get banned and they

27:42

didn't even know why they were banned

27:44

and the groups would be banned and

27:46

they didn't know what triggered it. The

27:49

ones we were hit with was no

27:51

appeal. The same happened to Brett Weinstein.

27:53

No appeal, and if you try to

27:56

appeal, it says, no appeal, you're dead.

27:58

I go, what the fuck kind of

28:00

sick Kafka-ass bullshit is this, right? Yeah,

28:03

no, this way, you know, when I

28:05

was talking about the long night stuff,

28:07

I was saying, okay, how we guard

28:10

against this years ago, you get some

28:12

kind of digital rights to kind of

28:14

make your version of yourself, your digital

28:16

self, protected, right? Make it more real.

28:19

And that you could be banned for

28:21

excessive kind of threats of threats and

28:23

personal stalking and that kind of stuff.

28:26

Anything that would cause you to be

28:28

muted or some kind of charge against

28:30

you you had to be informed of

28:33

it and there had to be a

28:35

process I laid all this out in

28:37

great detail in a article I wrote

28:40

for Quillit called Musk in Moderation where

28:42

I even allowed a marketplace. You'll love

28:44

this. If they had to ban you

28:47

based on a rule, a regulation in

28:49

their legislation, they had to quote what

28:51

you said and the relevant section of

28:53

their legislation. And if you wanted to

28:56

challenge it, you could, you'd take it

28:58

to an American Arbitration Association arbitrator, and

29:00

you could put up a steak. And

29:03

if they lost, they had to pay

29:05

you 10X. So if you thought you

29:07

had been, your post, that they said

29:10

violated section 1.3.7.1, the arbitrator's rule did

29:12

not violate 1.3.7.1, then you put up

29:14

1,000 bucks, then they had to pay

29:17

you 10,000. And even better, you could

29:19

syndicate your bets. So you could say,

29:21

all right, here is a marketplace for.

29:23

bets against Facebook's censorship, I put a

29:26

limit of a million dollars so that

29:28

you could syndicate your claim with the,

29:30

with the complainant getting 20% and the

29:33

betters getting 80% of the win. So

29:35

even if you had no money at

29:37

all, you could still put together a

29:40

big war chest for a particularly weak

29:42

claim. And I thought that'd be a

29:44

great self-policing system, because, you know, Facebook

29:47

doesn't want to lose 10 million dollars.

29:49

I had making a wrong decision on

29:51

something like this. Well, I mean, it

29:54

was going towards that kind of long

29:56

night scenario to pretty rapid clip until

29:58

2022. It's getting worse and worse and

30:00

worse, more oppressive. You know, we saw

30:03

the terms of service replace the bill

30:05

of rights and, you know, those terms

30:07

of service could change any moment that

30:10

Zuckerberg or anyone else wanted to kind

30:12

of flip them, tweak them. And that

30:14

was the key part of the term

30:17

of service. We can change this at

30:19

any time with or without your permission.

30:21

Yeah, I imagine a bill of rights

30:24

based like that, right? The president could

30:26

change the bill of rights any time

30:28

you, you know, you know, saw a

30:30

need. And the people that they had

30:33

selected to kind of moderate or kind

30:35

of arbitrate. changes to that term of

30:37

service were all from activist groups and

30:40

it was just nutty and it was

30:42

getting it's getting more more restrictive the

30:44

you can't talk in the major in

30:47

the major media was supporting it and

30:49

and social networking was suppressing it everything

30:51

turned around I mean any kind of

30:54

turnaround you thought was coming at the

30:56

grassroots level was going to get squashed

30:58

eventually as he's especially as you know

31:00

If it progressed, then AI would arrive

31:03

and then it would be employed to

31:05

kind of make, moderate all those billions

31:07

of people on Facebook and others at

31:10

a very deep level simultaneously. So the

31:12

change happened with Musk acquiring Twitter. I

31:14

would say it happened a little earlier.

31:17

It happened with a strong sense across

31:19

society that this had overreached by that

31:21

summer of 2021. And I would say

31:24

that Musk was then the next step.

31:26

Yeah, I mean, I think there was

31:28

a lot of people, there was a

31:31

lot of like brewing discontent with it,

31:33

but I do think that the system

31:35

as it was set up, the kind

31:37

of the way that the networks were

31:40

operating, was that it was going to

31:42

crush that, and it could well have,

31:44

without some big change, that any kind

31:47

of discontent would eventually have been squeezed

31:49

out and anyone who expressed this content

31:51

would become evil by definition and existential

31:54

threat. But I think the thing with

31:56

Musk was that the trigger for him

31:58

was watching that swarm response to Ukraine.

32:01

And my first response is, holy shit,

32:03

this is dangerous shit. It's bringing us

32:05

up to global war and pushing us

32:07

up. People were like, this swarm. thinking

32:10

was that we have to win at

32:12

all costs, we have to defeat Russia,

32:14

we have to depose Putin, Putin was

32:17

tied to Trump, he's evil, and therefore

32:19

Putin is the ultimate evil, just like

32:21

Trump is, therefore we have to completely

32:24

cause a regime change in Russia, and

32:26

you know nuclear war doesn't matter anymore,

32:28

it's just threats, it just bluffs, you

32:31

know, all the additional, you know, the

32:33

way we handled that in the Cold

32:35

War, and prevented a World War III

32:37

with nuclear weapons, it doesn't matter anymore,

32:40

and... I was like, holy, just step

32:42

back a little bit here, let's get

32:44

a little bit more rational in our

32:47

approach. Must saw that, and he said

32:49

in an interview, that was the reason

32:51

that he decided to make the bid.

32:54

There's no circuit breakers. There's no way

32:56

of controlling, and it's frustration that being

32:58

kind of edited and censored, but that

33:01

was the thing that kicked it over,

33:03

and once Twitter was acquired, everything changed.

33:05

In November of 2022, right. Let's go

33:08

back to Ukraine, though, for a moment.

33:10

I didn't necessarily see it as a

33:12

swarm. I know you did at the

33:14

time, a pro-Ukraine support. I thought of

33:17

it more of an elite action from

33:19

the traditional internationalist parts of both parties.

33:21

There was the Neicons that kind of

33:24

had seized control of the aimless United

33:26

States and brought us into wars in

33:28

the Middle East and then started expanding

33:31

NATO with the intent of completely encircling

33:33

Russia. And as we got closer and

33:35

closer to the borders, they started to

33:38

be... more reactive, just like, you know,

33:40

Ken and wrote in the late 90s

33:42

that, you know, this would eventually turn

33:44

a Democratic capitalist society. It was just

33:47

emerging into an enemy. It was an

33:49

inevitable kind of freight train, and it

33:51

did. You know, as we hit their,

33:54

you know, got right to their critical,

33:56

right around their border, they started to

33:58

have more resistance and more military response.

34:01

But those Neicons were in charge of

34:03

things and had us on this kind

34:05

of pathway that was inexorable, towards a

34:08

conflict with Russia. The swarm response was

34:10

different for me is that it was

34:12

based on Putin's connection to... Trump and

34:15

a lot of the kind of hostility

34:17

and attempt to kind of punish Putin

34:19

for getting Trump into office, you know,

34:21

they came to office, said, you know,

34:24

we're going to really make it tough

34:26

for Putin because he got Trump in.

34:28

They sparked it, but the swarm response

34:31

to it was so much bigger. It

34:33

was, I mean, all these little companies,

34:35

I mean, everybody, disconnecting simultaneously in just

34:38

weeks. And the whole framing of it

34:40

turning from, you know, we're opposed to

34:42

Russian invasion, which happened before, and we

34:45

didn't respond like this, it turned into

34:47

a new Cold War. I think that's

34:49

overstated. You know, I make this point

34:51

regularly. Russia has the GDP of Spain

34:54

or Canada. It's not a... giant country

34:56

anymore. It no longer has the whole

34:58

eastern block as its allies, right? It

35:01

no longer has the rest of the

35:03

USSR. This idea of painting Russia as

35:05

this amazing threat is basically bogus. Which

35:08

actually then says it's perfectly reasonable to

35:10

stop Putin and Ukraine because he is

35:12

stopable and he doesn't really have a

35:15

credible threat to do anything else. He's

35:17

not going to invade the Baltics. And

35:19

he does. We'd whip his ass and

35:21

heartbeat. And the Ukraine war is shown

35:24

the Russian army. He's a paper tiger.

35:26

Couldn't beat the 22nd rated military in

35:28

the world in a flat-out fight. Ridiculous.

35:31

So I don't buy a new Cold

35:33

war. I think the issue. is that

35:35

it's a huge mistake to let Putin

35:38

win a war of territorial aggression in

35:40

the 21st century. And that he needs

35:42

to be stopped for that reason. But

35:45

we provoked it. I mean, we offered

35:47

them, we offered them NATO membership, pushed

35:49

it, the name of membership to kind

35:52

of, you know, punish Putin for support

35:54

of Trump. And they said it was

35:56

a red line, any more than we

35:58

would say it was a red line

36:01

if China's decided to make ally with

36:03

Mexico and start putting troops and weapons

36:05

there. We wouldn't. Remember, the USSR did

36:08

the same thing in Cuba, and we

36:10

managed to avoid war there. You know,

36:12

by the Nick. By the Nick. Yeah.

36:15

A lot of mistakes that were kind

36:17

of, wow, I don't want to, I

36:19

just don't want to see a repeat

36:22

of that. I mean, they still have

36:24

more nukes than we do. And as

36:26

much as there are paper tiger and

36:28

everything else, even if half those nukes

36:31

don't work. It's still not something to

36:33

do. They're not going to use their

36:35

nukes though. It's insane. It's not a

36:38

credible threat. Well, it depends on whether

36:40

they see it as an existential threat.

36:42

I mean, having Ukraine in NATO, they

36:45

considered existential. I doubt it. I think

36:47

it's geopolitical chess playing. They lose some

36:49

ground if Ukraine becomes EU and NATO.

36:52

But it's not an existential threat. NATO's

36:54

never attacked anybody. put troops and nuclear

36:56

weapons in Mexico, we consider an existential

36:58

threat. We would be pissed off, but

37:01

we wouldn't consider an existential threat. Oh

37:03

yeah, we would. We would not nuke

37:05

China over that. We would build up

37:08

one hell of a big defensive army

37:10

down there, and we'd probably try to

37:12

invade them in a second. And probably

37:15

try to cut the sea lines, that's

37:17

what we'd probably do. So they cut

37:19

the sea lanes, and then invade. Anyway,

37:22

regime change and all this stuff. Anyway,

37:24

we could we could probably just we're

37:26

just going to fundamentally disagree about your

37:29

create. So let's move on to the

37:31

the red network reconfiguration after you know,

37:33

and I will say that one thing

37:35

that I'd like to add that it

37:38

really wasn't in this piece is it

37:40

wasn't just Twitter. The other thing that

37:42

was happening the same time was the

37:45

emergence of lots of other independent media

37:47

on from the team red perspective. Somebody

37:49

recently published this bubble graph that shows

37:52

the sizes of various podcasters, for instance.

37:54

The poor Jim Rutte show was too

37:56

small to show up on the graph

37:59

besides they wouldn't have any idea where

38:01

to put it. We're not either blue

38:03

or red, or we're silver, but it

38:05

was clear that the red-oriented podcasters. totally

38:08

outnumbered the Blue ones. And you know,

38:10

you talk about Fox News. Fox News

38:12

at its best draws four and a

38:15

half million viewers to its prime time

38:17

shows. Joe Rogan's interview with Trump, 57

38:19

million. His interview with J.D. Vance, 25

38:22

million. Right. And so under appreciated part

38:24

of this new red network reconfiguration is

38:26

this grassroots, entrepreneurially driven, right? People are

38:29

doing this because it's a business, but

38:31

they're also ideologically informed, you know, Bannon's

38:33

war room is very classic example, has

38:36

also fundamentally changed the media battlefield along

38:38

with. Twitter and also the alternative mostly

38:40

red networks as well. Yeah, no, that

38:42

was the red network's response to the

38:45

problem with persistence. I mean, the blue

38:47

network solved the persistence problem with these

38:49

networked organizations that they usually fall apart

38:52

once they achieve some goal or fail

38:54

it achieving a goal. And by creating

38:56

this kind of moral framework, they had

38:59

this enemy, they had a narrative that

39:01

created a sense of tribalism and unity.

39:03

that allowed them to operate as a

39:06

cohesive whole, you know, for a long

39:08

periods of time. And so the Red

39:10

Network didn't have it. It was more

39:12

loose. And what it came up with,

39:15

what it evolved into, what it merged

39:17

into, was these large accounts, these large

39:19

kind of digital ledgers, kind of a

39:22

view of the world that kind of

39:24

sense-making perspective based on individual personalities. that

39:26

were, became very popular, like Elon's Musk,

39:29

you know, 200 million plus flowers on

39:31

X. And these digital ledgers were highly

39:33

evolved. I mean, they changed, tweaked, and

39:36

improved the perspectives over time based on

39:38

user, you know, follow our feedback, pushback.

39:40

And they got really good, and they

39:42

post a ton, and they're all kind

39:45

of truth seekers. you know taking on

39:47

establishment narratives and taking on you know

39:49

what's being dictated through the media and

39:52

challenging it and coming up with alternative

39:54

sense-making perspectives and so there are a

39:56

ton of these counts and they're all

39:59

dominated by the red network I mean

40:01

there's right a little push right now

40:03

for the blue network to try to

40:06

do that they're Michelle Obama and others

40:08

but they're all fizzling because it's they're

40:10

not challenging the kind of any official

40:13

narrative and they're just not the same

40:15

dynamic so you had these big accounts

40:17

And they took the war to the

40:19

Democratic Party in the 2024 election when

40:22

Musk supported Trump in the summer. He

40:24

eliminated any kind of stigma associated with

40:26

supporting Trump, you know, which was still

40:29

there, that kind of shame, that kind

40:31

of disconnection, that kind of moral kind

40:33

of framework that the blue was using

40:36

to kind of keep people out. And

40:38

all these accounts started falling. Like, Rogan

40:40

officially came in and all these others

40:43

started to kind of give their... explicit

40:45

support for Trump and they put him

40:47

in office. You know, Trump wasn't really

40:49

running much of a campaign. I mean,

40:52

you know, the traditional sense, I don't

40:54

even think the campaign was even worth

40:56

it. It was really these accounts that

40:59

were driving. They came up in the

41:01

narratives, they ran the maneuver warfare, they

41:03

were constantly changing subjects, they were throwing

41:06

ideas out, amplifying things that Trump picked

41:08

up and ran, and that carried over,

41:10

that victory. Those big accounts didn't just

41:13

go away at the end. They were

41:15

still there and they started to take

41:17

over the positions in the government. I

41:19

mean, you know, you have a Tulsi

41:22

Cabard, you know, who has a big

41:24

account, is a true secret, podcast, constantly,

41:26

you know, it has your own ledger,

41:29

you know, it's put into, you know,

41:31

position of power, and, you know, the

41:33

same thing with all the other major

41:36

agencies for the most part. These gray

41:38

suited, unknowns, you know, you know, these

41:40

insiders that you used to use that

41:43

you used to use to use to

41:45

use to use to use to use

41:47

to use to put in use to

41:50

put in use to put in there.

41:52

vetted by the establishment as the kind

41:54

of people that would do the kind

41:56

of things that that throw money this

41:59

way and do this and maintain these

42:01

programs. Well, they were replaced by people

42:03

who are known. You know, had a

42:06

position that people, you know, they were,

42:08

they had some kind of public visibility.

42:10

And, you know, the most famous of

42:13

that is, of course, Elon Musk is

42:15

Doge, coming in with a, you know,

42:17

trying to act the system. And he

42:20

was, you know, that's how much to

42:22

disrupt it in the, I'm going to

42:24

break you sense, is finding the kind

42:26

of, what I call, system books, the

42:29

points inside the target network that will

42:31

allow you to exercise the greatest control

42:33

and influence. And he found that in

42:36

the kind of treasury data. I found

42:38

that in the OPM records, in the

42:40

communication records of the various agencies, the

42:43

email logs, and that data allowed him

42:45

to. Yep. It's happened something you mentioned,

42:47

and this is, you know, along the

42:50

lines, you're talking about these big ledgers.

42:52

the red digital warriors became more professionalized

42:54

essentially. Right, right. We're talking about 2016.

42:57

This was all hobbyists and, you know,

42:59

people in their mother's basements, you know,

43:01

typing away at three o'clock in the

43:03

morning. But now we have actually an

43:06

analog to mainstream media. You know, mainstream

43:08

media used to put people into administrations,

43:10

right? And now what we're seeing is

43:13

that it's called professionalized. You use professionalized

43:15

digital warriors. are now candidates to be

43:17

members of administrations. And many of them

43:20

are now more powerful than any mainstream

43:22

media voices. I'm sure far, far more

43:24

people listen to Steve Bannon than read

43:27

one of the. columnists in the New

43:29

York Times. This is quite interesting and

43:31

new. And Team Red has gotten their

43:33

first. To your point, Team Blue is

43:36

trying to figure out what to do

43:38

about it, but they kind of have

43:40

a hard problem. And this is something

43:43

else you talked about. So I think

43:45

it's worth surfacing. Blue, the way it

43:47

has defined itself in kind of rigid

43:50

moralistic terms, has a very hard time

43:52

adapting. As soon as you make any

43:54

change, one of the factions goes. woke

43:57

violated in some way or other. And

43:59

so that makes you very vulnerable to

44:01

maneuver warfare. And someone like the red

44:03

team having a faster oodle loop is

44:06

they can change and you can't adapt,

44:08

you get left behind. Yeah, I mean,

44:10

the red has its own problems. I

44:13

mean, it's not a unified view of

44:15

the truth or of reality. I mean,

44:17

each account has a sound kind of

44:20

unique approach. So it's kind of a

44:22

disparate in that regard. and that may

44:24

end up being what you know limits

44:27

the power of red is that because

44:29

there are very different factions you know

44:31

you have the You have the silicone

44:34

valley boys who, to my mind, are

44:36

aiming at neofutilism, probably, and then you

44:38

have the, you know, the bannonites and

44:40

the Christian nationalists are looking for theocracy,

44:43

and then you have a big chunk

44:45

in the middle of normal Republicans looking

44:47

for lower taxes and less regulation, but

44:50

basic sanity, and how those three totally

44:52

incompatible factions work things out is still

44:54

very much to be determined. Well, it's

44:57

less the factions, it's just accounts. So

44:59

you really want to kind of parse

45:01

it up is, you know, who's following

45:04

who? There is a certain amount of

45:06

influence that these big accounts have over

45:08

their followers is that they can take

45:10

them into new areas of sense making

45:13

pattern matching and then override the ledgers

45:15

of their followers. They push them and

45:17

they do it. They do it, you

45:20

know, unless they face massive resistance, they

45:22

usually goes through. So they can actually

45:24

change and shift the faction that's, you

45:27

know. Not so much faction on their

45:29

own, but people who follow them. The

45:31

blue is like interesting. I've been trying

45:34

to figure out how they're going to

45:36

oppose this. Their biggest flaw in their

45:38

kind of moral warfare. A moral warfare

45:41

usually works by constantly pointing out how

45:43

evil the opposition is and forcing them

45:45

to break into non-cooperative centers of gravity

45:47

where people say, oh, I can't be

45:50

associated with that people and start to

45:52

divide out. The problem with the blue

45:54

network was it had a maximus moral

45:57

framework. what it really needs and what

45:59

we need generally I think in a

46:01

to get a you know if we

46:04

look at the longer view of this

46:06

is we want a network decision-making at

46:08

a societal level in terms of our

46:11

politics, in terms of our culture, in

46:13

terms of society, that has a set

46:15

of standards, basic standards, both for us

46:17

and potentially what we could globalize, that

46:20

represent our values, but it's not maximistic.

46:22

So it's like, I always think of

46:24

the internet, you know, going with the

46:27

very simple TSPIP protocols that if you

46:29

do that little thing, you can connect

46:31

to us. And it spread, right? If

46:34

they went out with this kind of

46:36

maximalist kind of vision of the way

46:38

things work, and not HTML, but as

46:41

you know, with extras, and it's like,

46:43

you have to do this, people just,

46:45

oh, you know, they'll stick to their

46:47

guns, not fully connect, not fully integrate.

46:50

So we need kind of a simple

46:52

standard, and we use the kind of

46:54

red network for new ideas, new sense

46:57

making. Networks great for solving complex problems,

46:59

you know, which are. more and more

47:01

prevalent in this kind of wild system

47:04

that we're in because you need to

47:06

be able to come up with those

47:08

wild ideas that may solve the problem

47:11

where you're planning and your expertise may

47:13

prove fruitless. So we need to mix

47:15

the two. So what the blue network

47:18

probably going to have to do is

47:20

come up with a minimalist standard space,

47:22

you know, standard system. And not quite

47:24

sure how they're going to pull off.

47:27

Right now they're a disarray. I mean,

47:29

popularity that the Democratic Party is way

47:31

below the Republican's aisle, lowest has ever

47:34

been in history. There's no clear leadership

47:36

because they don't really work on a

47:38

leadership basis. Their moral frameworks are disarray

47:41

they're not on, and they've left, just

47:43

like a lot of the work I

47:45

had to leave to parlor and stuff

47:48

they're off on a stump network or

47:50

collection of networks, and they're not having

47:52

the influence that they should have, or

47:54

they need to have, actually, in order

47:57

to kind of, revive themselves. They have

47:59

to go on X and X is

48:01

the biggest one and play that kind

48:04

of wild environment. and build themselves up.

48:06

So I still haven't seen what they're

48:08

going to do to kind of create

48:11

that kind of new approach to kind

48:13

of revive themselves. Just like, you know,

48:15

it wasn't really apparent in 2020 what

48:18

the red network would do to revive

48:20

itself. And it emerged naturally. All of

48:22

a sudden, the robot's like huge and

48:24

Musk is huge and all those other

48:27

multi-million person accounts and 10 million, 20

48:29

million tuckers and all these others get

48:31

massive. It wasn't apparent in 2020 with

48:34

that response. I do think we're going

48:36

to get it. It's just it's coming.

48:38

These networks evolve. They improve that decision-making

48:41

system that they're using to kind of

48:43

improve their capabilities, tribalize better, create more

48:45

cohesion, is on its way, but I

48:48

haven't seen a lot of evidence as

48:50

to what it will be yet. And

48:52

I will say the... Trump administration has

48:55

been making the job more difficult by

48:57

operating at an amazing tempo once they

48:59

got into office. I mean I'm an

49:01

amateur military scholar you're a more serious

49:04

one. You know we know that in

49:06

war tempo is hugely important and I've

49:08

never seen tempo exercised in the political

49:11

realm like the current administration is doing.

49:13

Yeah, they're doing a lot of delaying

49:15

tactics for anything with the legal or

49:18

the law fair and other things like

49:20

that to try to slow it down.

49:22

I'm delay it. I mean, of course,

49:25

some things have to be done. I

49:27

mean, running a $2 trillion a year

49:29

deficit is not sustainable. Yeah, so we're

49:31

going to end up in bankruptcy before

49:34

you know it at that kind of

49:36

clip. So there has to be stuff

49:38

cut. So yeah, now what I, you

49:41

know, That tempo, of course, maneuver warfare,

49:43

you constantly are shifting topics, you're constantly,

49:45

it's very decentralized now because they have

49:48

all these different agency heads that have

49:50

their own kind of true-seeking agendas, and

49:52

they're going off and changing things. The

49:55

only thing really stopping them or slowing

49:57

them at this point is the legal

49:59

and regulatory structures that bind this, what

50:02

I call a hollow state, together. Because

50:04

the nation state dies. 20 years ago

50:06

when we started to go globalization and

50:08

globalize and focus outward and it's turned

50:11

kind of into a hollow state where

50:13

there's all these it's really a mechanism

50:15

for looting what's left of the nation-state

50:18

and they built a lot of defenses

50:20

a lot of influence networks a lot

50:22

of legal regulations and you know if

50:25

you've ever done any government contracting you

50:27

know how absolutely corrupt it is. I'll

50:29

pay If I pay $200,000, I mean

50:32

$200,000, I can get disinserted to a

50:34

bill and it will specifically be specific

50:36

enough that it will address me in

50:38

particular. You know, my company, I'll make

50:41

two or three million or ten million

50:43

or twenty million because I can sell,

50:45

I'm the only one selling this. It's

50:48

awful. I certainly hope that. Elon digs

50:50

into the procurement process, particularly the defense

50:52

procurement process, which is like the rottenest

50:55

pile of sludge. ever. I had a

50:57

narrow, intentionally my business career avoided government

50:59

contract. I remember once one of our

51:02

companies, the SEC wanted to use one

51:04

of our online financial products, and would

51:06

we go on the GSA register? We

51:08

said no. You can use it with

51:11

your credit card, because I'll be god

51:13

damned, I'm going to get immersed in

51:15

their horseshit. But I did have some

51:18

encounters later, and I'll tell you. Very

51:20

unpleasant very bad. They could probably find

51:22

their trillion dollars in defense procurement alone

51:25

If they dug deep enough, you know

51:27

over a ten year over ten years.

51:29

Oh, yeah, and you have these all

51:32

these fraudsters Yeah, my brother just retired.

51:34

He was like a he ran all

51:36

of the NSA's hardware Real a little

51:39

bit of it. He had all these

51:41

different ways to actually improve it. You

51:43

know, how you approved the contracting and

51:45

a lot of had to do with

51:48

holding people accountable for results and it

51:50

a lot of times it, you know

51:52

It dies, you know, if you try

51:55

to hold somebody accountable, it's just a

51:57

corporate shell that gets punished, but not

51:59

the people who are... you know, actually

52:02

running it. And they, as soon as

52:04

they fail, they transfer over to another

52:06

shell, right? And then they keep on

52:09

going. It's a serial fraud. It's a

52:11

serial kind of disaster. And even the

52:13

big guys will do the, you know,

52:15

billion dollar, you know, the IRS renovation

52:18

project, the big SAIC, information spy architecture

52:20

that rode off because the big, even

52:22

big boys, you know, fuck this shit

52:25

up. And of course, partially it's the

52:27

government's own fault, because what they really

52:29

should be doing is like, you know,

52:32

think about SpaceX versus... the defense contractor

52:34

method. Space X just said, what's the

52:36

problem, getting weight into orbit safely and

52:39

economically, rather than some fucking 2,000-page specification

52:41

that NASA writes and puts out the

52:43

bid between Boeing and Lockheed, right? And

52:45

when you try to tell people how

52:48

to do things, particularly if you're relatively,

52:50

you know... lower level bureaucrat, you're very

52:52

likely to grossly over specify, make wrong

52:55

assumptions about the best way to do

52:57

it. You know, if I were the

52:59

dictator, it's probably not a bad idea,

53:02

I would put, you know, defense stuff,

53:04

okay, I want air supremacy, I don't

53:06

give a fuck how we have air

53:09

supremacy, I want air supremacy, and we're

53:11

willing to pay blah for air supremacy,

53:13

and let people figure out how to

53:16

do air supremacy, what's the mixture of

53:18

man versus drone? Maybe it's all drone,

53:20

maybe it's a mix, maybe it's very

53:22

close satellite drone plus man, but let

53:25

there be some mission-solving orientation to these

53:27

bids rather than specification-based bids. Yeah, first

53:29

principle is kind of analysis. It's not

53:32

done nearly enough. People don't go back

53:34

to the basics and try to look

53:36

at what, you know, we're way downstream

53:39

of what that was, you know, that

53:41

first principle is kind of a development

53:43

cycle, and we're like... We have so

53:46

many baked in assumptions to what we

53:48

do now, and they're hidden in programs

53:50

and layers of thinking, and a lot

53:52

of them. run, right? I mean, you

53:55

know, stay in Afghanistan because we have

53:57

to be there, you know, they should

53:59

building two and a half trillion dollars

54:02

later in a failure in an evacuation.

54:04

I mean, these things just persist. No

54:06

one goes back to kind of dig

54:09

those suckers, you know, dig those vault

54:11

assumptions out, those rotten ones that are

54:13

locking us into things. You try to

54:16

talk about it about it with people.

54:18

They'll go, you can't even address it

54:20

because, you know, you try to get

54:23

into a serious form, they go, You

54:25

know, we can't talk about that. That's

54:27

too complex. That's never going to happen.

54:29

It's never going to, you know, it's

54:32

not even worthy of consideration because it's

54:34

so outside the realm of possibility that

54:36

that could be changed and they avoid

54:39

it. Rather than, you know, they'll want

54:41

to talk about the edges and, you

54:43

know, little tweaks and things like that.

54:46

And it's like, you can't fix it

54:48

until we do this. We'll end up

54:50

being a net positive for our countries.

54:53

There'll no doubt break some things, but

54:55

there is so much rot there that

54:57

needs to be shoveled out that it's

54:59

worth the occasional break. I mean, there's

55:02

a reason why the counties around the

55:04

beltway are the wealthiest counties in the

55:06

United States. I mean, those folks... and

55:09

their kids. Yeah, I pointed that out

55:11

more than one. I grew up in

55:13

the DC Burbs and they were not

55:16

the richest counties in the 50s and

55:18

60s, right? One of them was Montgomery

55:20

County, but none of the rest. But

55:23

by today, most of all the richest

55:25

counties in United States around DC's, feeding

55:27

at the trough. Yeah, I mean, I

55:29

can understand like I live in Middlesex

55:32

and we have all the, you know.

55:34

software and finance there and all that

55:36

other stuff and I can understand the

55:39

wealth that was created but the DC

55:41

area it doesn't you know you have

55:43

and then you start seeing these like

55:46

USAID and other things you see these

55:48

people you know like the head of

55:50

USAID is government employee and goes from

55:53

a million dollars that worth the 20

55:55

million dollar that worth in three years

55:57

of actually running it how do you

56:00

How is that even possible? And you

56:02

see these again and again and again

56:04

and again and the kids of the

56:06

connected inside there are making multi-hundred thousand

56:09

multi-million dollar deals doing basically nothing or

56:11

advancing an agenda that is contrary to

56:13

the interests of the United States. I

56:16

mean I think we did this on

56:18

the last podcast of what went wrong

56:20

with America is that we have to

56:23

become slightly more nationalistic, I mean more

56:25

Washingtonian. and lead more by example, you

56:27

know, just trying to improve the prosperity

56:30

of Americans, coming up with ways to

56:32

use all this new technology that we're

56:34

developing for the most part in a

56:36

way that's positive and, you know, something

56:39

everybody else wants to emulate, just like

56:41

we do with electricity, just like we

56:43

did with indoor plumbing, just like we

56:46

did with all these other things that

56:48

we pushed out into the general society

56:50

and everyone else copied us. So I

56:53

want to see that. You know, appliances,

56:55

everything. All. all these things I want

56:57

to see this. So I don't know

57:00

how we get there from here with

57:02

these kind of network organizations, but I

57:04

hope we can find a way to

57:06

turn network decision-making into something that can

57:09

solve the kind of breakdown in hierarchical

57:11

and market-based decision-making that got us to

57:13

this point. It's kind of dysfunction, this

57:16

kind of holliness. Low cohesion, low trust,

57:18

dysfunction, corruption, inability to solve complex problems,

57:20

and more. So as chaotic it is

57:23

on the network front, it's... Something we

57:25

have to go through it in order

57:27

to solve these bigger problems that we

57:30

have with the way we have been

57:32

making decisions in the past. A hopeful

57:34

future with some question marks around it.

57:37

Want to thank John Rob for yet

57:39

another very interesting conversation here on the

57:41

Jim Rutcho and or check out the

57:43

episode page at Jim Rutcho.com to links

57:46

to many of the things we talked

57:48

about. All right, thanks Jim. Audio

57:53

production and editing by Andrew

57:55

Blevins Productions Music by Tom

57:58

Mueller at Modern Space music.com

Rate

Join Podchaser to...

  • Rate podcasts and episodes
  • Follow podcasts and creators
  • Create podcast and episode lists
  • & much more

Episode Tags

Do you host or manage this podcast?
Claim and edit this page to your liking.
,

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features