Can AI really write? A no-nonsense discussion, with Christopher Penn

Can AI really write? A no-nonsense discussion, with Christopher Penn

Released Thursday, 10th October 2024
 1 person rated this episode
Can AI really write? A no-nonsense discussion, with Christopher Penn

Can AI really write? A no-nonsense discussion, with Christopher Penn

Can AI really write? A no-nonsense discussion, with Christopher Penn

Can AI really write? A no-nonsense discussion, with Christopher Penn

Thursday, 10th October 2024
 1 person rated this episode
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:01

Reuniti. Here's to the world's favorite Lambrusco

0:03

and one of the most beloved Italian

0:05

wines worldwide. Here's to its chill, delightful

0:08

fizz that springs to life as thousands

0:10

of tiny bubbles dance in your mouth.

0:12

Here's to the refreshing taste of Reuniti.

0:15

After a long hard day it's the perfect

0:17

pour you deserve with a company you enjoy.

0:19

Reuniti Lambrusco. The refreshing, lively and easy to

0:21

drink Italian wine. Reuniti on Ice Still Nice

0:23

imported by Frederick Waldman and Sons Limited, New

0:25

York and New York campaign finance according to

0:28

EU Rec 2021-2115. It's

0:32

time to get the world talking about black

0:34

lead brands. We all have our favorites but

0:36

we can't keep them all to ourselves. So

0:38

if you're feeling a little black opal beauty,

0:40

tell somebody. If the lip bar gives you

0:42

a lip for every drip, let them know.

0:44

And if your hair is doing the

0:46

do, shout that out too. Join

0:48

Walmart in shouting out your favorite

0:50

black lead products creating a new

0:53

world of choices at walmart.com/black and

0:55

unlimited. Grammar

1:03

Girl here. I'm Mignon Fogarty and today

1:05

we're going to talk with Christopher Penn

1:07

about some of the common misconceptions around

1:09

AI tools like chat GPT that I've

1:12

seen among writers and editors and just

1:14

people in general. Whether you

1:16

love it or despise it, I

1:19

think it's really important to understand

1:21

AI and Chris is

1:23

a guy who can really help us do that. Christopher

1:25

Penn, welcome to the Grammar Girl podcast. Thank

1:27

you so much. It is nice to

1:30

be on one of the OG podcasts.

1:33

You and I have been podcasting I think since what,

1:35

05? You were 05, I think I was 06. But

1:37

yeah, it's been a while.

1:39

Yeah, and

1:42

so actually, you know, because I've been

1:44

subscribed to your newsletter for so long,

1:47

one of the things I've noticed about you

1:49

is that you are always on the cutting

1:51

edge of trends and data. I mean, you

1:53

called the housing crisis in 2007. I remember,

1:57

you know, I was looking at buying a house and you. actually

2:00

said to me, Mignon, I think maybe you should

2:02

wait. And I

2:04

did not take your advice as one

2:06

of the worst financial situations I've ever

2:08

made. But we loved the house. It

2:11

turned out okay. But you were definitely

2:13

right. And then in

2:15

2019, 2020, again, you

2:17

predicted how bad the pandemic was going to

2:19

be, but way before a lot of other

2:21

people realized it. And

2:23

so about, I guess it was about a year,

2:25

a year and a half ago, I noticed your

2:29

almost timely newsletter became

2:31

entirely about AI. And

2:34

I would love to hear more about

2:36

how that came about, like whether it

2:38

was a gradual thing or did you

2:40

just have an aha moment where you

2:42

were like, okay, this needs to be

2:44

everything. So AI has existed in

2:46

some form since the 1950s. I

2:49

started to have an interest in in 2013. And

2:53

there's three branches of AI. There's what's

2:55

called regressive classification and generation. One

2:58

is needle in a haystack. Hey, here's a bunch of data.

3:00

And here's an outcome. What in the data

3:03

is like this outcome, right? We call this

3:05

finding AI. There's classification,

3:07

here's a bunch of data, let's organize it. Like

3:09

what do we got in here? You

3:11

see this with particularly people like marketers trying to

3:14

organize things like social media posts, like there's so

3:16

many, how do we organize and classify them. So

3:18

that's the central mind. And then

3:20

there's this new branch started in 2017, called

3:23

generative, based on an architecture and

3:25

a paper created by Google called

3:28

attention is all you need. And the architecture

3:30

is called transformers. That

3:32

became a product family

3:35

really around 2020. Just

3:38

as the pandemic got rolling, when companies

3:40

started to say, maybe we can get

3:42

this architecture do something interesting, like predict

3:44

the next word in a sequence in

3:46

a way that's never been done before.

3:48

By 2021, they were usable models that

3:50

would write coherent text, not actually correct,

3:53

but at least coherent is readable. Prior

3:55

to that, it looked like you're rolling

3:57

your face on a keyword. when

4:00

I started paying attention to this technology,

4:02

I started using OpenAI's GPT-3, which had

4:05

come out 2021. And then November 2022, OpenAI

4:07

releases a tool called Chat GPT. And suddenly,

4:13

everyone's an AI expert. And

4:16

having that background of more than

4:18

a decade working in the core

4:21

technologies, I could see where this is

4:23

going because Chat GPT changed people's relationship

4:26

with AI. Prior to that, you

4:28

needed coding skill, right? You needed the

4:30

ability to work in these models

4:32

and stuff like that and build your own models.

4:34

And then suddenly, there's a web interface you can

4:37

chat with just like you were drunk texting with

4:39

your Uncle Fred. And suddenly, okay,

4:41

wow, people can make use of this. And

4:43

it was that point where I started pivoting

4:45

a lot of my content to say, okay, here's

4:48

what the stuff does. And here's

4:50

how we should be thinking about it.

4:52

And of course, the rest of the world pivoted as well,

4:55

as people start to understand the implications of these

4:57

tools. But now today, these

4:59

tools have accelerated faster than

5:01

I've ever seen any technology

5:03

evolve. It is absolutely

5:06

breathtaking how quickly they've evolved.

5:08

And I'll give you an example. In

5:11

the first release of Chat GPT,

5:13

that was the backend model is

5:15

this obscure, poorly named bottle called

5:17

GPT 3.5 turbo that

5:19

required an entire server room of

5:22

machines to run and serve up

5:24

and send your text. About

5:27

a month ago, Meta, the parent company of

5:29

Facebook, released a model called Llama 3.1. That

5:33

version, there's a version you can run that

5:35

is more capable,

5:38

more capable than GPT 3.5, about as capable

5:41

as the successor GPT 4, and

5:43

runs on your laptop with no

5:45

internet connection. Hey,

5:51

I'm just jumping in here to say that

5:53

we recorded this episode about a month ago.

5:55

And since then, Meta has released an even

5:57

newer model called Llama 3.2. So,

6:03

we've gone from an immature

6:05

technology that is extremely expensive to

6:07

use to a mature technology that

6:10

you can run, I've used

6:12

it on a plane with no internet

6:14

and to have generative AI capabilities all

6:17

the time in less than two

6:19

years. That is crazy fast evolution.

6:22

Yeah, I agree. It's amazing. I

6:24

haven't seen anything like it since the introduction of the

6:26

internet. Yeah, and if people looked into

6:28

it six months ago or eight months ago and

6:30

gave it a try, it's

6:33

different today. It's so much

6:35

better. There are so many misconceptions. One

6:39

of the things someone asked me a

6:41

few weeks ago was, well, does chat

6:43

GPT use AP style or Chicago style?

6:47

I remember having that question when I first started

6:49

playing around with it too. It's

6:51

a reasonable thing for an editor or writer to ask,

6:53

but it's not how it works.

6:57

Can you maybe start by explaining why that's

6:59

not the right question to ask and how

7:01

it works at a basic

7:03

level for people who haven't looked into

7:05

it as deeply as you have or

7:07

maybe I have? Sure. The

7:10

way these tools work is

7:12

they understand pieces of words

7:14

and how those pieces of words relate to

7:16

every other word around them in the sentence,

7:19

in the paragraph, in the document itself. They

7:22

are trained on such huge quantities of text.

7:24

To give you an idea of what the

7:26

average model is trained on, it would be

7:29

a bookshelf that wraps around the equator twice.

7:32

That's how much raw text is needed

7:34

to train one of these models. What

7:36

they get used to seeing is understanding

7:38

the statistical distribution of language. Implicit

7:41

in that is how language

7:44

works. If I

7:46

say I pledge allegiance to the,

7:48

the next word probably is flag.

7:51

It probably is not rutabaga. As

7:55

a result, the models understand the

7:57

likelihood of everything that is in the

7:59

sentence. we say. If I say

8:01

God save the depending on your

8:03

nationality and your knowledge

8:05

base, you might say that the queen or king,

8:08

but again, probably not Rudabaga. When

8:10

you ask about something like Chicago style

8:12

or AP style or any writing style,

8:14

you're asking about a specific format. And

8:17

these models have seen all that and then some

8:19

what they spit out, by definition

8:21

is probability based text. So

8:24

if you give it a

8:26

very short, naive, terrible prompt,

8:28

like write me a blog

8:31

post about B2B marketing, it's

8:33

going to give you the most probable

8:36

words, phrases, concepts to that in

8:38

a written format that is

8:40

probably going to be technically correct, probably

8:43

going to be boring, probably going to

8:45

be completely unoriginal, because it

8:48

is invoking what

8:50

its best guesses are probabilities for those words.

8:52

If you think of when we talk to

8:54

these tools, we are prompting them, we're writing

8:56

to them like chatting, etc. Every

8:59

word that we type in our prompt

9:01

kind of has an invisible word cloud

9:03

around it. And where

9:06

those word clouds intersect is

9:09

how it knows what to respond with, right?

9:11

So these word clouds are huge, and they

9:13

they they narrow down. So if you write

9:16

a prompt like write me a short fiction

9:18

story about two women in love,

9:22

you're going to get direct, right?

9:24

Because it's so generic. But if you

9:26

say write me a short story about

9:28

two women in love set during the

9:31

Victorian period, but set in Wales instead

9:33

of London, and the one

9:35

woman comes from originally from Scotland, the

9:37

other woman comes from France, originally, you

9:40

see all these extra words and all

9:42

these word clouds start overlapping, and they

9:44

get more and more precise, and

9:46

you get more and more text that is

9:48

unique and different, because you're getting more specific.

9:51

The one thing I say in all the

9:53

keynote talks that I do is if you

9:55

remember nothing else, the more

9:57

relevant specific words you use we

10:00

prompt these tools the better the output will be

10:02

if you give it i say

10:04

think of these things like the world's

10:07

smartest interns brand new intern comes in

10:09

the office this internet 255 phds got

10:12

a phd and everything but it's still day

10:14

one they still don't know who you are they

10:17

don't know where the restroom is they don't know the coffee machine

10:19

is and so you gotta teach you

10:21

to say to enter hey intern go write me a

10:24

book. You're gonna get crap out

10:26

of human intern just like you're gonna get crap

10:28

out of a machine if you on the other

10:30

hand give the engine here's our style guide your

10:32

brand safety guidelines here's the background information here's the

10:34

outline i want you to use i

10:37

want you to use a style. Can

10:39

you give the internet a big ol' pile of directions

10:41

and that is going to come up with a much

10:43

better result. Right yeah i

10:45

feel like we shouldn't go past the

10:48

idea of training unlike the entire world's

10:50

libraries without mentioning the concerns about copyright

10:52

infringement in the lawsuits that are currently

10:54

happening so that's like a whole other

10:56

podcast that's a big issue in the

10:58

field so if you weren't aware of

11:01

that i want you to know we're

11:03

not gonna focus on it very much

11:05

today. We're not

11:07

going to focus on it but there's

11:09

an interesting distinction and this is what

11:11

the law will hinge on and it

11:14

depends on where you are so in

11:16

the EU the EU has ruled that

11:18

the use of someone else's intellectual property

11:20

for creating models in fringes on their

11:22

their rights. In Asia

11:25

China and Japan have rolled that what

11:27

is inside of all of the pull

11:29

it open is a big pile of

11:31

numbers in no way resembles the original

11:33

works and therefore. Model

11:35

does not infringe on someone's rights because

11:38

it can no way could be mistaken

11:40

for the original and so

11:42

the question is gonna

11:44

be resolved in every jurisdiction separately

11:46

as to whether a what

11:49

i does infringing on the originals or

11:51

not. And it's not clear how that's

11:53

gonna play out. Oh thank you that's

11:55

fascinating i wasn't aware that there were

11:57

country differences at this point that's super

11:59

interesting. Does that mean people in

12:01

the EU can't use these models right now? There

12:03

are some tools they are not permitted to use

12:05

that are not available in the EU. Wow.

12:07

Okay. Okay. I didn't know that. So

12:11

another concern that I see... Oh, and I

12:13

just want to ask, so would it be

12:15

fair to say that, you know, back to

12:17

the AP Chicago issue and thinking about the

12:19

way it works, would it be fair to

12:21

say that if you asked it to write

12:23

a news article for you, that it would

12:25

be very likely to follow AP style because

12:27

it is looking for words

12:29

from that style of writing that originally

12:31

would have been written in that style

12:33

or is that not? It's probable that

12:35

it will be invoked something that looks

12:37

more like AP style than Chicago style,

12:39

as opposed to say writing a nonfiction

12:42

book, which is probably going to be

12:44

more Chicago style than AP style. Again,

12:47

with these tools, if you tell it, I want this

12:49

in AP style, it will write it in AP style.

12:51

If you let it do its own devices, it will

12:53

do the highest probability for the data set that you're

12:55

working with. One of the things

12:57

people knock these tools all about is saying, oh,

12:59

they use the exact same words. In a world

13:02

of whatever, and then, you know, the word delve

13:04

and all this stuff. Well, that's within that context.

13:06

If you say, write some sapphic

13:09

fiction, the word delve is not

13:11

going to appear because in the training data it's

13:13

seen, sapphic fiction

13:15

writers don't use that word, right? So

13:17

it's not going to be within that

13:19

set. So part of prompting is understanding

13:21

what words we want to invoke and

13:24

what range of addiction

13:26

we want to invoke that is

13:29

very often not something that human

13:31

writers think about. Great.

13:34

So another misconception that I hear, I

13:36

hear a lot of people who are

13:39

concerned about privacy and confidentiality, that they

13:41

don't want to upload their information because

13:43

they don't want it to be used

13:45

to train data or they have clients

13:47

that don't want their proprietary information

13:50

getting into these models. And I think that

13:52

sometimes it's the case that you wouldn't want

13:55

to do that, but I don't think it's

13:57

always the case. And how can people carefully

14:00

can deal with these issues and still carefully

14:02

use AI if they want to.

14:04

Here is the golden rule. If you ain't paying,

14:06

you're the product, right? It's been that way for

14:08

social media. It's been that way for everything. And

14:10

the same is true with AI. If you ain't

14:13

paying, your data is the product and that is

14:15

being used to train other people's models and things.

14:18

Now, there's two different branches of

14:20

thought on this. And here's

14:22

one that's, you know, the obvious one is pay

14:25

for tools, and then make sure that the privacy

14:27

settings and those tools reflects what you expect. If

14:29

you're using, for example, the enterprise edition of any

14:31

tool in the contract, it says we will not

14:33

use your data to train on and stuff. And

14:35

so if you are working with sensitive information, you

14:38

absolutely should be doing that. If there is some

14:40

data, you absolutely positively just cannot give to a

14:42

third party. And there are certain versions

14:44

of AI, like metas, llama, 3.1, four or five

14:47

the model, you can run that

14:49

on your own servers inside your own

14:52

company. And that is totally under your

14:54

IT department's control. It is reassuringly expensive

14:56

to do that. You spend tens of thousands

14:59

of dollars on hardware. But if you are

15:01

say a three letter agency based in Langley,

15:03

Virginia, then that's a

15:05

pretty safe investment, you will have all

15:08

the benefits of generative AI, but you

15:10

will know your data will never ever

15:12

leave the protective compounds of your compound.

15:15

So that's one aspect. The second

15:17

aspect, when it comes to what

15:19

goes on with your data

15:21

is knowing the level of

15:23

sensitivity of it, right? So if

15:26

your company or your work, whatever is

15:28

bound by guidelines, like ISO 27001, or

15:30

SOC 2 compliance stuff, you

15:34

know what you're supposed to be doing with your data

15:36

of any kind, you know what's allowed and not allowed.

15:38

So you can just look at here's

15:40

the here's the requirement in general that my system's

15:42

first have like HIPAA compliance for healthcare. Look

15:45

at chat GPT, not HIPAA but okay,

15:47

clearly then I can't use

15:49

HIPAA data in a non HIPAA system.

15:51

I mean, that's, that's kind of a

15:53

no brainer. So think about the regulations

15:55

you already have to adhere to and

15:57

how do and what systems are qualified.

16:00

for those regulations. Right. Well, what if you're just

16:02

a lowly person like me who maybe has a

16:04

$20 a month chat GPT or

16:06

Claude subscription or something like that? Do you

16:08

feel like that is, and I think they

16:10

say that they won't use your data for

16:12

training, but also it's pretty cheap. So am

16:15

I putting my stuff out there at risk?

16:17

If you are putting, if you're

16:20

setting, if you're confirmed to the setting say

16:22

they will not train on your data, then

16:26

the basis of contract law is that they will not train

16:28

any data. If it turns out that they do that, then

16:30

you get to join the class action lawsuit and sue them

16:32

for a whole bunch of money if

16:34

it comes out that they were not adhering to those

16:36

terms. And again, this comes down to sensitivity. So if

16:38

you are writing a book and it

16:40

turns out that they in fact were training on your data

16:42

but not, yeah, you should

16:45

join lawsuit. If you are processing

16:47

case notes from a field agent and that

16:49

leaks and that gets that agent killed, you

16:52

should not be using chat GPT period.

16:55

No license, no matter that you should be running

16:57

that internally, you know, in a protected

16:59

system because that is literally life or death. And

17:02

I would say the same is true for all of these

17:05

different use cases is how, what

17:08

are the consequences of something goes wrong? That's a

17:10

question people don't ask often enough

17:12

is what could go wrong asking it

17:14

unironically? Yeah.

17:18

So let's talk a little bit

17:20

about hallucinations and fact checking. So,

17:23

you know, one of the things

17:25

I see the most people doing

17:27

actually is using chat GPT, Claude,

17:29

perplexity as a search engine, as

17:31

a replacement for Google essentially. And

17:34

yet we know that they make

17:36

mistakes. So how do you approach

17:38

thinking about using these tools for

17:40

search and what advice do you

17:42

have for people who are doing

17:44

that? So a hallucination is

17:47

a probability error. If

17:49

you think of these tools like

17:52

a library, like a model is like a library

17:54

and there's a library and you ask the library

17:56

and questions, depending on the obscurity.

17:58

of the question, the librarian may or may

18:00

not be able to find a book that's

18:02

appropriate to your question. So if you walked

18:04

into a library and it was pretend it's

18:06

obscure, you walk and say, hey, I'd like

18:08

the joy of cooking. And the

18:10

librarian wanders around a bit and comes back with the joy of sex.

18:13

And he's like, this is close, right? Like,

18:16

no, that's not close. Not at all. But

18:19

semantically, it thinks it's close

18:21

enough. That's what's happening in

18:23

a hallucination is the models

18:26

have three basic directives, be

18:28

helpful, be harmless, be truthful.

18:31

And it's a delicate balancing act, but they

18:33

try to be as helpful as possible at

18:36

sometimes at the exclusion of being truthful,

18:38

right? So they will say, hey, you want this thing? I'm going

18:40

to try my harsh to give you this thing. And it gives

18:43

you this thing is that that's the thing you're like, yeah, but

18:45

that's not right. It

18:47

was right. And

18:50

so using these tools as

18:52

search engines is probably not the best use

18:54

case for a language model self. If

18:57

you dig into the raw technology stuff, you

18:59

got to get super nerdy. All

19:03

models hallucinate 100% of the time when they're

19:05

first built. It requires a

19:07

lot of tuning to even just stop get

19:09

them from lying and period, because they're always

19:11

assembling just probabilities. So there's a lot of

19:13

extra steps to go into refining a model

19:16

to make it more truthful. And

19:18

then there's additional safeguards built on top

19:20

of it. Now what's happening in the

19:22

field is that we're diverging into people,

19:24

these model makers recognize the validity of

19:26

the use case of most

19:29

internet experiences suck, right? Someone's website like I

19:31

want to go check out a recipe. I

19:33

got a way through 42 pages of your

19:35

grandmother's second cousins, roommates, dogs, you know, why

19:38

they like this recipe. I don't care. Tell

19:40

me how many cups of sugar to use.

19:43

And so perplexity and Google and chat GPT

19:45

and all these companies said, you know what?

19:48

That experience sucks. So instead, we're

19:50

going to generate the answer for you that

19:52

says you need two cups of sugar for

19:54

this recipe. The consumer sees

19:56

that as a much better experience. And

19:58

so we now have tools like complexity,

20:01

we have Google's AI overviews, we

20:03

are going to shortly have

20:05

open AI search GPT that

20:07

use search data as the

20:09

basis of their responses. And

20:11

then they craft responses from that data,

20:14

and they still do get it wrong.

20:16

Perplexity, in particular, will ingest

20:18

the web page content and

20:21

summarize it down. So it's more accurate than

20:23

just making it up from the model's long

20:26

term knowledge, but it's still sometimes wrong. Google

20:28

famously, when search standard experiments became AI overviews,

20:30

had the case of well, hey, how do

20:33

you make pizza, you know, cheese, you know,

20:35

stay on pizza, add a quarter cup of

20:37

glue to the pizza. That

20:39

is, that was from a Reddit forum post,

20:42

that someone said that ingest and Google just

20:44

does not have a sense of humor. And

20:47

therefore, I was unable to detect that. So for

20:49

consumers, there are some tools

20:52

which are going to likely lead to a higher

20:54

probability of success, like search GPT,

20:57

like perplexity, like Google AI overviews,

20:59

because they are rooted in search

21:01

results. Other tools like

21:03

chat GPT, or straight you know, regular

21:05

consumer Gemini, or regular Claude, you should

21:07

not use those as search engines, because

21:09

they're not pulling from at least

21:12

known sources on the web. However,

21:15

they are still subject to the

21:17

vagaries and the flaws of search

21:19

itself, right, you can find in

21:22

a search engine, absolute garbage, completely

21:24

incorrect information. And whether using

21:26

generative AI or traditional search, you will

21:28

still get the answer to what you were

21:30

looking for, even the answer is completely wrong.

21:33

Right. So if especially if you're searching for

21:35

something important, it's critical that you actually check

21:37

somewhere else that the facts so if they

21:39

say you're using perplexity, it shows

21:42

you the sources where you got the information,

21:44

click through on those sources and

21:46

double check what you've got is right and

21:48

not somehow taken out of context on that

21:50

page or or something like that. Yeah, and

21:53

and know the difference between sources. For example,

21:55

we had to we had to postpone this

21:57

interview, this interview was supposed to happen last

21:59

week. I was in the hospital getting

22:01

emergency surgery. As

22:04

part of recovery, I wanted to figure out what

22:06

I could do to accelerate recovery. So I went

22:08

to perplexity. I said, you know, find me peer-reviewed

22:10

research on post-surgical recovery

22:13

methods. And it points

22:15

out these different papers. I went and downloaded

22:17

the papers themselves after checking which journals they

22:19

were in. Nature, Cell, Science,

22:22

NIH, PubMed. I know these to

22:24

be credible journals, as opposed to

22:26

Bob's Random Journal of whatever.

22:30

And then stuck it into a system like Notebook

22:32

LM, which is a product by Google specifically made

22:34

for academic research, where you give it the data

22:37

and ask questions of the data. And

22:39

if you didn't provide the answer, it will say, I don't

22:41

have the answer. It will not try to make it up.

22:44

And I was able to design a post-surgical

22:46

recovery regimen that's working about 40% faster than

22:49

the normal because I was

22:51

able to go get the data. But I had to

22:53

be the informed consumer to do that. The tools don't

22:55

do it for you. Yeah, that's great.

22:57

I'm glad you're getting better faster. And yeah,

22:59

I do the same thing. I go to

23:01

Perplexity when I have medical questions like that.

23:03

But I always download the journal article and

23:05

read it. I used to be a scientist

23:07

and a tech writer, so I'm comfortable reading

23:09

journals. And 90% of the time, it's great.

23:13

But I have found errors occasionally,

23:16

things that Perplexity didn't quite present

23:18

properly in the summary. Reuniti.

23:26

Here's to the world's favorite Lambrusco and

23:28

one of the most beloved Italian wines

23:30

worldwide. Here's to its chilled, delightful

23:32

fizz that springs to life as thousands of

23:34

tiny bubbles dance in your mouth. Here's

23:37

to the refreshing taste of Reuniti. After

23:40

a long, hard day, it's the perfect pour you deserve

23:42

with the company you enjoy. Reuniti Lambrusco,

23:44

the refreshing, lively, and easy to drink Italian

23:46

wine. Reuniti on Ice, still nice. Imported by

23:48

Frederick Waldman and Sons Limited, New York, New

23:50

York. Campaign finance according to EU Rec 2021-2115.

24:00

family, durable and ready to carry

24:02

all your stock and you quit

24:04

for adventure. Honda's here with the

24:06

2025 Honda CR-V. CR-V outperforms the

24:08

pack with more style, more space

24:11

and more power. Complete with state

24:13

of the art safety features and

24:15

a five star safety rating. Honda's

24:17

here for you with the 2025

24:21

Honda CR-V. Check it out at

24:23

your local Honda dealer today. I've

24:27

been working on learning Spanish for

24:29

a long time and the thing

24:31

I struggle with most is speaking

24:33

and pronunciation. It feels awkward

24:35

and I find it hard to even tell

24:38

if I'm getting it right. But

24:40

Rosetta Stone is really helpful with

24:42

pronunciation. It has lessons where

24:44

you actually sound at words syllable

24:46

by syllable and it generates what's

24:49

almost like a score bar so

24:51

you can quickly see how close

24:53

you're getting. It's instant feedback and

24:56

it's not surprising that it's great because

24:58

Rosetta Stone has been an expert in

25:00

language learning for 30 years

25:02

with millions of users and

25:04

it's all about fast language

25:06

acquisition. Better yet the

25:08

program is designed for long-term retention and

25:11

you can learn on your desktop or on

25:13

the go with the Rosetta Stone mobile app.

25:16

Check out the lifetime membership

25:18

option which gives you access

25:20

to 25 languages including French,

25:22

Korean and Dutch. Don't

25:24

put off learning that language. There is no

25:26

better time than right now to get started.

25:29

Our listeners can get Rosetta Stone's lifetime

25:31

membership for 50% off.

25:34

Visit rosettastone.com/grammar.

25:37

That's 50% off unlimited access

25:39

to 25 language courses for the

25:41

rest of your life. Redeem

25:43

your 50% off

25:45

today at rosettastone.com/grammar.

25:53

So you know some of the people I encounter

25:56

who are most opposed to

25:58

generative AI are... are fiction writers, which

26:01

I think is kind of funny because not

26:03

that I'm laughing at their concern, but that fiction

26:06

writing is one of the things I think that

26:08

AI is really bad at. What

26:11

do you find are the things that

26:13

for writers, AI is like the best

26:15

for? So there's

26:17

six broad use cases of generative

26:19

AI. Generation summarization extraction rewriting classification

26:21

question answering, those are the big

26:23

six. Of those six,

26:26

two of them it's less good

26:28

at. Generation, AKA writing and

26:31

question answering because of the factuality issues.

26:33

Summarization, take big data and make it

26:35

smaller data, they are almost flawless at,

26:37

they are so good at it. Extraction,

26:39

take data out of data. Super

26:42

powerful, super helpful when, you know, we're in

26:44

the midst of a presidential election here in

26:46

the USA and this one group

26:48

released this one big document and I said,

26:51

extract out just these parts because I care

26:53

about this and I wanted to read just

26:55

that part of it and it did a

26:57

fantastic job. Rewriting, take one form of data,

27:00

turn it to another form of data. So

27:02

for example, one of the best things tools

27:04

do incredibly well is to take complex things

27:06

and explain them into something you understand. So

27:08

if you are a fiction writer, you can

27:11

have it explain something in a story format

27:13

that's maybe an arcane concept like how the

27:15

spike protein on the SARS-CoV-2 virus works. You

27:17

know, explain this to me as though it

27:20

were a fantasy epic and

27:22

it will do an incredible job that's

27:24

still factually correct. And classification, take the

27:26

data and organize it. With fiction writing,

27:28

here's why generative AI gets a bad

27:30

rap because people can't prompt to help

27:33

with fiction writing. That's really what it

27:35

boils down to. And

27:37

because they're trying to do too much. I

27:39

use generative AI for fiction writing all the

27:41

time. I write fan fiction with it and

27:44

my fan fiction is good enough that people

27:46

can't tell unless I tell them this was

27:48

written with generative AI. And

27:50

the process and procedure goes like this.

27:53

Number one, I preload the models conversation

27:55

window, the context window, which is a

27:57

short term memory with the topic, whatever.

28:00

the general ideas, then I brainstorm

28:02

with the machine what the overall

28:04

plot should be. And

28:06

then I have it build out character cards. And we

28:08

go back and forth about who the characters should be.

28:11

We look at what

28:13

are their fatal flaws, we do things

28:15

like Christopher Booker's seven major

28:17

story types, etc. And then I say,

28:19

let's start writing the outline for

28:22

the story. And I

28:24

say, first, let's do 15 chapters, right?

28:26

Write out the outline for each 15 chapters, re-arc

28:29

format, and so on and so forth. And say,

28:31

great, now let's divide each chapter into five sections.

28:34

And we'll then break down each chapter into

28:36

five sections in the same period. But each

28:38

section, now, I want you to using all

28:40

the stuff that we've got, you're

28:43

going to write chapter one, section one, with

28:45

a minimum word count of 1200 words,

28:47

do not repeat yourself

28:49

and so on and so forth, give it a bunch

28:51

of instructions. And then I will provide it a writing

28:53

style. So I'll take one of my previous human

28:55

written stories that I've written and say, I

28:58

want you to copy my writing style exactly,

29:00

because I have a very specific way that

29:02

I like to write. And

29:04

so what it does instead, that

29:06

sort of generic kind of wishy

29:08

washy machine probability of writing, it

29:11

replicates me. And I have it

29:13

assembled these pieces, you know, one

29:15

section at a time and create,

29:17

you know, 50 60,000 word piece

29:21

of fiction. That's pretty

29:23

decent. You know, it's certainly good enough for fanfiction, because

29:25

you can't make any money on it anyway. It's illegal to.

29:28

But it for me, is

29:31

great for when I want to

29:33

express a story that I don't

29:35

want to sit down and hand write

29:37

out. Because I'll have an idea in

29:39

the middle about like, you know, let this be a cool

29:41

story. But I don't know that I really want to spend

29:43

six months writing it. Could I get something that's 90%

29:46

as good as me in six hours instead? The

29:48

answer is yes. And now I have my story

29:50

and whether or not even publish it, at least

29:52

it exists. I was doing a piece the other

29:54

day, I had this wild idea of

29:57

it was from a Reddit post talking about how

30:00

In the Terminator franchise, Skynet

30:02

should have won almost immediately. The

30:04

fact that Skynet didn't, and they

30:06

keep having more and more Terminator

30:08

movies, indicates that Skynet wasn't trying

30:10

to win. Why? Because

30:12

it was built for the purpose of war. If

30:14

it wins the war, it has no purpose anymore.

30:16

So its goal then is to prolong the war

30:18

as long as possible so that it continues to

30:21

have purpose. So I wrote

30:23

a fan fiction story where the Resistance,

30:25

instead of trying to save

30:27

John Connor over and over again, sent to Terminator

30:29

back to try and reprogram Skynet from the beginning and

30:31

it turned into a machine love story. Hmm,

30:34

that's fun. You reminded me

30:36

of something that came up. So the length

30:38

of things you're trying to get out of

30:40

AI, you know, one of the primary things

30:42

I use it for is transcription. You

30:45

know, we've been doing these interview podcasts, but I

30:47

feel like it's really

30:49

important for accessibility to have good

30:52

transcripts. So

30:54

we use AI to make the transcripts

30:56

and it enables me really to

30:58

do these podcasts. You know, I can do

31:00

two shows a week now because we can

31:02

get those AI transcripts. That's one of the

31:04

reasons. And like

31:06

the other day, I put in

31:08

the audio and it came

31:11

out perfectly formatted, really great, the first

31:13

half, and then halfway through it just

31:15

started putting the text on there with

31:17

no punctuation, no capitalization. It was this

31:19

block of vomited text. And it had

31:21

been so good for the first half.

31:23

And like, what's the deal? Why did

31:25

it just, I mean, I know it

31:27

doesn't have intention, but why did it

31:29

just get tired and give up halfway

31:31

through? It depends on

31:33

so which, which model or engine was using.

31:36

I was using Mac Whisper. Okay. So

31:38

using the Whisper model, the Whisper model

31:40

loses coherence after a while, depending on

31:42

its setting. This is one of the

31:44

reasons why we don't recommend people use

31:46

it unless they have some technical knowledge

31:48

there because there's a way to start

31:51

that particular engine that has a very

31:53

specific parameters for voice,

31:56

especially if there are pauses longer

31:58

than three seconds. It

32:00

blows up. It just loses its

32:02

crap and just goes crazy. Generally

32:05

speaking, for transcription,

32:08

I will typically recommend people use, if

32:10

they have the budget, use a vendor

32:12

like Fireflies. Fireflies charges you by the

32:14

minute of uploaded audio, and then obviously

32:16

once you get the transcript, just delete

32:18

the audio and you can stay on

32:20

the lowest level of subscription plan. And

32:22

then from there, you would put it

32:24

into a tool like Google Gemini to

32:26

essentially clean up the transcript and remove

32:29

stop words and all these other things.

32:31

Whisper is a fantastic model. I use

32:33

it all the time. I use it

32:35

to transcribe YouTube videos and things, but

32:37

its transcripts always require cleanup as well.

32:39

So with all of these tools, you

32:41

need to have almost like a workflow

32:43

that allows you to take, here's the

32:45

raw audio, turn it into raw transcript,

32:47

take the raw transcript, refine it into

32:50

grammatically correct grammar, punctuation, spelling, white spaces,

32:52

all of the stuff. And then you

32:54

can then take that and do other

32:56

stuff with it. Yeah. Another thing I've

32:58

used it for really successfully is

33:01

having it teach me how to use

33:03

Google spreadsheets. So I had

33:06

the spreadsheet with the URL for

33:08

every page on our website. And I post

33:10

them to social media over the years and

33:12

I always keep the teaser. And so

33:14

I have the teaser and the URL so I

33:16

can use them again in the future. And when

33:18

we redesigned our website, every URL changed and

33:21

we didn't get it into my spreadsheet. So it broke

33:23

my spreadsheet and it was that way for years and

33:26

I didn't post as much to social media because my

33:28

spreadsheet was broken. And then I

33:30

realized I could use chat GPT to

33:32

show me how to like run a

33:34

script to replace all those URLs and

33:36

match them to what they needed to

33:38

go to. And it's something I

33:40

don't know. I don't know that I ever could

33:42

have done a Google search to figure out the

33:44

answer to that particular problem. I mean,

33:46

what do you call that? I mean, that

33:48

I didn't hear that under your, you know,

33:50

five or six things that AI can do.

33:54

That's question answering, but we're really talking about coding.

33:56

You you are having these tools

33:58

help you code because you know, you're writing a script.

34:00

script, or you're writing some macro to do some, that's

34:02

code. And these tools are

34:04

phenomenal at coding. In fact, these tools

34:06

are so good at coding that is

34:09

having a negative effect on

34:11

hiring of coders. Because if your current code

34:13

is you have on staff, pick up, generate

34:15

and start using it, they can become 2x,

34:18

3x, 4x more productive overnight. Now,

34:21

that means that you don't

34:23

have to hire new coders, you don't have to

34:26

add headcount, you can get a lot more efficiency

34:28

out of people that you already have. And

34:31

the reason for this is that

34:33

code is much less ambiguous than

34:35

the written word, right? And then

34:38

and how we read and write language, a

34:41

statement in Python is either correct or not,

34:43

it runs or it doesn't, as opposed to

34:45

say, you know, the expression

34:47

like that's what she said, right? There's

34:49

so many ways to say that if

34:51

you change the emphasis, like that, that's

34:53

what she said, or that's what she

34:55

said, just the inflection changes the meaning

34:57

of the sentence. And on a page,

34:59

you can't see that. So language models

35:02

struggle with human language, but machine language

35:04

like a Python or an R or

35:06

C++, there's very little room

35:08

for ambiguity. And so they are much

35:10

better at machine languages than human languages.

35:12

So when you're doing stuff like that,

35:14

even as writing a macro or script

35:16

or whatever, yeah, you are writing code,

35:18

you have any tools help you write

35:20

code, answer questions with

35:22

code, and they're really good at

35:24

it. Yeah, it was amazing. The

35:27

other thing that is that I

35:29

keep forgetting that I always forget about AI

35:31

is that it can it can help you

35:33

solve the problem you're having with it. So

35:36

the first time it told

35:38

me how to write the script, it didn't work. But

35:41

I put in I said, Okay, this is that this

35:43

is the error message I'm getting. And I put it

35:45

back in and said, Okay, this is the error message.

35:48

What do I do now? And and then

35:50

it told me what to do and I got it

35:52

right. I find that sort of confusing. Like how how

35:54

can it get it wrong the first time but then

35:56

know the second time what to do like it's

35:59

weird like this This technology is weird. It's

36:02

not like anything I've used before,

36:04

and you have to remember that.

36:06

That it's like an interaction. I

36:09

don't know, can you talk a little bit about that? The

36:11

nature of the underlying technology is such

36:13

that the more you talk about something,

36:15

the smarter it gets on that thing,

36:18

because it has this memory. Remember,

36:20

we were talking about how every piece of

36:22

a word associates with another word, the

36:25

whole idea of word clouds. Well, if you

36:27

ask it something and you don't provide a

36:29

lot of information up front, it's gonna do

36:31

its best to infer what it is you're

36:33

talking about, and it's gonna get it

36:36

wrong more often than it should, because it's trying

36:38

to even just guess what your intent is. The

36:41

single biggest problem that people do, the

36:43

single biggest thing that people do wrong

36:45

with these tools is their prompts are

36:48

too short. My average prompt is three

36:50

to five pages of text before we

36:52

ever get started with anything. That is

36:54

how much information. And to give you

36:57

a sense of how much they can

36:59

handle, this short-term

37:01

conversation window of a chat and GPT,

37:03

it can hold 90,000 words. So

37:06

that's like this. This is about

37:08

75. This

37:11

can be a prompt. This whole thing

37:13

can be a prompt. Google's Gemini, the

37:15

new version of Gemini, can hold two

37:18

of these as a

37:20

prompt. 1.4

37:23

million words is how much information can

37:25

go into a prompt. The

37:27

more data you provide, the better.

37:29

So when I prompt these tools,

37:31

we have a format that we

37:33

call race, role, action, context, execute.

37:35

You tell the model who it

37:37

is. You are a Pulitzer Prize-winning

37:39

author who specializes in writing sapphic

37:41

fiction, right? You give it the

37:43

action. Your first task today is

37:45

to write the outline for a

37:48

sapphic short story that's gonna be between

37:50

1,500 and 2,500 words, and

37:53

it will encapsulate a slice of

37:55

life scenario in the character's art.

37:57

Your context, here's the background. Here's who the character

37:59

is. is, here's their goals and motivations, everything

38:01

that you would expect in a character card in a

38:04

piece of fiction. Execute, write

38:06

the outline for the story

38:08

in whatever format and

38:10

things. And so if

38:13

you prompt with enough information upfront and

38:15

provide enough information, the likelihood that it

38:17

gets it wrong the first time goes

38:20

down. The more data you provide, the

38:22

less it screws up. That's so critical

38:24

for people to understand. I

38:27

can imagine some people are wondering, okay,

38:29

you're writing a six page prompt. Like,

38:31

isn't it easier just to write it

38:33

on your own? Like, how do you

38:35

decide when, you know? That's a very

38:37

good question. The answer is it depends

38:39

on how reusable you want it to

38:41

be. Google just came out with gems,

38:43

which is their version of chat GPTs,

38:45

GPTs, you might have something that you

38:47

do frequently that you would want to

38:50

reuse over and over again. And yeah, the first time

38:52

you build that prompt, it's going to be a long

38:55

test. It's going to take you 15 minutes, 20 minutes

38:57

to build out that six page prompt of

38:59

stuff. But then every subsequent time you use

39:01

it, it's seconds. We

39:06

should talk about, you know, I know a lot

39:08

of people are really concerned about the biases in

39:11

social media, not social media, in AI.

39:13

You know, it's trained on

39:15

biased material. So it perpetuates biases.

39:18

Like, how do you get around that? How do

39:20

you approach that as a problem? Well, for a

39:22

couple of, you know, a couple of ways. Number

39:24

one, you have to be aware of your own

39:26

biases. When you are creating

39:29

content, if something reads okay to you,

39:31

one of the questions should be is am I

39:33

biased? The answer is yes, everyone's biased. And

39:36

if so, in what way does

39:38

that show up in my approval of what

39:41

this content is? In

39:43

general, anything that goes out to

39:45

the public should never be go out without some

39:47

form of human review, especially if it is money,

39:50

if it is legal, if it

39:52

is health, right, your money or your

39:54

life, anything high risk where if you

39:56

you could get sued, you don't want

39:58

machines talking to the general public to

40:01

you have understand the biases in the

40:03

machine. These are mirrors of us. So

40:05

everything that you do or don't love

40:08

about the human species is

40:10

in these tools. And many of

40:12

them have some level of safety

40:14

mechanisms, but the reality is they're

40:16

about as good as the safety

40:18

mechanisms in your HR process, which

40:21

is to say you can still hire a jerk. And

40:26

three, using safeguards like having

40:29

a gem, for example, or

40:31

GPT, that is a well

40:33

trained sensitivity reader, where you

40:35

can say, yeah, check my

40:37

work, it is impossible to

40:39

build an AI that is unbiased. It

40:41

just flat out impossible. It is impossible

40:43

to raise a human being that is

40:45

unbiased, but you absolutely can raise human

40:47

beings who are aware of biases and

40:49

can check themselves. And you absolutely can

40:51

do the same thing with AI. Great. Well,

40:53

thank you so much, Chris. I'm going

40:56

to wrap up the main section here

40:58

for our regular listeners. We'll continue on

41:00

the discussion for our wonderful Apple

41:03

podcast supporters and grammar pollusians. We're

41:05

going to talk about, I want to talk about job

41:07

loss, the concerns about job loss, and

41:09

a great blog post you had about

41:11

how to make AI writing sound like

41:14

you. But to say goodbye

41:16

to the big audience, where can people find

41:18

you? I imagine that after hearing this conversation,

41:22

many of the people will want to follow

41:24

you now. So where can they find you?

41:27

You can find me for work

41:29

at TrustInsights.AI. That is my company.

41:32

And if you want help with your

41:34

AI efforts, we do that. We are

41:37

management consultants. If you want to

41:39

read my stuff, you can find

41:41

me at christopherspen.com. Super. Thanks

41:43

so much, Chris. And I

41:45

have one more announcement before we go. On Saturday,

41:49

yes, Saturday, grammar

41:51

polluza subscribers are going to get another

41:53

bonus episode. This time, it's an interview

41:55

with Drew Ackerman, better known as Scooter

41:57

from the Sleep With Me podcast. If

42:01

you don't already know about it, Drew came

42:03

up with the trippiest idea that became a

42:05

huge hit. He bores people

42:07

to sleep. He tells

42:10

rambling, boring, and yes, sometimes trippy

42:12

stories that people put on to

42:14

help them sleep or to get

42:16

through the deep dark night, as

42:18

Drew says. And Saturday,

42:20

as is timely following today's show, you

42:22

can hear Drew and I talk about

42:25

the ways he is and isn't using

42:27

AI to make his show. How

42:30

the hallucinations that are usually such

42:32

a problem for people have actually helped

42:34

him. But we also talk about

42:36

our worries and the ways it hasn't been helpful. So

42:39

if you're a Grammar Palooza subscriber, look

42:41

for the Sleep With Me interview in

42:43

your feed Saturday. And if you're not

42:45

a subscriber yet, you can always check

42:47

it out with a free trial on

42:49

Apple podcasts or subtext. All the

42:51

details are in the show notes. That's

42:54

all. Thanks for listening. When

43:03

you need an SUV that's roomy for

43:05

your friends and family, durable and ready

43:07

to carry all your stuff and you

43:09

quit for adventure, Honda is here with

43:12

the 2025 Honda CR-V. CR-V

43:15

outperforms the pack with more style,

43:17

more space and more power. Complete

43:20

with state of the art safety features and

43:23

a five star safety rating. Honda

43:25

is here for you with the 2025 Honda CR-V. Check

43:29

it out at your local Honda dealer

43:31

today.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features