Fired by a bot? What happens when AI takes over performance reviews and layoffs

Fired by a bot? What happens when AI takes over performance reviews and layoffs

Released Monday, 3rd March 2025
Good episode? Give it some love!
Fired by a bot? What happens when AI takes over performance reviews and layoffs

Fired by a bot? What happens when AI takes over performance reviews and layoffs

Fired by a bot? What happens when AI takes over performance reviews and layoffs

Fired by a bot? What happens when AI takes over performance reviews and layoffs

Monday, 3rd March 2025
Good episode? Give it some love!
Rate Episode

Episode Transcript

Transcripts are displayed as originally observed. Some content, including advertisements may have changed.

Use Ctrl + F to search

0:00

Brought to you by Purdue

0:02

University. The only university

0:04

on fast companies brands that

0:06

matter list. Four years running.

0:09

Honored as a heritage brand,

0:11

Purdue pushes innovation and partnership

0:14

to make a consequential impact

0:16

on the world. Learn more

0:18

at Purdue.edu/partner with

0:20

us. I

0:23

think any worst-case scenario

0:26

starts with leaders

0:28

in an organization

0:30

outsourcing the implications of

0:33

AI to AI. Welcome to

0:35

Fast Companies, the new way

0:38

we work, where we take

0:40

listeners on a journey

0:42

through the changing landscape

0:44

of our work lives

0:46

and explain exactly what

0:48

we need to build

0:50

the future we want.

0:52

I'm Fast Company Deputy

0:54

Editor Kathleen Davis. I've

0:56

been covering workplace culture

0:59

for over a decade and

1:01

there are a few topics

1:03

that consistently bring out strong

1:05

feelings. Performance reviews are at

1:07

the top of that list.

1:09

While most people would agree that it's

1:12

a good thing to have a tool

1:14

to measure how employees are doing at

1:16

their jobs and a time for managers

1:18

to discuss career advancement, very few

1:20

seem to think that the

1:22

way performance reviews are currently

1:24

set up is working. In fact,

1:27

a Gallup survey last year found

1:29

that only 2% of human resource

1:31

officers at major companies think their

1:33

performance management system is working, and

1:36

66% of employees feel the same.

1:38

One of the biggest complaints

1:40

employees have about performance reviews

1:42

is that they're so subjective.

1:44

What it takes to be considered good

1:46

at your job or eligible for a

1:49

raise or promotion is often down to

1:51

the opinion of just a couple of

1:53

people, which means it's fertile ground for

1:55

bias. So if both employees and leadership think

1:57

performance reviews are broken... Could artificial

1:59

intelligence be the magic bullet that

2:01

fixes it? The new crop of

2:03

startup selling AI-powered performance management tools

2:05

certainly think so. But is it

2:07

just swapping out human bias for

2:09

AI bias? Is something as nebulous

2:11

as being good at your job

2:13

quantifiable? And are humans ready to

2:15

be evaluated by a robot? To

2:17

help me parse out the benefits

2:19

and drawbacks of using technology in

2:21

both performance reviews and layoffs is

2:23

Brian Ackerman. The head of AI

2:25

strategy and transformation at Corn Ferry,

2:27

a management consulting firm. Brian, thank

2:30

you so much for being here.

2:32

Thanks for having me. So I

2:34

opened this episode with some stats

2:36

on how disliked performance reviews are.

2:38

That's what we're here to talk

2:40

about today. They're both disliked by

2:42

leadership and employees. And one of

2:44

the biggest complaints that we've seen

2:46

in our coverage is that they

2:48

often feel meaningless or only punitive

2:50

and that they're incredibly subjective and

2:52

open to bias. For those reasons,

2:54

it's easy to see why using

2:56

an AI tool could be tempting.

2:58

What's your view on how AI

3:00

can help solve those problems? Well,

3:02

I would not argue with any

3:04

of those findings or that analysis.

3:06

It feels a little bit like

3:08

we've lost the plot on performance

3:10

reviews. You know, who does this

3:12

help really? So I think that

3:14

needs to be the foundation of

3:16

the discussion for where best to

3:18

use AI. Let's pull back for

3:20

a second and say, what is

3:22

the fundamental thing we're trying to

3:24

change? What is a valid, meaningful,

3:26

meaningful input? for the enterprise and

3:28

then quite frankly can we reintroduce

3:30

so to speak a what's in

3:32

it for the employee getting this

3:34

review because those connections you know

3:36

I think I would agree with

3:38

you are pretty tenuous so I

3:40

think that is the entry point

3:42

for the discussion about AI we

3:44

tend to think about this topic

3:46

with a couple of dimensions one

3:48

would be just the official of

3:50

the process, we can talk through

3:52

that, and that's got an employee

3:54

side of that as well as

3:56

a reviewer or a waiter perspective

3:58

and HR perspective. Also, I think

4:00

there's an opportunity to change using

4:02

AI the way performance reviews are

4:04

consumed by the person, to make

4:06

them more understandable, more impactful, more

4:09

relevant, and then probably even more

4:11

importantly turn these back into. development

4:13

tools, turn these back into tools

4:15

that help you as a professional,

4:17

move along in your career, all

4:19

of which is kind of left

4:21

unsaid with the, hey, you're a

4:23

three out of five or a

4:25

one out of two, or you

4:27

know, whatever your organization's approach to

4:29

this happens to be, in box

4:31

seven of a nine box. Yeah,

4:33

I mean, it's exactly that. I

4:35

think one of the things that

4:37

a lot of people hate about

4:39

performance reviews is that they do

4:41

seem meaningless in that like, what

4:43

the hell does a three mean

4:45

versus a two? What does like

4:47

super exceeds expectations versus exceeds expectations

4:49

mean? I guess on like a

4:51

basic level, when you think of

4:53

AI, it's data driven. Is there

4:55

a pitfall or chance that AI

4:57

kind of goes too far in

4:59

that direction and turns everything in

5:01

what feels like a very human

5:03

process, right into just a set

5:05

of data? Absolutely, exclamation point. You

5:07

can kind of look at these

5:09

three categories in turn. From an

5:11

efficiency perspective, let's talk about that

5:13

for a minute. Tools now, we'll

5:15

use General AI to help draft,

5:17

have the manager draft notes, which

5:19

is all, okay, subject to the

5:21

same challenge of what the quality

5:23

of the draft is and, you

5:25

know, how much work I'm going

5:27

to have to do to make

5:29

that actually meaningful as a manager.

5:31

Am I saving any time anyway?

5:33

Right. allow a manager to have

5:35

a bigger, more robust set of

5:37

data, maybe engagement data, maybe operational

5:39

data to add to the mix

5:41

when they're crafting a report. Another

5:43

good use case, but as you

5:46

said, absolutely subject to the quality

5:48

and accessibility of that kind of

5:50

data that can go and create

5:52

more of a problem. than it

5:54

solves if the data quality isn't

5:56

great. The question of how are

5:58

we making a manager's life easier

6:00

or harder, and since a part

6:02

of the challenge here is with

6:04

respect to managers of the world,

6:06

the ones who kind of do

6:08

it by road. right, who aren't

6:10

really investing in the time to

6:12

form a basis for a good

6:14

performance review, is this going to

6:16

help or hurt? All of that

6:18

around is generative AI and adding

6:20

more data into the mix, you

6:22

know, making this process easier and

6:24

more efficient and more consistent and

6:26

safer and less bias? Or is

6:28

it just adding complexity that then

6:30

the manager is somehow supposed to

6:32

still make sense around? Yeah, when

6:34

we talk about kind of adding

6:36

complexity and the manager role in

6:38

it, I'm a manager. I use

6:40

a software system for... our performance

6:42

reviews and of course like everything

6:44

else it's popped up when I'm

6:46

writing something and I'll say like

6:48

make this better with AI and

6:50

as a writer I never use

6:52

it of course you know I'm

6:54

like no you will not make

6:56

me better with AI but I

6:58

wonder from an employee standpoint does

7:00

it take subjectivity and bias I

7:02

really want to dig into and

7:04

that is certainly a complaint with

7:06

performance reviews but does it feel

7:08

differently as an employee than to

7:10

get a performance review that's written

7:12

by a bot and maybe better

7:14

in a good way, maybe because

7:16

it takes that subject to video.

7:18

I think we have to be

7:20

cognizant that it's February 12, 2025,

7:23

and this technology is as bad

7:25

as it's ever going to be,

7:27

right? It only gets better from

7:29

here. So right this second, if

7:31

you go into those tools and

7:33

they're drafting language, they will come

7:35

across a bit stilted, a bit

7:37

sterile, right? In the interest of

7:39

pulling... bias out of it or

7:41

at least attempting to, of being

7:43

consistent, they're absolutely coming across as

7:45

sterile and need a lot of

7:47

input. Does that particular element change

7:49

as these tools improve as you

7:51

start seeing the new models make

7:53

their way into these products? probably,

7:55

right? So you could see that

7:57

the quality of these drafts are

7:59

going to improve, but I think

8:01

that is different from that it's

8:03

going to not sound or act

8:05

like it's written by a bot.

8:07

We are skeptical that this is

8:09

going to be good enough that

8:11

you're still not going to need

8:13

essentially the same human review that

8:15

you want, especially as you said.

8:17

this process is screaming for better

8:19

human interaction. So I think it'll

8:21

improve, but I'm not sure that

8:23

it's really going to become any

8:25

significantly easier, as it will in

8:27

some other areas where you're using

8:29

generative AI to generate content. Yes,

8:31

it sounds like you're saying, in

8:33

which makes a lot of sense.

8:35

Is this a performance review? Is

8:37

it? essentially human process. AI in

8:39

the best case scenario can help

8:41

improve it, but it's not going

8:43

to replace it. You do not

8:45

envision a future where my performance

8:47

review is with a computer and

8:49

not with a human at all.

8:51

Yeah, that was easy, at least

8:53

for me. No, I hope not,

8:55

and I kind of dread the

8:57

day that we go down that

8:59

path. That is not to say

9:02

that you can't give the human

9:04

superpower, you can't... give them access

9:06

to a broader swath of data

9:08

to make them a better manager

9:10

and do a better review, but

9:12

it's still the person providing feedback

9:14

to another individual. As I said,

9:16

I think you got to keep

9:18

coming back to why are we

9:20

doing this process again? We've lost

9:22

the plot. If the purpose is,

9:24

as it should be, to provide

9:26

meaningful guidance that addresses the individual's

9:28

growth and the organization's requirements in

9:30

the context of the performance of

9:32

their role, then let's go do

9:34

that. And none of that screams,

9:36

let me take the human out

9:38

of the equation and give it

9:40

all to a generative AI capability.

9:42

You mentioned that today in 2025,

9:44

these tools are not very advanced,

9:46

but they obviously like all of

9:48

these. get more advanced. Are there

9:50

specific tools that you've seen that

9:52

are more common than others, things

9:54

that people are using, and kind

9:56

of what are the challenges and

9:58

opportunities with those? Every performance management

10:00

module of every HR tech software

10:02

company is introducing, just as you

10:04

said, Kate, the hey, make this

10:06

better with AI, that's really becoming

10:08

commonplace very quickly. There is a

10:10

generation of bespoke tools coming up

10:12

from employee engagement world that are

10:14

using generative AI to aggregate information

10:16

together for the reviewer to make,

10:18

you know, surface insights from data

10:20

as part of that drafting process.

10:22

To be honest, I think we

10:24

see the most. use in the

10:26

space in the general purpose tool,

10:28

so the Microsoft copilots, the chat

10:30

GPTs, right, the ones that were

10:32

also using to create presentations and

10:34

word documents and emails these days,

10:36

that's still the most common use,

10:39

and then cutting and pasting it

10:41

into reviews. But they're all that

10:43

GPT-40 generation of underlying foundation model,

10:45

so they all act that way.

10:47

And that's the part as the

10:49

models get reintegrated into the HR

10:51

text act. the quality of the

10:53

answer will improve, but right now

10:55

it's a drafting tool. You could

10:57

even argue, I know not all

10:59

companies do this, but a lot

11:01

of companies then do legal and

11:03

compliance reviews on this stuff. There's

11:05

no indication that that's going to

11:07

go away, right, because the level

11:09

of quality of the results isn't

11:11

quite there to pass muster on

11:13

a legal review either. I can

11:15

see a use case for AI

11:17

in performance reviews when it is

11:19

data focused, when it's like your

11:21

goal was to write 50 articles

11:23

a month rather than me going

11:25

in counting and seeing how many

11:27

articles I wrote a month to

11:29

see if I reach that goal.

11:31

We use AI to like do

11:33

that kind of number analysis and

11:35

data analysis. Are there tools or

11:37

use cases in that way? And

11:39

is that a good use of

11:41

AI for performance reviews? You're beginning

11:43

to see the easy ones, the

11:45

two easy ones are engagement, you

11:47

know, 360 degrees. an engagement kind

11:49

of data about the individual being

11:51

brought into performance management systems to

11:53

complement the narrative. We're beginning to

11:55

see operational data like sales performance

11:57

data coming in. Hey, you made

11:59

your numbers or you didn't make

12:01

your numbers and here's how being

12:03

incorporated. It's still pretty minor. That's

12:05

pretty new because the integrations that

12:07

have to be done to those

12:09

data sources is still very much

12:11

evolving. The agentic AI capabilities that

12:13

are popping up now will improve

12:16

that. But as of right now,

12:18

that's still pretty nascent. It is

12:20

promising, right to your point. Today,

12:22

if I've got to do that,

12:24

there are a lot of managers

12:26

that just don't. And then those

12:28

that do have to do a

12:30

ton of legwork to have good

12:32

data. to serve as the basis.

12:34

So that is clearly a place

12:36

where you would expect to see

12:38

efficiencies. The magic next step, though,

12:40

of turning that into a robust,

12:42

meaningful review is still, you know,

12:44

the human is still that manager,

12:46

right? I don't see any indications

12:48

of automated performance ratings happening in

12:50

any kind of scale. We see

12:52

point examples, right? So if you're

12:54

a coder. you've got code quality

12:56

tests that can creep their way

12:58

into these kind of reviews but

13:00

that's still pretty niche. Yeah it

13:02

does seem like the best use

13:04

case because as you say a

13:06

lot of managers just kind of

13:08

go on for lack of a

13:10

better word vibes you know like

13:12

oh I think you're meeting your

13:14

expectations I'm not going to do

13:16

the extenuous legwork of figuring it

13:18

out so I have a general

13:20

feeling that you're meeting your expectations

13:22

I'm going to rank you at

13:24

meet expectations rather than have the

13:26

hard data that you can't argue

13:28

with, like, yes, you met your

13:30

goals, no, you didn't meet your

13:32

goals, you exceeded your goals by

13:34

exactly this much. We've talked so

13:36

far a lot from the manager

13:38

perspective. From the employee perspective, I

13:40

can see both intimidating to think

13:42

of it this way and maybe

13:44

promising, depending on maybe your relationship

13:46

with your manager. How do you

13:48

think employees should be thinking about?

13:50

this change? So it's a great

13:52

area first off for an employee

13:55

to look to their employer for

13:57

transparency right it's not a bad

13:59

question to ask how is that

14:01

organization using AI in the performance

14:03

review process that shouldn't be a

14:05

secret because this is already a

14:07

fairly low-trust process for all the

14:09

reasons we talked about so if

14:11

AI is yet another black box

14:13

into this it's not going to

14:15

help that so it's a great

14:17

opportunity for an organization to be

14:19

transparent to gather the messaging that

14:21

is being used to the benefit

14:23

of the employee. But I think

14:25

this is maybe a little away

14:27

from the efficiency side of all

14:29

this. I look at what tools

14:31

are like Nopagelam for the Google

14:33

has, are doing, consuming a report,

14:35

a document, creating an easy to

14:37

understand in the moment, learning asset,

14:39

like a short-form podcast, as Nopagelam

14:41

does an example, and using that

14:43

as a way to make the

14:45

performance review more consumable. by the

14:47

individual, not relying on that PDF

14:49

and, you know, sometimes quite fairly,

14:51

we say, PDF on the head,

14:53

give that employee an asset that

14:55

not only helps them understand what

14:57

a three is and how it

14:59

relates to the mission and their

15:01

job, you know, really use it

15:03

as a way to give a

15:05

different way to consume the performance

15:07

review in addition to, maybe as

15:09

pre-work or follow-up to the manager

15:11

conversation. Google's just introduced the ability

15:13

to actually interact with that. They

15:15

bring real-time voice back into that

15:17

equation. So you can ask your

15:19

performance report a question in between

15:21

your conversations with your manager, again,

15:23

not replacing that manager. So I

15:25

think there's a lot of potential

15:27

if we think differently about how

15:29

do we make this valuable to

15:32

the colleague, to the employee. There's

15:34

some really interesting things that we

15:36

can do with not a lot

15:38

of engineering to... Not have the

15:40

oh it's it's the performance review

15:42

time and I'm going to be

15:44

a 3.2 and I won't know

15:46

why Actually creating a tool out

15:48

of it that the person has

15:50

a better chance to be able

15:52

to use and actually develop from

15:54

it. And that's what excites me.

15:56

Yeah, you know, relatedly something that

15:58

we talk about a lot when

16:00

we talk about AI and also

16:02

we talk about a lot when

16:04

we talk about performance reviews is

16:06

bias and human bias and both

16:08

the pros that like does AI

16:10

perpetuate the bias, does it solve

16:12

against it? I will share with

16:14

you and our listeners that back

16:16

in 2014, one of the first

16:18

articles that I wrote for Fast

16:20

Company that went viral. The headline

16:22

was, this is the one word

16:24

men never see in their performance

16:26

reviews, and the word was abrasive.

16:28

Humans in performance reviews obviously are

16:30

open to a lot of bias.

16:32

There's potential there for AI to

16:34

perpetuate it, but there's potential to,

16:36

for AI to perpetuate it, but

16:38

there's potential to, for AI to

16:40

flag those sorts of things. What's

16:42

your experience with how AI can

16:44

help or hinder bias review? Jenny,

16:46

I should do perfectly fine on

16:48

that. And he already is with

16:50

the drafting tools that are out

16:52

there. We're seeing them make suggestions,

16:54

hey, we're this a little differently.

16:56

And so from that perspective, I

16:58

think that's a good thing and

17:00

that's a benefit. But that is

17:02

surface. If you ask me the

17:04

question that's on everybody's mind, these

17:06

models are only as good as

17:09

the data they're trained on, if

17:11

the data has bias, so does

17:13

the model, this use case is

17:15

not going to solve for that

17:17

obviously. So the need for the

17:19

human review, the need for the

17:21

HR professional review, that's one of

17:23

the reasons why I don't think

17:25

it's going away. Because that's going

17:27

to show up more subtly in

17:29

the review text than the George

17:31

Carlin words I shouldn't say, right?

17:33

And I don't think there's a

17:35

magic answer to that quite yet.

17:37

You would hope that's something that

17:39

makes its appearance as the model,

17:41

underlying frontier and foundation models continue

17:43

to improve. but they still require

17:45

an oversight and an implementation of.

17:47

of testing and architecture and monitoring

17:49

tools around the core application, which

17:51

is absolutely essential in this case.

17:53

I want to ask about the

17:55

role of tech in AI-powered tools,

17:57

not just in performance reviews, but

17:59

also in PIP's performance improvement plans,

18:01

such as much dreaded, and in

18:03

layoffs, both meta and Microsoft announced

18:05

recently that they're conducting layoffs based

18:07

on. poor performance, meta even gave

18:09

it that wonderful jargon name, non-regrettable

18:11

attrition. On the surface, it appears

18:13

the fairest way to make cuts,

18:15

if you have to make cuts,

18:17

but it also assumes that there's

18:19

kind of well-defined goals in equal

18:21

metrics for everyone in a clear

18:23

understanding of what good performances versus

18:25

what poor performance looks like. And,

18:27

you know, as we've talked about,

18:29

that is not... always the case

18:31

or frequently not the case at

18:33

most companies. What are the pros

18:35

and cons of using AI tools

18:37

to help make those sorts of

18:39

decisions? Who's an underperforming employee? It'll

18:41

mirror a little bit but only

18:43

to appoint the discussion around the

18:45

review itself as a tool that

18:48

augments the aggregation of data that

18:50

you are going to as the

18:52

human used to make that employment

18:54

decision. Sure, subject to the accuracy

18:56

of the data and the accuracy

18:58

of the interpretation, right, by the

19:00

AI tool. So do I support

19:02

bringing in if you're evaluating the

19:04

performance of the developer, subjective team

19:06

performance 360 degrees for reviews from

19:08

his team and coding statistics about

19:10

how that individual is actually coding

19:12

and bringing that together using generative

19:14

AI as a draft to have

19:16

a hypothesis? Okay, yeah, I can

19:18

certainly go with that. Are we

19:20

anywhere near that being a filter

19:22

or a cutoff or a decision

19:24

point? Not that I'm aware of,

19:26

and I think that would be

19:28

a dangerous place to be. You

19:30

can see a word. If there

19:32

are roles in an organization that

19:34

are so clear on, just as

19:36

you said, so clear on the

19:38

metrics, the KPIs, by the way,

19:40

that probably can't be too many

19:42

of them, that are easily measured.

19:44

Did I make a number? Did

19:46

I make a number? Did I

19:48

not make a number? Did I

19:50

have a certain defect rate if

19:52

I'm a coder or not? Did

19:54

I write the thing, the items

19:56

I was supposed to put into

19:58

the world, or did I not?

20:00

Do I need AI in the

20:02

mix? What's AI bringing in the

20:04

mix, right? Yeah. You know, again,

20:06

I don't see organizations using AI

20:08

to identify poor performing salespeople. They

20:10

look in their CRM, right? You

20:12

can sense in my them if

20:14

we're just beginning to use them

20:16

effectively as information gatherers in a

20:18

one-point performance review. Again, where's the

20:20

plot? I think we're a ways

20:22

away from... using it as an

20:25

objective criteria to eliminate and having

20:27

that be the one and only

20:29

decision poor. We do get it,

20:31

we get us, hey, is AI

20:33

going to fire me? It's a

20:35

valid question and a valid worry.

20:37

Yeah, I think the more that

20:39

we see these sorts of headlines

20:41

and the more that AI becomes

20:43

such a big part of our

20:45

working lives, it does become a

20:47

real fear. And something I keep

20:49

going back to and I talked

20:51

about when I talked about AI

20:53

and hiring is just the nuance

20:55

that is missed and I think

20:57

when it's strictly database. and we

20:59

saw some of the meta employees

21:01

are posting on LinkedIn saying you

21:03

know actually I was a good

21:05

performer or I was on parental

21:07

leave and that's why I didn't

21:09

make my numbers or you know

21:11

all of these very human things

21:13

that happen you definitely need a

21:15

human in there to help understand

21:17

them rather than being judged solely

21:19

as a number. I mean the

21:21

reality is the vast vast majority

21:23

of our clients today will make

21:25

that as a subjective that as

21:27

a one of the most challenging

21:29

difficult impactful leadership decisions that have

21:31

to be made period full stop

21:33

and they are made by humans

21:35

they are almost never black and

21:37

white. They are almost never so

21:39

cut and dried that you can

21:41

just go by the objective data,

21:43

right? Just as you said, there's

21:45

a constellation of reality that informs

21:47

the way a person performs in

21:49

their job that goes far beyond

21:51

their objective metrics. So it's just

21:53

not a space that should, in

21:55

our opinion, lend it to being

21:57

delegated or outsourced to generative AI.

21:59

I don't think any ever, but

22:02

we'll see how this goes. So

22:04

speaking of, when a business approach

22:06

you for a guidance on AI

22:08

and performance management, what are their

22:10

hopes and how do you advise

22:12

them? We get a lot of

22:14

questions right along the lines we're

22:16

discussing. Right now, usually from the

22:18

operational efficiency perspective, right, because managers

22:20

don't like to do it, the

22:22

systems that do it aren't great,

22:24

the feedback is usually pretty obtuse

22:26

and... delivering it is usually a

22:28

challenge. Other than that the process

22:30

is amazing and incredibly well received

22:32

by the participants. We try to

22:34

again get our clients and work

22:36

with our clients to pull back

22:38

a bit and get at the

22:40

transformation that employees are looking to

22:42

make to move along in their

22:44

career and how does that align

22:46

with what the organization needs to

22:48

achieve from the talent to get

22:50

to their business strategy. When we

22:52

do that effectively, the conversation expands

22:54

from how do I make it

22:56

as sufficient as I can be,

22:58

that's still part of it. But

23:00

begins to talk about how do

23:02

we help the leader or the

23:04

manager deliver it more effectively, how

23:06

do we coach these managers at

23:08

scale, but sometimes using AI, to

23:10

be a better feedback giver. And

23:12

then from an individual perspective, what

23:14

do you do with it? Great,

23:16

you got it, you know. But

23:18

a lot of organizations are giving

23:20

ratings without development guidance still, right?

23:22

So you try and bring those

23:24

together. A little bit trying to...

23:26

Move the conversation away from what

23:28

your organization needs is they need

23:30

to objectively rate their employees and

23:32

move it back to a development

23:34

question. How do you improve the

23:36

capabilities and the performance of your

23:38

workforce? And they are a different

23:41

question, even though performances Emily. Well,

23:43

and I think, you know, as

23:45

you're saying, this one thing I

23:47

think that we cover a lot,

23:49

that's a way that people approach

23:51

performance reviews that's flawed, is that

23:53

This is sure, I'll do it

23:55

this one time, we'll get it

23:57

done with, and then we won't

23:59

think about it at all until

24:01

we have to do it again

24:03

next year. And performance management is

24:05

supposed to be an ongoing conversation.

24:07

It's supposed to be, especially with

24:09

career advancement and opportunities that comes

24:11

in along the way. Are there

24:13

effective ways or are there AI

24:15

tools that can help with that

24:17

kind of engagement tools, things that

24:19

kind of go along throughout the

24:21

year and other than this one

24:23

process at this one point? challenge

24:25

and answer enabled by technology has

24:27

been a thing for a while

24:29

now, right? So moving from annual

24:31

performance reviews or quarterly performance reviews

24:33

to something that's more in the

24:35

moment, that's as close to in

24:37

real time as, you know, an

24:39

organization can feasibly make it, that

24:41

has led to kind of a

24:43

coalescing of traditional performance management systems

24:45

and employee engagement tools. That's been

24:47

a technology-driven initiative for some time,

24:49

and it introduces a lot more

24:51

data into the equation when it's

24:53

done well. It can also just

24:55

be yet another thing for a

24:57

manager to do. So now they're

24:59

not only doing the annual performance

25:01

review, they're doing periodic 360s, and

25:03

they're doing project-based assessments. And again,

25:05

it doesn't make the process any

25:07

easier. It adds more touch points

25:09

if it's done well. So you

25:11

would expect and beginning to see...

25:13

AI tools to make that easier,

25:15

as I said, easier to get

25:18

that data together. As an engagement

25:20

stops being surveys and starts being

25:22

more pulses, that's another thing that's

25:24

been going on for a while,

25:26

make it easier to say how

25:28

somebody's doing. You can see AI

25:30

beginning to make that process more

25:32

efficient. We still believe that unless

25:34

you then still consider, does it

25:36

actually help the person? Or are

25:38

you just gathering more data? Fundamentally,

25:40

the reception or the septivity of

25:42

the organization to the performance management

25:44

process is not going to approve.

25:46

And objectively, you probably won't be

25:48

able to look at the process

25:50

overall in the aggregate as an

25:52

organization and say it's making any

25:54

more of an impact just because

25:56

I've added more touch points. That

25:58

rubicross is not cross-cut. So we've

26:00

talked a lot about both the

26:02

pros and the cons and you

26:04

have a little bit alluded to

26:06

where things could go wrong. I'd

26:08

like to end though with a

26:10

worst case scenario and your most

26:12

optimistic scenario for where all of

26:14

this the use of increased technology

26:16

and particularly AI in performance reviews,

26:18

performance management and layoffs in those

26:20

decisions. What's your worst case scenario?

26:22

What's your best case scenario? as

26:24

it applies to performance management, but

26:26

just broadly, are introducing some of

26:28

the most profound leadership decisions organizations

26:30

are going to have to make

26:32

over the next several years, right?

26:34

This is the Pandora's box is

26:36

open, the cover's been ripped off,

26:38

it's been thrown down the hall.

26:40

This isn't changing, this is here.

26:42

So I think any worst case

26:44

scenario starts with leaders in an

26:46

organization outsourcing the implications of AI

26:48

to AI, right? No more. felt

26:50

more obtusely in this kind of

26:52

process, right? So the worst case

26:55

for me absolutely is organizations that

26:57

believe that AI is going to

26:59

take this burden off their plate

27:01

and decide who is going to

27:03

be an employee of my company

27:05

and who will not. There's no

27:07

upside I can see in any

27:09

way you choose to look at

27:11

the problem. So clearly is a

27:13

worst case. Similar worst cases is

27:15

a... that it's a one-sided equation,

27:17

right? That this makes it incredibly

27:19

efficient for the organization. and we

27:21

lose yet another opportunity to make

27:23

this meaningful, relevant, actionable to the

27:25

employee because then trust won't go

27:27

up and we have the opportunity

27:29

to really change the way feedback

27:31

is given, give the human superpowers,

27:33

give them the data and the

27:35

robust feedback tools. make it pre-read,

27:37

make it part of the process

27:39

so that when you are interacting

27:41

with your manager or your reviewer,

27:43

you are having that human conversation

27:45

and not spending your time, well,

27:47

what does a three mean, right?

27:49

I think that's the opportunity here

27:51

to really make it a tool

27:53

for development and advancement of the

27:55

individual and of the organization that

27:57

is happening in the flow of

27:59

work, is happening in the way

28:01

these people are doing their jobs

28:03

every day. That's the potential. I

28:05

hope the nightmare scenario doesn't happen.

28:07

Yeah, I think if I could

28:09

summarize that for myself, it's that

28:11

the worst case scenario is we

28:13

hand it all over to AI,

28:15

the best case scenario is AI

28:17

helps us fix the parts that

28:19

we don't like and have more

28:21

time to improve the human parts

28:23

that we do need to have

28:25

there. If you knew that the

28:27

performance review was a point, hopefully

28:29

of many, that you're getting meaningful

28:31

feedback and guidance and help moving.

28:34

your career law and help do

28:36

your job better? I'm fairly certain.

28:38

You would look forward to it.

28:40

Or at least, maybe that's too

28:42

optimistic, at least not dread it,

28:44

right? At least not dread it,

28:46

right? Yeah, yeah, yeah, okay, that's

28:48

fair, that's fair. But you got

28:50

to do that. If we do

28:52

anything short of that, you will

28:54

look back in this and go,

28:56

you know, all right, we still

28:58

lost the plot. We haven't really,

29:00

you know, you know, found that,

29:02

I guess. Yeah, it has the

29:04

potential then to become yet another

29:06

management fad that we look back

29:08

on and say like, oh, remember

29:10

when we all tried to outsource

29:12

this to AI and what a

29:14

disaster that was. Yeah. Well, Brian,

29:16

thank you so much for being

29:18

here with me. This gave us.

29:20

a lot to

29:22

think about, especially

29:24

in this hotly

29:26

debated topic. I

29:28

really appreciate your

29:30

input. Thank you

29:32

for inviting me,

29:34

Ms. appreciate your input. Thank you

29:36

for inviting me. It was fun.

29:38

For more on AI and performance

29:40

reviews, check more on

29:42

AI for reviews,

29:44

check out the

29:46

show notes for

29:48

this episode for

29:50

related fast company

29:52

articles. And the next

29:54

episode of The

29:56

New Way We

29:58

Work, I'll talk

30:00

about AI and

30:02

AI rights. rights. sure

30:04

to subscribe to

30:06

The New Way

30:08

We Work Way

30:11

you listen so

30:13

you never miss

30:15

an episode. And

30:17

if you like

30:19

this episode, leave

30:21

us a rating

30:23

and review on

30:25

Apple leave us a rating

30:27

The New Way

30:29

We Work is

30:31

hosted by me,

30:33

Kathleen Davis, We and

30:35

produced by Henry

30:37

by me, Kathleen Nelson, and

30:39

and Joshua by Henry

30:41

Mixing by Nicholas

30:43

Torres. and Joshua Christensen with

30:45

mixing by Nicholas Torres.

Unlock more with Podchaser Pro

  • Audience Insights
  • Contact Information
  • Demographics
  • Charts
  • Sponsor History
  • and More!
Pro Features