Episode 6

Executives under pressure from AI hype

Published on: 22nd April, 2025

Are you feeling overwhelmed by the relentless noise around Artificial Intelligence?

Is your board pushing for an AI strategy?

Are consultants selling fear while employees worry about job losses?

In this eye-opening episode of Impact with Digital, host Jay Tikam speaks directly to CEOs, founders, and senior leaders navigating the chaotic, high-stakes world of digital and AI transformation. This is Part 1 of a powerful two-part series that breaks through the confusion—and reveals why most AI and digital initiatives fail before they even begin.


🔍 What You’ll Learn:

  • Why 70–85% of AI initiatives fail—and what that means for your business


  • The six types of AI failure every executive must understand


  • Real-world cautionary tales from Air Canada, Amazon, and Wells Fargo


  • The internal and external pressures pushing companies into AI before they’re ready


  • The silent fears executives feel—but rarely voice

💡 Why This Episode Matters:


Digital transformation isn't about being the first to adopt AI, it’s about being the first to get it right. Whether you're a corporate executive, SME owner, or tech leader, this episode brings clarity, empathy, and strategy to a world drowning in buzzwords and fear-based selling.


👉 This is your map through the chaos, so you can lead with confidence instead of confusion.


⏱️ Episode Timestamps:


  • 00:00 – Feeling the Pressure: The overwhelming AI noise in the boardroom
  • 01:00 – The Inner Dialogue: What executives really feel
  • 02:30 – It’s Not Just You: AI anxiety is everywhere
  • 03:00 – Noise from All Sides: Boards, vendors, regulators, and staff
  • 06:00 – Public Confidence vs Private Doubt: Who do you trust?
  • 09:00 – What If It Fails? Real fears, real risks
  • 11:40 – Six Ways AI Projects Fail: From adoption issues to ethical landmines
  • 16:00 – AI Gone Wrong: Case studies from major brands
  • 21:30 – Why AI Initiatives Crash: The hype, the data gaps, the poor planning
  • 25:00 – The Late-Night Questions: What keeps leaders awake at night
  • 27:00 – The Real Risk: Doing AI for the wrong reasons

🎧 Next up in Part 2:

Jay shares a step-by-step framework to make your AI and digital strategy actionable, people-first, and impact-driven.

🔔 Subscribe & Follow:


Spotify | Apple | Google Podcasts → @ImpactWithDigital

Learn more: Vedanvi.com

💬 Let’s Keep the Conversation Going:


Tag @ImpactWithDigital and use #AIOverload #ExecutiveLeadership #DigitalTransformation

Transcript
Jay Tikam:

You are an executive, and right now you're feeling the weight of the AI

2

:

revolution pressing down from every side.

3

:

Competitors seem miles ahead.

4

:

Consultants constantly warn

you're falling behind your board.

5

:

Anxiously wonders if you're doing enough

and your employees whisper nervously

6

:

about losing the jobs to algorithms.

7

:

It's relentless.

8

:

I'm sure it's exhausting

and frankly overwhelming.

9

:

I get it.

10

:

So let's stop for a second.

11

:

How are you really feeling

about all of this AI hype?

12

:

What's running through your mind

when you're lying awake at night?

13

:

Are you cautiously optimistic, quietly

anxious, or just plain confused about

14

:

where to begin in this two part series?

15

:

Just like episode three and four,

I am stepping into your shoes.

16

:

Today we'll explore exactly what's

on your mind as an executive facing

17

:

ai, your hopes, your fears, your

frustrations, and lay them bare.

18

:

No jargon, no judgment, just empathy,

clarity, and real world understanding.

19

:

And in the next episode, I

promise clarity and confidence.

20

:

I'll provide a clear step-by-step

blueprint to confidently embrace AI

21

:

without falling into the dreaded 70%.

22

:

Of digital transformation failures.

23

:

Stick with me.

24

:

In the next two episodes, I will turn

the madness of AI into manageable,

25

:

actionable steps, bringing not only

method, but genuine excitement,

26

:

performance, and impact to your journey.

27

:

I am your host, Jay Tikam.

28

:

Welcome to episode six of Impact with

Digital, where we explored digital

29

:

strategies from a human perspective.

30

:

Making sense out of the noise, finding

a way to drive performance and impact.

31

:

Making sure you are on the right track

32

:

in this part, one of the two part series.

33

:

I'll get into your mind trying

to understand the dilemma that

34

:

executives face when confronted

with AI thinking they need to do

35

:

something, not knowing where to start.

36

:

No matter how big your business is or how

small it is, no matter what sector you

37

:

are in, we all face the same pressures.

38

:

I'll try my best to understand your

dilemma, and in part two, we'll

39

:

provide a step-by-step approach to

dealing with AI in the most optimal

40

:

way, ensuring that you get the best

return on your investment, and that

41

:

the AI actually works in practice.

42

:

So let's get started.

43

:

I'm sure you're constantly

surrounded by AI noises.

44

:

Some make perfect sense

while others seem like noise.

45

:

Some messages raise important issues

while others make you concerned.

46

:

Let's explore some of the AI messages

coming at you from all directions.

47

:

On a daily basis from different

types of stakeholders.

48

:

Of course, I use AI to

generate these voices.

49

:

So your board is getting concerned

because , it seems like your

50

:

company just isn't doing enough.

51

:

This puts more pressure on you.

52

:

You feel compelled to

start something, anything.

53

:

The other board members and

I are increasingly concerned

54

:

about our AI strategy.

55

:

We are reading about these advancements

daily, and frankly, we need to see

56

:

a clear plan for how we're going

to leverage AI to drive innovation

57

:

and future-proof the business.

58

:

Are we truly keeping pace?

59

:

You read articles on how AI can

make your business more competitive.

60

:

You wonder whether you're missing out.

61

:

You wonder whether your slow response

may see competitors getting ahead of you.

62

:

In the crazy world of ai.

63

:

You're bombarded by consultants and

software vendors urging you to act

64

:

and spend money on their solutions.

65

:

They put the fear of God into you hinting

that if you don't do anything, you're just

66

:

going to be left behind maybe becoming

the next Kodak in the camera market.

67

:

Or Blockbuster in the Netflix market.

68

:

The window of opportunity

with AI is rapidly closing.

69

:

If you are not aggressively

implementing solutions now, you

70

:

risk being fundamentally outcompeted

within the next 18 to 24 months.

71

:

Laggards will simply become irrelevant.

72

:

You're bombarded with executive level

courses from prestigious institutions.

73

:

You wonder whether you

should take a course.

74

:

Just to make sense of this AI stuff

or look competent in front of your

75

:

board and staff, you wonder whether

you should be arranging training for

76

:

your board and maybe also all staff.

77

:

Maybe if they have the skills.

78

:

They could be coming up with ideas

on AI rather than consultants.

79

:

Or maybe you want your employees

to take the Batten and learn about

80

:

AI so that they can implement

solutions in their functional areas.

81

:

Staff are disillusioned with

yet another digital change

82

:

program following the latest fad.

83

:

They fear job losses or at

least disruption to adopt yet

84

:

another new shiny tech solution.

85

:

I.

86

:

We've been promised all sorts

of efficiency gains with new

87

:

technology before, and it

rarely pans out as advertised.

88

:

This AI stuff just feels like another

buzzword that will create more work

89

:

for us in the long run because we'll

have to fix buggy systems yet again.

90

:

New regulation on AI emerging regularly.

91

:

They promised to put guardrails on

something that has huge potential

92

:

for good, but can also cause harm.

93

:

They also put restrictions

on what you can and can't do.

94

:

You feel compelled to follow

regulation because after all, you

95

:

don't want to end up taking your

firm on the wrong side of the law.

96

:

New AI principles, standards, and best

practices emerge daily, and you and

97

:

your organization need to find a way

to stay on top of these standards.

98

:

You see so many cases of

AI failures emerging in the

99

:

press and in research papers.

100

:

This can really be scary,

especially when you're just on the

101

:

path to doing something with ai.

102

:

So there's a lot of noise

out there in the AI world.

103

:

There's a lot of hype as well.

104

:

And you're not imagining

this noise, it's real.

105

:

The pressure is mounting from within

and from outside on you as an executive,

106

:

everyone is shouting all around you.

107

:

What are you meant to do?

108

:

How do you feel?

109

:

With so much noise going on around you,

some of this could be good, some just

110

:

hype, some serving personal interests

rather than serving your interests.

111

:

As an executive owner of the business

or a founder, you have a public face.

112

:

You have to show your shareholders,

investors, the market and staff that

113

:

you have everything under control,

that you're on top of AI opportunities.

114

:

You need to look like you are in control

and at the forefront of the AI wave.

115

:

However, in the quiet of your office

or in the dead of the night when

116

:

you can't sleep 'cause of all these

worries, how do you really feel?

117

:

What are you thinking?

118

:

What should you do next?

119

:

Should you get internal

advice from your IT team?

120

:

Should you get independent

external advice?

121

:

Well, advisors just try and sell

you something you just don't need.

122

:

Should you raise it at the board meeting?

123

:

And seek the board's

advice on what to do next.

124

:

Should you organize a planning

session with your board?

125

:

If you are the owner of a small

business, should you consult your

126

:

peers, ask a similar business like

yours, what they are doing, should

127

:

you raise it at the next networking

meeting of your industry association?

128

:

Should you just launch that pilot, you've

contemplated or implement the system.

129

:

The vendors have been

forcing you to trial.

130

:

You know you need to do something.

131

:

Your board management and staff are

looking to you to take the lead.

132

:

Ensuring your company is not

being left behind in the AI race.

133

:

You're expected to act

fast, deploy something, show

134

:

results, but you're worried.

135

:

You're unsure.

136

:

In the previous episode, I've already

highlighted evidence by a reputable

137

:

research, which shows that more than

70% of digital transformations fail.

138

:

Is AI transformation going to be any

different if you launch any initiative?

139

:

What if it fails?

140

:

Honestly, I don't wanna be the bearer of

bad news, but we have got to face reality.

141

:

Let me get some facts on whether AI

initiatives are succeeding or failure.

142

:

What are the lessons we

learned from failure of others?

143

:

If you wanna make sure you don't

end up making the same mistakes you

144

:

wanna be forearmed and forewarned.

145

:

Now, research highlights, very

high failure rates for ai, just as

146

:

they do for digital transformation.

147

:

Gartner's research highlight

that 85% of AI projects fail.

148

:

Many other sources report

between 70 and 80% failure rates.

149

:

Such high failure rates will give

you cause for concern and rightly so.

150

:

What if you have to abandon the

AI program a few months into it?

151

:

It's going to waste money, but it's also

going to take up time and the attention

152

:

of management and employees that could

have been better used on other strategic

153

:

projects that drive the company forward.

154

:

If it fails, your reputation is

on the line, and maybe your CEO or

155

:

executive job may also be at risk.

156

:

We talk about high failure

rates of AI programs and also

157

:

of digital transformation.

158

:

But what does this

actually mean in practice?

159

:

So let's try to define what failure means.

160

:

Does it mean a complete loss?

161

:

Is there still some benefit

of AI initiatives, in which

162

:

case it's worth pursuing?

163

:

Let's dive into what research defines

as failure, and more importantly,

164

:

how to spot the signs of failure.

165

:

So let's look at six different

definitions of failure ranging

166

:

on a scale from complete loss to

systems we are deployed, but they

167

:

don't really produce the benefits.

168

:

So number one, we have complete

abandonment of the AI project.

169

:

In this case, the AI project is halted.

170

:

Resources invested are written off, and

the intended solution is never deployed.

171

:

If it's deployed, it's decommissioned.

172

:

And the telltale signs of this kind

of a failure is project termination,

173

:

reassignment of team members,

and no deployment to production.

174

:

So the second type of failure is a failure

to deliver expected business value.

175

:

Now, in this case, the AI solution

is functional and is deployed, but

176

:

it doesn't generate the anticipated

returns, improve efficiencies, or

177

:

increase revenue, or it doesn't achieve

the strategic goal it was intended to.

178

:

The telltale signs of such a

failure is really low user adoption.

179

:

Minimal impact on key business metrics,

or the bottom line high maintenance

180

:

costs, which outweigh the benefits.

181

:

So a third example of failure could be

technical failure, poor performance.

182

:

Again, in this case, the AI

solution is actually deployed.

183

:

But it faces some technical issues

such as inaccurate predictions, biases,

184

:

instability, and integration issues.

185

:

And you can often tell these with,

telltale signs or indicators, such

186

:

as high error rates, inconsistent

outputs, inability to handle

187

:

real world data variability,

security vulnerabilities, and user

188

:

dissatisfaction due to poor performance.

189

:

At number four, we have a pilot

purgatory and failure to scale.

190

:

What does pilot purgatory

mean in this case?

191

:

It it means that the pilot actually

shows promise in a controlled environment

192

:

or in a proof of concept environment,

but it fails to be successfully

193

:

deployed in business as usual.

194

:

You can tell, or this kind of a failure,

when you find that the projects, uh, or

195

:

the pilot is prolonged for a long, long

period of time, there's a difficulty

196

:

in integrating with existing systems.

197

:

The organization is just not ready

for widespread adoption of this

198

:

AI system, and there's no business

value beyond this test case.

199

:

So in the fifth case, and this is a

significant one, failure can mean ethical

200

:

failure and negative consequences.

201

:

So the system, again, is deployed

into business as usual, but it starts

202

:

generating unintended consequences such

as discrimination, privacy violation,

203

:

lack of transparency, and it erodes trust

or actually invites regulatory scrutiny.

204

:

The telltale signs of such a

failure is really biased, outputs,

205

:

regulatory, non-compliance, damage,

or loss of reputation and trust.

206

:

Maybe because customers start to complain.

207

:

And lastly, failure could mean a

failure to achieve user adoption.

208

:

So in this case, the system is

developed, implemented, and actually

209

:

proves valuable to the organization.

210

:

However, users don't adopt such

a system because of resistance.

211

:

Maybe they lack trust.

212

:

Maybe they have insufficient training

and or poor user experience, and you can

213

:

tell this kind of a failure often by low

usage rates, negative feedback from users.

214

:

There's also workarounds being

developed by employees or customers

215

:

to avoid using such a system.

216

:

Lack of integration in daily

life is a clear sign that you're

217

:

facing this kind of a failure.

218

:

These are crucial signs that you have

to watch out for and hopefully this

219

:

gives you some framework to monitor

whether your AI project when deployed

220

:

is going down the path to failure.

221

:

And at least once you understand the

telltale signs, you can do something

222

:

about it to make sure that you bring

things back on track before wasting a

223

:

huge amount of cost, time, and effort.

224

:

So let's explore some real world

case studies of AI going wrong,

225

:

and what are the consequences when

AI solutions actually go wrong.

226

:

It's always great idea to keep track of

AI initiatives failing and succeeding.

227

:

They provide valuable learning lessons.

228

:

An interesting case study popped

up in research for this podcast.

229

:

It involved Air Canada.

230

:

They developed an AI chatbot to

help customers improve customer

231

:

experience and reduce reliance on

staff, maybe also to reduce costs.

232

:

However, in February, 2024, it ended

up costing the company money, its

233

:

reputation and time in fighting

a small claims court case brought

234

:

about by one of its customers.

235

:

The chatbot gave the

customer wrong information.

236

:

About the bereavement fair policy.

237

:

So in this case, a customer, Jake Moffitt,

visited Air Canada's website to book

238

:

a flight from Vancouver to Toronto.

239

:

He needed to travel urgently to

attend his grandmother's funeral.

240

:

He suspected that the airline

provided discounted rates for

241

:

travel related to bereavement.

242

:

Not sure he opened the chatbot.

243

:

To ask if a discounted rate applied.

244

:

The chatbot presented the following text

message, which I will read out to you.

245

:

Air Canada offers reduced bereavement

fares if you need to travel

246

:

because of an imminent death or

a death in your immediate family.

247

:

If you need to travel immediately

or have already traveled and

248

:

would like to submit your ticket

for a reduced bereavement rate.

249

:

Kindly do so within 90 days of the date

your ticket was issued by completing

250

:

our ticket refund application form.

251

:

So just to be sure, Jake took

screenshots of the full AI chat bot

252

:

conversation and he proceeded to apply

for a refund after booking the flight.

253

:

Jake was surprised when his

claim for a refund was rejected.

254

:

Not once, but multiple times.

255

:

Air Canada's actual policy was

not to issue refunds after the

256

:

ticket had already been booked.

257

:

Remember, Jake applied for a

refund after booking the ticket.

258

:

Air Canada maintained that the

chatbot had included a link to the

259

:

actual policy and that the customer

should have read the policy document

260

:

instead of relying on the chatbot.

261

:

Naturally, Jake wasn't happy

and he launched a civil claim

262

:

in the small claims court.

263

:

The chatbot clearly did not

follow the company's policy and

264

:

provided incorrect information.

265

:

In their defense, air Canada said

that they cannot be held liable

266

:

for the information provided

by one of its agent servants or

267

:

representatives, including a chatbot.

268

:

However, in the court, air Canada

lost, they were ordered to pay

269

:

a partial refund and damages.

270

:

A small financial cost.

271

:

However, a heavy hit on their

reputation for poor customer service.

272

:

While the financial loss was small, it

highlights the fear that the company,

273

:

its staff and management, can sometimes

have little to no control over what

274

:

AI produces, and they have to bear the

consequences of large scale errors.

275

:

So let's go to a second example of Amazon.

276

:

Amazon was desperately

looking for developers.

277

:

They developed an AI recruitment

tool, which was later ditched.

278

:

So the recruitment engine was meant to

go out and search profiles on social

279

:

media, as well as accumulate cvs.

280

:

Go through those cvs.

281

:

Against the rules.

282

:

They were expected to come up with

five of the top candidates out of

283

:

every, for example, a hundred cvs.

284

:

Because of the way the AI was trained

on historical data, they created biases

285

:

against women, and therefore, they

were not referring or highlighting

286

:

developers that were female.

287

:

Amazon naturally shut this project down.

288

:

With wasted time, cost, and effort.

289

:

So let's look at the last

case study of Wells Fargo.

290

:

They created an automated calculator

to understand whether clients facing

291

:

foreclosure could afford the repayments.

292

:

After the loan was modified, for

example, the term of the loan was, uh,

293

:

increased, extended, or the interest

rates dropped just to help out the

294

:

customers, and obviously they had to

factor in the attorney costs, which also

295

:

they needed to claim from their clients.

296

:

As a result of this miscalculation around

400 customers, were denied mortgage

297

:

modifications even if they qualified,

and sadly, many lost their homes.

298

:

why do these AI initiatives fail?

299

:

I.

300

:

In the past episode, I covered the reasons

for digital transformation failures.

301

:

Let's explore the research to learn some

common reasons for AI initiatives failing.

302

:

I'll touch on what research

highlights, and there are many

303

:

research papers out there.

304

:

This topic, which I urge you to look at.

305

:

These are just some of the examples, and

I'm sure you will find many other reasons

306

:

which I don't cover in this episode.

307

:

So, firstly, as I already highlighted

before, the AI project could

308

:

completely be on the wrong path.

309

:

It could be misaligned

with business objectives.

310

:

The AI could also be technology led.

311

:

A solution looking for a

problem lacking clear goals.

312

:

When the uh AI system is adopted, perhaps

the executives or the organization

313

:

gets caught up in the hype and the

slick sales by vendors and consultants.

314

:

Data issues, poor quality of data

or lack of data is also a cause

315

:

for ai, uh, initiatives failing.

316

:

Insufficient expertise is one

of the common reasons for AI

317

:

projects failing or halting.

318

:

There's a lot of demand for AI skills and

there's a lack of supply in the market.

319

:

You also need to train up employees,

for example, with specialized

320

:

knowledge, which takes time.

321

:

Um, and that slows down the, uh, adoption

and deployment of the AI solution.

322

:

Inadequate infrastructure.

323

:

AI is very hungry for computer power,

and if your systems can't cope, you need

324

:

to upgrade that which takes you time.

325

:

Alternatively, the system that

you implement just doesn't work.

326

:

Um, there are also some issues integrating

with legacy systems for example.

327

:

Oftentimes unrealistic expectations

is the cause for ai project failures,

328

:

getting caught up in the hype and

expecting AI to solve all problems

329

:

and when it doesn't actually deliver

what you intended to deliver.

330

:

This can be partial failure,

or executives may decide to can

331

:

the whole project altogether.

332

:

Change management is something I've

covered in the past, and whilst

333

:

the system could be all flashy, all

singing, all dancing, if employees,

334

:

uh, customers, staff, partners,

don't use it, is then pretty useless.

335

:

And finally, AI is not the

panacea solution to everything.

336

:

In some cases, a simple system

could solve the problem.

337

:

And executives get caught up again in

the hype and develop an AI solution.

338

:

It's an overkill.

339

:

Um, and executives find out later on

that they could have just implemented a

340

:

potentially much cheaper solution, which

may, they may well do so in future because

341

:

it's simpler to implement, simpler to

adopt, and simpler for users to, use.

342

:

So we've heard the noise, we've seen

the failures, we've tried to understand

343

:

the causes of failures, and really

you are bombarded by the noise driving

344

:

you to do something related to AI

or risks staying behind forever.

345

:

You see the groundbreaking opportunities

of AI to shift your business.

346

:

However, you can't ignore the risks,

challenges, and most importantly,

347

:

the very high failure rates.

348

:

It's okay to feel overwhelmed,

349

:

so what are you thinking right now?

350

:

What questions come to mind?

351

:

In the quiet of your office,

in the dead of the night.

352

:

When worries of AI keep you awake.

353

:

Many questions come to mind.

354

:

Let's dive into the mind of the executive

CEO, founder or owner of the business.

355

:

Let's listen to some of the self-talk

expressing fears, uncertainties,

356

:

and confusion of what to do next.

357

:

What if I'm the only one holding

back and I miss the one opportunity

358

:

that could have changed everything.

359

:

We've been burned before by

big promises and buzzwords.

360

:

How is this any different?

361

:

I tell my board and my team we're

exploring ai, but the truth is I don't

362

:

really know what that even means.

363

:

If this fails, what do

I really stand to lose?

364

:

Not just money, but the trust of my

team, the confidence of my board, maybe

365

:

even my own belief in what I'm doing.

366

:

So you've heard the noise, you've seen

the headlines, you've felt the pressure,

367

:

and maybe just maybe you ask yourself

some of those same late night questions.

368

:

The real risk isn't just in doing

AI wrong, it's in doing it for the

369

:

wrong reasons, without clarity,

without alignment, without purpose.

370

:

That's what part two of the series is all

about because AI isn't the starting point.

371

:

Your problem is.

372

:

In the next episode, we'll

show you how to cut through the

373

:

noise and begin with clarity.

374

:

How do you find the right

problems worth solving?

375

:

And how do you know if

AI is really the answer?

376

:

Step by step?

377

:

No jargon, no hype.

378

:

It's time to replace anxiety with

alignment, panic with purpose,

379

:

and confusion with clarity.

380

:

See you in part two.

381

:

This is your host, Jay Tikam, signing off.

All Episodes Previous Episode

Listen for free

Show artwork for Impact with Digital

About the Podcast

Impact with Digital
Innovating, Adapting, and Succeeding in the Digital Age
Impact with Digital, hosted by Jay Tikam, explores the strategies and insights behind successful digital transformations in the financial services and public sectors. Each episode features expert-led discussions on how businesses can adapt to the digital age by focusing on people-first strategies, rather than just technology. From practical advice on frameworks like the SWIFTER model to in-depth case studies, Jay and his guests share real-world experiences and lessons on how to achieve sustainable change. Whether you're navigating digital transformation or seeking to understand its impact, this podcast offers valuable perspectives for leaders and decision-makers.

About your host

Profile picture for Jay Tikam

Jay Tikam

Jay Tikam is the visionary host of Impact With Digital and the CEO of Vedanvi, a strategic advisory firm that helps companies achieve high-impact digital transformations. With a global career spanning over two decades, Jay has worked with financial institutions, scale-ups, and regulatory bodies across both developed and emerging markets.

Renowned for cutting through digital hype with clarity, Jay brings a unique perspective that blends strategy, stakeholder-centric thinking, and actionable insight. He’s passionate about helping CEOs and boards move beyond buzzwords—like "digital-first mindset" or "culture change"—to make smarter, high-ROI digital decisions. Through his stakeholder-first framework and deep experience with transformation audits, he equips leaders to avoid the 70% failure rate that plagues most digital initiatives.

Whether he’s unpacking billion-dollar missteps or highlighting real-world case studies, Jay's mission is simple: empower business leaders to listen, learn, and lead meaningful transformation that creates lasting impact—for the business, its people, and society at large.