Episode 6
Executives under pressure from AI hype
Are you feeling overwhelmed by the relentless noise around Artificial Intelligence?
Is your board pushing for an AI strategy?
Are consultants selling fear while employees worry about job losses?
In this eye-opening episode of Impact with Digital, host Jay Tikam speaks directly to CEOs, founders, and senior leaders navigating the chaotic, high-stakes world of digital and AI transformation. This is Part 1 of a powerful two-part series that breaks through the confusion—and reveals why most AI and digital initiatives fail before they even begin.
🔍 What You’ll Learn:
- Why 70–85% of AI initiatives fail—and what that means for your business
- The six types of AI failure every executive must understand
- Real-world cautionary tales from Air Canada, Amazon, and Wells Fargo
- The internal and external pressures pushing companies into AI before they’re ready
- The silent fears executives feel—but rarely voice
💡 Why This Episode Matters:
Digital transformation isn't about being the first to adopt AI, it’s about being the first to get it right. Whether you're a corporate executive, SME owner, or tech leader, this episode brings clarity, empathy, and strategy to a world drowning in buzzwords and fear-based selling.
👉 This is your map through the chaos, so you can lead with confidence instead of confusion.
⏱️ Episode Timestamps:
- 00:00 – Feeling the Pressure: The overwhelming AI noise in the boardroom
- 01:00 – The Inner Dialogue: What executives really feel
- 02:30 – It’s Not Just You: AI anxiety is everywhere
- 03:00 – Noise from All Sides: Boards, vendors, regulators, and staff
- 06:00 – Public Confidence vs Private Doubt: Who do you trust?
- 09:00 – What If It Fails? Real fears, real risks
- 11:40 – Six Ways AI Projects Fail: From adoption issues to ethical landmines
- 16:00 – AI Gone Wrong: Case studies from major brands
- 21:30 – Why AI Initiatives Crash: The hype, the data gaps, the poor planning
- 25:00 – The Late-Night Questions: What keeps leaders awake at night
- 27:00 – The Real Risk: Doing AI for the wrong reasons
🎧 Next up in Part 2:
Jay shares a step-by-step framework to make your AI and digital strategy actionable, people-first, and impact-driven.
🔔 Subscribe & Follow:
Spotify | Apple | Google Podcasts → @ImpactWithDigital
Learn more: Vedanvi.com
💬 Let’s Keep the Conversation Going:
Tag @ImpactWithDigital and use #AIOverload #ExecutiveLeadership #DigitalTransformation
Transcript
You are an executive, and right now you're feeling the weight of the AI
2
:revolution pressing down from every side.
3
:Competitors seem miles ahead.
4
:Consultants constantly warn
you're falling behind your board.
5
:Anxiously wonders if you're doing enough
and your employees whisper nervously
6
:about losing the jobs to algorithms.
7
:It's relentless.
8
:I'm sure it's exhausting
and frankly overwhelming.
9
:I get it.
10
:So let's stop for a second.
11
:How are you really feeling
about all of this AI hype?
12
:What's running through your mind
when you're lying awake at night?
13
:Are you cautiously optimistic, quietly
anxious, or just plain confused about
14
:where to begin in this two part series?
15
:Just like episode three and four,
I am stepping into your shoes.
16
:Today we'll explore exactly what's
on your mind as an executive facing
17
:ai, your hopes, your fears, your
frustrations, and lay them bare.
18
:No jargon, no judgment, just empathy,
clarity, and real world understanding.
19
:And in the next episode, I
promise clarity and confidence.
20
:I'll provide a clear step-by-step
blueprint to confidently embrace AI
21
:without falling into the dreaded 70%.
22
:Of digital transformation failures.
23
:Stick with me.
24
:In the next two episodes, I will turn
the madness of AI into manageable,
25
:actionable steps, bringing not only
method, but genuine excitement,
26
:performance, and impact to your journey.
27
:I am your host, Jay Tikam.
28
:Welcome to episode six of Impact with
Digital, where we explored digital
29
:strategies from a human perspective.
30
:Making sense out of the noise, finding
a way to drive performance and impact.
31
:Making sure you are on the right track
32
:in this part, one of the two part series.
33
:I'll get into your mind trying
to understand the dilemma that
34
:executives face when confronted
with AI thinking they need to do
35
:something, not knowing where to start.
36
:No matter how big your business is or how
small it is, no matter what sector you
37
:are in, we all face the same pressures.
38
:I'll try my best to understand your
dilemma, and in part two, we'll
39
:provide a step-by-step approach to
dealing with AI in the most optimal
40
:way, ensuring that you get the best
return on your investment, and that
41
:the AI actually works in practice.
42
:So let's get started.
43
:I'm sure you're constantly
surrounded by AI noises.
44
:Some make perfect sense
while others seem like noise.
45
:Some messages raise important issues
while others make you concerned.
46
:Let's explore some of the AI messages
coming at you from all directions.
47
:On a daily basis from different
types of stakeholders.
48
:Of course, I use AI to
generate these voices.
49
:So your board is getting concerned
because , it seems like your
50
:company just isn't doing enough.
51
:This puts more pressure on you.
52
:You feel compelled to
start something, anything.
53
:The other board members and
I are increasingly concerned
54
:about our AI strategy.
55
:We are reading about these advancements
daily, and frankly, we need to see
56
:a clear plan for how we're going
to leverage AI to drive innovation
57
:and future-proof the business.
58
:Are we truly keeping pace?
59
:You read articles on how AI can
make your business more competitive.
60
:You wonder whether you're missing out.
61
:You wonder whether your slow response
may see competitors getting ahead of you.
62
:In the crazy world of ai.
63
:You're bombarded by consultants and
software vendors urging you to act
64
:and spend money on their solutions.
65
:They put the fear of God into you hinting
that if you don't do anything, you're just
66
:going to be left behind maybe becoming
the next Kodak in the camera market.
67
:Or Blockbuster in the Netflix market.
68
:The window of opportunity
with AI is rapidly closing.
69
:If you are not aggressively
implementing solutions now, you
70
:risk being fundamentally outcompeted
within the next 18 to 24 months.
71
:Laggards will simply become irrelevant.
72
:You're bombarded with executive level
courses from prestigious institutions.
73
:You wonder whether you
should take a course.
74
:Just to make sense of this AI stuff
or look competent in front of your
75
:board and staff, you wonder whether
you should be arranging training for
76
:your board and maybe also all staff.
77
:Maybe if they have the skills.
78
:They could be coming up with ideas
on AI rather than consultants.
79
:Or maybe you want your employees
to take the Batten and learn about
80
:AI so that they can implement
solutions in their functional areas.
81
:Staff are disillusioned with
yet another digital change
82
:program following the latest fad.
83
:They fear job losses or at
least disruption to adopt yet
84
:another new shiny tech solution.
85
:I.
86
:We've been promised all sorts
of efficiency gains with new
87
:technology before, and it
rarely pans out as advertised.
88
:This AI stuff just feels like another
buzzword that will create more work
89
:for us in the long run because we'll
have to fix buggy systems yet again.
90
:New regulation on AI emerging regularly.
91
:They promised to put guardrails on
something that has huge potential
92
:for good, but can also cause harm.
93
:They also put restrictions
on what you can and can't do.
94
:You feel compelled to follow
regulation because after all, you
95
:don't want to end up taking your
firm on the wrong side of the law.
96
:New AI principles, standards, and best
practices emerge daily, and you and
97
:your organization need to find a way
to stay on top of these standards.
98
:You see so many cases of
AI failures emerging in the
99
:press and in research papers.
100
:This can really be scary,
especially when you're just on the
101
:path to doing something with ai.
102
:So there's a lot of noise
out there in the AI world.
103
:There's a lot of hype as well.
104
:And you're not imagining
this noise, it's real.
105
:The pressure is mounting from within
and from outside on you as an executive,
106
:everyone is shouting all around you.
107
:What are you meant to do?
108
:How do you feel?
109
:With so much noise going on around you,
some of this could be good, some just
110
:hype, some serving personal interests
rather than serving your interests.
111
:As an executive owner of the business
or a founder, you have a public face.
112
:You have to show your shareholders,
investors, the market and staff that
113
:you have everything under control,
that you're on top of AI opportunities.
114
:You need to look like you are in control
and at the forefront of the AI wave.
115
:However, in the quiet of your office
or in the dead of the night when
116
:you can't sleep 'cause of all these
worries, how do you really feel?
117
:What are you thinking?
118
:What should you do next?
119
:Should you get internal
advice from your IT team?
120
:Should you get independent
external advice?
121
:Well, advisors just try and sell
you something you just don't need.
122
:Should you raise it at the board meeting?
123
:And seek the board's
advice on what to do next.
124
:Should you organize a planning
session with your board?
125
:If you are the owner of a small
business, should you consult your
126
:peers, ask a similar business like
yours, what they are doing, should
127
:you raise it at the next networking
meeting of your industry association?
128
:Should you just launch that pilot, you've
contemplated or implement the system.
129
:The vendors have been
forcing you to trial.
130
:You know you need to do something.
131
:Your board management and staff are
looking to you to take the lead.
132
:Ensuring your company is not
being left behind in the AI race.
133
:You're expected to act
fast, deploy something, show
134
:results, but you're worried.
135
:You're unsure.
136
:In the previous episode, I've already
highlighted evidence by a reputable
137
:research, which shows that more than
70% of digital transformations fail.
138
:Is AI transformation going to be any
different if you launch any initiative?
139
:What if it fails?
140
:Honestly, I don't wanna be the bearer of
bad news, but we have got to face reality.
141
:Let me get some facts on whether AI
initiatives are succeeding or failure.
142
:What are the lessons we
learned from failure of others?
143
:If you wanna make sure you don't
end up making the same mistakes you
144
:wanna be forearmed and forewarned.
145
:Now, research highlights, very
high failure rates for ai, just as
146
:they do for digital transformation.
147
:Gartner's research highlight
that 85% of AI projects fail.
148
:Many other sources report
between 70 and 80% failure rates.
149
:Such high failure rates will give
you cause for concern and rightly so.
150
:What if you have to abandon the
AI program a few months into it?
151
:It's going to waste money, but it's also
going to take up time and the attention
152
:of management and employees that could
have been better used on other strategic
153
:projects that drive the company forward.
154
:If it fails, your reputation is
on the line, and maybe your CEO or
155
:executive job may also be at risk.
156
:We talk about high failure
rates of AI programs and also
157
:of digital transformation.
158
:But what does this
actually mean in practice?
159
:So let's try to define what failure means.
160
:Does it mean a complete loss?
161
:Is there still some benefit
of AI initiatives, in which
162
:case it's worth pursuing?
163
:Let's dive into what research defines
as failure, and more importantly,
164
:how to spot the signs of failure.
165
:So let's look at six different
definitions of failure ranging
166
:on a scale from complete loss to
systems we are deployed, but they
167
:don't really produce the benefits.
168
:So number one, we have complete
abandonment of the AI project.
169
:In this case, the AI project is halted.
170
:Resources invested are written off, and
the intended solution is never deployed.
171
:If it's deployed, it's decommissioned.
172
:And the telltale signs of this kind
of a failure is project termination,
173
:reassignment of team members,
and no deployment to production.
174
:So the second type of failure is a failure
to deliver expected business value.
175
:Now, in this case, the AI solution
is functional and is deployed, but
176
:it doesn't generate the anticipated
returns, improve efficiencies, or
177
:increase revenue, or it doesn't achieve
the strategic goal it was intended to.
178
:The telltale signs of such a
failure is really low user adoption.
179
:Minimal impact on key business metrics,
or the bottom line high maintenance
180
:costs, which outweigh the benefits.
181
:So a third example of failure could be
technical failure, poor performance.
182
:Again, in this case, the AI
solution is actually deployed.
183
:But it faces some technical issues
such as inaccurate predictions, biases,
184
:instability, and integration issues.
185
:And you can often tell these with,
telltale signs or indicators, such
186
:as high error rates, inconsistent
outputs, inability to handle
187
:real world data variability,
security vulnerabilities, and user
188
:dissatisfaction due to poor performance.
189
:At number four, we have a pilot
purgatory and failure to scale.
190
:What does pilot purgatory
mean in this case?
191
:It it means that the pilot actually
shows promise in a controlled environment
192
:or in a proof of concept environment,
but it fails to be successfully
193
:deployed in business as usual.
194
:You can tell, or this kind of a failure,
when you find that the projects, uh, or
195
:the pilot is prolonged for a long, long
period of time, there's a difficulty
196
:in integrating with existing systems.
197
:The organization is just not ready
for widespread adoption of this
198
:AI system, and there's no business
value beyond this test case.
199
:So in the fifth case, and this is a
significant one, failure can mean ethical
200
:failure and negative consequences.
201
:So the system, again, is deployed
into business as usual, but it starts
202
:generating unintended consequences such
as discrimination, privacy violation,
203
:lack of transparency, and it erodes trust
or actually invites regulatory scrutiny.
204
:The telltale signs of such a
failure is really biased, outputs,
205
:regulatory, non-compliance, damage,
or loss of reputation and trust.
206
:Maybe because customers start to complain.
207
:And lastly, failure could mean a
failure to achieve user adoption.
208
:So in this case, the system is
developed, implemented, and actually
209
:proves valuable to the organization.
210
:However, users don't adopt such
a system because of resistance.
211
:Maybe they lack trust.
212
:Maybe they have insufficient training
and or poor user experience, and you can
213
:tell this kind of a failure often by low
usage rates, negative feedback from users.
214
:There's also workarounds being
developed by employees or customers
215
:to avoid using such a system.
216
:Lack of integration in daily
life is a clear sign that you're
217
:facing this kind of a failure.
218
:These are crucial signs that you have
to watch out for and hopefully this
219
:gives you some framework to monitor
whether your AI project when deployed
220
:is going down the path to failure.
221
:And at least once you understand the
telltale signs, you can do something
222
:about it to make sure that you bring
things back on track before wasting a
223
:huge amount of cost, time, and effort.
224
:So let's explore some real world
case studies of AI going wrong,
225
:and what are the consequences when
AI solutions actually go wrong.
226
:It's always great idea to keep track of
AI initiatives failing and succeeding.
227
:They provide valuable learning lessons.
228
:An interesting case study popped
up in research for this podcast.
229
:It involved Air Canada.
230
:They developed an AI chatbot to
help customers improve customer
231
:experience and reduce reliance on
staff, maybe also to reduce costs.
232
:However, in February, 2024, it ended
up costing the company money, its
233
:reputation and time in fighting
a small claims court case brought
234
:about by one of its customers.
235
:The chatbot gave the
customer wrong information.
236
:About the bereavement fair policy.
237
:So in this case, a customer, Jake Moffitt,
visited Air Canada's website to book
238
:a flight from Vancouver to Toronto.
239
:He needed to travel urgently to
attend his grandmother's funeral.
240
:He suspected that the airline
provided discounted rates for
241
:travel related to bereavement.
242
:Not sure he opened the chatbot.
243
:To ask if a discounted rate applied.
244
:The chatbot presented the following text
message, which I will read out to you.
245
:Air Canada offers reduced bereavement
fares if you need to travel
246
:because of an imminent death or
a death in your immediate family.
247
:If you need to travel immediately
or have already traveled and
248
:would like to submit your ticket
for a reduced bereavement rate.
249
:Kindly do so within 90 days of the date
your ticket was issued by completing
250
:our ticket refund application form.
251
:So just to be sure, Jake took
screenshots of the full AI chat bot
252
:conversation and he proceeded to apply
for a refund after booking the flight.
253
:Jake was surprised when his
claim for a refund was rejected.
254
:Not once, but multiple times.
255
:Air Canada's actual policy was
not to issue refunds after the
256
:ticket had already been booked.
257
:Remember, Jake applied for a
refund after booking the ticket.
258
:Air Canada maintained that the
chatbot had included a link to the
259
:actual policy and that the customer
should have read the policy document
260
:instead of relying on the chatbot.
261
:Naturally, Jake wasn't happy
and he launched a civil claim
262
:in the small claims court.
263
:The chatbot clearly did not
follow the company's policy and
264
:provided incorrect information.
265
:In their defense, air Canada said
that they cannot be held liable
266
:for the information provided
by one of its agent servants or
267
:representatives, including a chatbot.
268
:However, in the court, air Canada
lost, they were ordered to pay
269
:a partial refund and damages.
270
:A small financial cost.
271
:However, a heavy hit on their
reputation for poor customer service.
272
:While the financial loss was small, it
highlights the fear that the company,
273
:its staff and management, can sometimes
have little to no control over what
274
:AI produces, and they have to bear the
consequences of large scale errors.
275
:So let's go to a second example of Amazon.
276
:Amazon was desperately
looking for developers.
277
:They developed an AI recruitment
tool, which was later ditched.
278
:So the recruitment engine was meant to
go out and search profiles on social
279
:media, as well as accumulate cvs.
280
:Go through those cvs.
281
:Against the rules.
282
:They were expected to come up with
five of the top candidates out of
283
:every, for example, a hundred cvs.
284
:Because of the way the AI was trained
on historical data, they created biases
285
:against women, and therefore, they
were not referring or highlighting
286
:developers that were female.
287
:Amazon naturally shut this project down.
288
:With wasted time, cost, and effort.
289
:So let's look at the last
case study of Wells Fargo.
290
:They created an automated calculator
to understand whether clients facing
291
:foreclosure could afford the repayments.
292
:After the loan was modified, for
example, the term of the loan was, uh,
293
:increased, extended, or the interest
rates dropped just to help out the
294
:customers, and obviously they had to
factor in the attorney costs, which also
295
:they needed to claim from their clients.
296
:As a result of this miscalculation around
400 customers, were denied mortgage
297
:modifications even if they qualified,
and sadly, many lost their homes.
298
:why do these AI initiatives fail?
299
:I.
300
:In the past episode, I covered the reasons
for digital transformation failures.
301
:Let's explore the research to learn some
common reasons for AI initiatives failing.
302
:I'll touch on what research
highlights, and there are many
303
:research papers out there.
304
:This topic, which I urge you to look at.
305
:These are just some of the examples, and
I'm sure you will find many other reasons
306
:which I don't cover in this episode.
307
:So, firstly, as I already highlighted
before, the AI project could
308
:completely be on the wrong path.
309
:It could be misaligned
with business objectives.
310
:The AI could also be technology led.
311
:A solution looking for a
problem lacking clear goals.
312
:When the uh AI system is adopted, perhaps
the executives or the organization
313
:gets caught up in the hype and the
slick sales by vendors and consultants.
314
:Data issues, poor quality of data
or lack of data is also a cause
315
:for ai, uh, initiatives failing.
316
:Insufficient expertise is one
of the common reasons for AI
317
:projects failing or halting.
318
:There's a lot of demand for AI skills and
there's a lack of supply in the market.
319
:You also need to train up employees,
for example, with specialized
320
:knowledge, which takes time.
321
:Um, and that slows down the, uh, adoption
and deployment of the AI solution.
322
:Inadequate infrastructure.
323
:AI is very hungry for computer power,
and if your systems can't cope, you need
324
:to upgrade that which takes you time.
325
:Alternatively, the system that
you implement just doesn't work.
326
:Um, there are also some issues integrating
with legacy systems for example.
327
:Oftentimes unrealistic expectations
is the cause for ai project failures,
328
:getting caught up in the hype and
expecting AI to solve all problems
329
:and when it doesn't actually deliver
what you intended to deliver.
330
:This can be partial failure,
or executives may decide to can
331
:the whole project altogether.
332
:Change management is something I've
covered in the past, and whilst
333
:the system could be all flashy, all
singing, all dancing, if employees,
334
:uh, customers, staff, partners,
don't use it, is then pretty useless.
335
:And finally, AI is not the
panacea solution to everything.
336
:In some cases, a simple system
could solve the problem.
337
:And executives get caught up again in
the hype and develop an AI solution.
338
:It's an overkill.
339
:Um, and executives find out later on
that they could have just implemented a
340
:potentially much cheaper solution, which
may, they may well do so in future because
341
:it's simpler to implement, simpler to
adopt, and simpler for users to, use.
342
:So we've heard the noise, we've seen
the failures, we've tried to understand
343
:the causes of failures, and really
you are bombarded by the noise driving
344
:you to do something related to AI
or risks staying behind forever.
345
:You see the groundbreaking opportunities
of AI to shift your business.
346
:However, you can't ignore the risks,
challenges, and most importantly,
347
:the very high failure rates.
348
:It's okay to feel overwhelmed,
349
:so what are you thinking right now?
350
:What questions come to mind?
351
:In the quiet of your office,
in the dead of the night.
352
:When worries of AI keep you awake.
353
:Many questions come to mind.
354
:Let's dive into the mind of the executive
CEO, founder or owner of the business.
355
:Let's listen to some of the self-talk
expressing fears, uncertainties,
356
:and confusion of what to do next.
357
:What if I'm the only one holding
back and I miss the one opportunity
358
:that could have changed everything.
359
:We've been burned before by
big promises and buzzwords.
360
:How is this any different?
361
:I tell my board and my team we're
exploring ai, but the truth is I don't
362
:really know what that even means.
363
:If this fails, what do
I really stand to lose?
364
:Not just money, but the trust of my
team, the confidence of my board, maybe
365
:even my own belief in what I'm doing.
366
:So you've heard the noise, you've seen
the headlines, you've felt the pressure,
367
:and maybe just maybe you ask yourself
some of those same late night questions.
368
:The real risk isn't just in doing
AI wrong, it's in doing it for the
369
:wrong reasons, without clarity,
without alignment, without purpose.
370
:That's what part two of the series is all
about because AI isn't the starting point.
371
:Your problem is.
372
:In the next episode, we'll
show you how to cut through the
373
:noise and begin with clarity.
374
:How do you find the right
problems worth solving?
375
:And how do you know if
AI is really the answer?
376
:Step by step?
377
:No jargon, no hype.
378
:It's time to replace anxiety with
alignment, panic with purpose,
379
:and confusion with clarity.
380
:See you in part two.
381
:This is your host, Jay Tikam, signing off.