AI hasn’t broken employee training. It exposed it.
- Glenn Martin

- 3 hours ago
- 4 min read
Employee training hasn’t failed because of AI.
It’s failed because AI has exposed how fragile most training models already were.
For years, employee training has followed a familiar pattern:
Set a company-wide goal
Design learning for all roles and functions
Prioritise virtual delivery for scale
Motivate participation through campaigns and incentives
Brief managers so they can support the rollout
Encourage employees to apply learning through projects
In principle, this sounds sensible. In practice, it relies on a set of assumptions that no longer hold once AI enters real workflows.
AI does not simply add a new skill requirement. It increases decision frequency, compresses time-to-action, and raises the cost of poor judgement or a lack of critical thinking. Training models designed for stable environments are going to buckle under this type of sustained pressure.
This blog isn’t a stress test, it is my diagnosis of where the traditional training model is breaking in the AI era.
1. Company-wide training goals
The assumption: One goal can meaningfully apply across functions and roles
Goals like “build AI literacy”, “upskill everyone”, or “prepare the workforce for AI” sound aligned, but they collapse into abstraction the moment people return to their desks. They do not translate into the lived decision-making reality of employees.
A recruiter using AI to shortlist candidates is managing bias, fairness, and candidate experience. A finance manager using AI for forecasting is managing risk exposure, auditability, and accountability. Treating these as equivalent learning needs looks nice on a slide, but it’s useless in practice.
What breaks: Training optimises for consensus instead of relevance. People agree with the goal, complete the training, and change nothing.
What must change: Anchor training goals to decisions, not populations. If you cannot clearly answer the question “Which decisions must improve in the next 90 days?”, you are not ready to train.
2. Designing learning for all functions and roles
The assumption: Broad applicability increases efficiency.
The trade-off is depth. The more universal the content, the less it changes behaviour. People recognise this immediately, even if they cannot articulate it. It is why training often scores highly on feedback and barely registers in day-to-day work.
This is how organisations end up training language rather than judgement. Employees can describe concepts confidently, then falter when context becomes ambiguous or pressure increases.
What breaks: Learning becomes performative. Vocabulary improves. Behaviour does not.
What must change: Design a shared foundation, then deliberately diverge the learning pathways. One common spine, followed by role-specific learning paths. If learning cannot show up differently for different roles, it is not finished.
3. Prioritising virtual delivery
The assumption: Scalability equals effectiveness.
Virtual delivery scales access. It does not guarantee absorption, judgement, or safe application. For information transfer, it works well enough. For behaviour change, it is insufficient on its own.
Virtual-first models also reward self-starters and quietly disadvantage those who need challenge, social calibration, or real-time correction to learn well.
What breaks: Capability gaps widen while reported success improves.
What must change: Use virtual delivery for knowledge. Introduce live, social, or coached moments for application, challenge, and correction. If your entire AI training strategy can be completed alone, on demand, it will not change behaviour.
4. Motivation via campaigns, awards, and incentives
The assumption: Motivation is the main barrier to learning.
Most employees are not unmotivated. Many are uncertain, anxious, or quietly defensive about what AI means for their role. That fear, whether conscious or not, shapes engagement far more than incentives do.
Awards and campaigns increase completion, they do not tell you whether people trust the learning enough to use it when the pressure to deliver is high.
What breaks: Participation rises. Judgement does not.
What must change: Remove friction before adding motivation. If learning clearly helps people solve problems they already face, you do not need a campaign. If it does not, no incentive will save it.
5. Managers understanding the offering
The assumption: Understanding the programme enables effective guidance.
Managers and leaders rarely fail because they misunderstand what training exists, they fail because they are unsure when to intervene, when to reinforce, and when to challenge behaviour.
Knowing what is available is not the same as knowing how to use it in context.
What breaks: Managers and leaders become signposts, not multipliers. Training is something they point to, not something they actively integrate into work.
What must change: Train managers and leaders on when learning should show up in real decisions and real work. If managers and leaders cannot name the behaviour they are looking for, the reality is, they cannot coach it.
6. Hands-on projects to compound learning
The assumption: Application will happen if encouraged.
This is the strongest part of the traditional model and the most under-designed. Encouragement without protection fails. People will not experiment if mistakes are punished, time is unprotected, or outcomes are vague.
Psychological safety is not a cultural nice-to-have here. It is an operational requirement.
What breaks: Only confident employees apply learning. Everyone else reverts to safer habits.
What must change: Give explicit permission to experiment. Co-create scoped experiments. Make tolerance for imperfect outcomes visible. Learning without psychological safety is performative.
What changes in the era of AI
The traditional model assumes training is something you roll out, then support, then apply.
That sequencing no longer holds.
In an AI-enabled organisation, capability degrades quickly unless it is continuously refreshed, contextualised, and challenged. Tools evolve. Use cases shift. Risk surfaces in new places.
Capability is not proven by completion. It is visible in behaviour.
If behaviour does not change, neither does capability.
For CHROs, CPOs, and People Leaders accountable for AI training, the redesign is non-negotiable:
Measure decision quality, not attendance
Build role-specific capability, not universal coverage
Equip managers to coach judgement, not just support participation
Design specific application, not optional projects
If you do not, the danger is not that training will fail visibly. It is that it will appear successful while quietly changing very little.
AI has not broken employee training. It has removed the excuses.




Comments