Tag: change-management

  • Deploy First, Fix Later: How Poor AI Rollouts Are Engineering Burnout Into the System

    Over the past week, maddaisy has examined two dimensions of the AI-burnout nexus: the work intensification mechanisms that make employees busier rather than better, and the organisational frameworks that might address the people side of the equation. But there is a third dimension that sits upstream of both: the technology deployment itself.

    Before burnout becomes a management problem, it is often an implementation problem. And a growing body of evidence suggests that the way organisations roll out AI tools — rushed timelines, inadequate capability-building, and a persistent belief that go-live equals readiness — is baking unsustainable working conditions into the system from day one.

    The capability gap that no one budgets for

    A detailed analysis published by CIO.com this month draws on workforce upskilling data from large-scale enterprise rollouts to make a blunt assessment: systems rarely fail because the technology does not work. They fail because the organisation has not built the infrastructure to support the people using them.

    The numbers are sobering. Organisations routinely experience productivity drops of 30 to 40% within the first 90 days of a major technology go-live when workforce capability has not been adequately addressed. Support tickets triple. Workaround behaviours — offline spreadsheets, manual reconciliations, shadow systems — proliferate as employees revert to what feels safe and controllable. Post-go-live support costs run 40 to 60% over budget. ROI timelines slip by six to 12 months.

    None of this is caused by employee resistance or technological failure. It is the predictable consequence of treating capability-building as a training event rather than a systemic requirement.

    The 10% problem

    Perhaps the most striking data point concerns training transfer. Research on technology implementation training suggests that only 10 to 20% of skills learned in formal programmes translate into sustained on-the-job performance. The issue is not poor course design. It is the assumption that a three-day training session two weeks before go-live can substitute for the repeated practice, pattern recognition, and feedback loops that genuine capability requires.

    This matters directly for the burnout conversation. When employees lack the capability to operate new systems confidently, every minor issue becomes a support ticket. Every exception becomes an escalation. The cognitive load of navigating unfamiliar tools while maintaining output expectations creates exactly the kind of sustained pressure that Harvard Business Review’s recent research identified as driving work intensification. The tools are not the problem — the deployment is.

    Where AI deployments differ from traditional IT rollouts

    Enterprise technology deployments have always carried capability risks. But AI introduces specific complications that amplify them.

    First, AI tools change the scope of work, not just the process. As maddaisy’s earlier analysis noted, employees using AI do not simply do the same tasks faster — they absorb new responsibilities, blur role boundaries, and take on work that previously justified additional headcount. A traditional ERP deployment changes how someone does their job. An AI deployment can change what their job is. Upskilling programmes designed around process training cannot address a shift that is fundamentally about role redesign.

    Second, AI output is probabilistic, not deterministic. An ERP system produces the same result given the same inputs. An AI tool might produce different outputs each time, requiring users to exercise judgement about quality, accuracy, and appropriateness. That judgement cannot be trained in a classroom — it is built through experience, and it demands a kind of cognitive engagement that is qualitatively different from following a process manual.

    Third, AI raises the visibility of individual output. When an AI tool enables someone to produce a first draft in minutes rather than hours, the speed becomes the new baseline expectation. Employees who are still building capability with the tool face pressure to match output norms set by early adopters or by the tool’s theoretical capacity. The result is the self-reinforcing acceleration cycle that researchers have now documented repeatedly.

    Governance as deployment infrastructure

    The CIO.com analysis proposes treating upskilling as a governance concern rather than a training administration task — embedding capability-building into the transformation workstream with the same rigour as data migration or integration testing. This means defining capability in behavioural terms tied to business processes, starting practice in sandbox environments as soon as process designs stabilise, and measuring performance data rather than training completion rates.

    One practical recommendation stands out: establishing a “performance council” distinct from the training team. This group — composed of process owners, frontline managers, and high-performing end users — meets weekly to review whether people are performing reliably in live operations. They examine error rates, support ticket patterns, workaround behaviours, and time-to-competency. Critically, they have the authority to pause rollouts when the data shows capability is not sticking.

    This is not a radical proposal. It is the kind of operational discipline that well-run technology programmes have always applied to infrastructure and integration. The fact that it needs to be articulated separately for workforce capability suggests how often organisations skip the people side of deployment planning.

    Connecting the implementation layer to the burnout evidence

    The burnout data that has emerged over recent weeks tells a consistent story. DHR Global reports that 83% of workers experience some degree of burnout, with engagement dropping 24 percentage points in a single year. Only 34% of employees say their organisation has communicated AI’s workplace impact clearly. Among entry-level staff, that figure falls to 12%.

    These are not just management failures. They are deployment failures. When organisations ship AI tools without building the capability infrastructure to support them — without adequate training transfer, without performance monitoring, without role redesign — they are not just creating a change management problem. They are creating a technical debt of human capability that compounds over time, manifesting as burnout, disengagement, and quiet resistance.

    As Deloitte’s 2026 AI report made clear, the gap between strategic confidence and operational readiness remains the defining feature of enterprise AI. The implementation evidence suggests that closing this gap requires treating workforce capability not as a soft skill or an HR deliverable, but as core deployment infrastructure — as essential as the data pipeline and as measurable as system uptime.

    What this means for the next phase

    The organisations that avoid engineering burnout into their AI deployments will share a common trait: they will treat go-live as the beginning of capability-building, not the end. They will budget for the 90-day productivity dip and plan for it rather than being surprised by it. They will measure whether people can perform reliably at scale under real business pressure, not whether they completed a training module.

    Most importantly, they will recognise that the fastest way to undermine an AI investment is not a technical failure — it is deploying capable technology to an unprepared workforce and then measuring success by adoption rates alone. The tools work. The question, as it has always been with technology transformations, is whether the organisation around them does too.

  • The Missing Half of AI Adoption: Why Organisations Are Failing the People Side of the Equation

    The evidence that AI tools can intensify work rather than lighten it has been building steadily. As maddaisy examined earlier this week, UC Berkeley researchers found that employees using AI worked faster but not better — absorbing more tasks, blurring work-life boundaries, and entering a self-reinforcing cycle of acceleration. The diagnosis is now well-documented. The question that remains largely unanswered is what organisations should actually do about it.

    New data suggests that most are doing very little — and that the gap between leadership confidence and frontline reality is wider than many boards realise.

    The engagement crisis hiding behind adoption metrics

    A 2026 workforce trends report from DHR Global paints a stark picture. Some 83% of workers report experiencing at least some degree of burnout, with the technology sector among the worst affected at 58%. That figure has held roughly steady since 2025 — but what has changed is burnout’s impact on engagement. More than half of employees (52%) now say burnout actively reduces their engagement at work, up from 34% just a year earlier.

    Meanwhile, overall engagement has dropped sharply. Only 64% of workers describe themselves as very or extremely engaged, down from 88% in 2025. That 24-percentage-point decline should concern any organisation that has been measuring AI success purely by deployment velocity.

    The report also reveals a telling asymmetry in how AI communication is handled. Only 34% of employees say their organisation has communicated AI’s workplace impact “very clearly.” Among entry-level staff, that figure falls to just 12%. Among C-suite leaders, it rises to 69%. The people making decisions about AI adoption and the people living with its consequences occupy different informational worlds.

    Resistance is not the problem — silence is

    A common framing in boardrooms is that employees resist AI because they fear change. The reality, according to recent research published in Harvard Business Review, is more nuanced. Workers’ responses to AI map onto three core psychological needs: competence (feeling effective), autonomy (feeling in control), and relatedness (maintaining meaningful connections with colleagues).

    When AI deployment satisfies these needs — by removing drudgery and enabling more skilled work — adoption tends to follow. When it frustrates them — by making expertise feel disposable, reducing worker agency, or replacing human collaboration with human-machine interaction — the response is not always visible protest. It is quiet disengagement. Some 31% of knowledge workers admit to actively working against their company’s AI initiatives. Among Gen Z workers, that figure rises to 41%.

    Perhaps more telling: 54% of workers say they would use unapproved AI tools, and 32% hide their AI use from employers entirely. The adoption numbers that leadership teams celebrate may obscure a more complicated picture of how AI is actually being used — and resisted — on the ground.

    From diagnosis to framework

    If the problem is well-documented, the response toolkit is still emerging. Two frameworks have gained traction in the consulting and HR strategy space this year, and both share a common starting point: organisations need to treat AI adoption as a change management challenge, not a technology deployment exercise.

    The first, proposed by Ranganathan and Ye in their HBR study on work intensification, is the concept of “AI practice” — deliberate organisational norms around how AI is used. This includes structured pauses before major decisions (to counteract the speed bias that AI creates), sequencing work to reduce context-switching, and protecting time for human connection. The researchers argue that without intentional guardrails, the default trajectory is escalation: faster output, rising expectations, and eventual burnout.

    The second is the AWARE framework, developed by researchers studying psychological resistance to AI. It stands for: Acknowledge employee concerns proactively; Watch for both adaptive and maladaptive coping behaviours; Align support systems with the psychological needs that AI may be disrupting; Redesign workflows around human-AI complementarities rather than simple substitution; and Empower workers through transparency and genuine participation in implementation decisions.

    Neither framework is complicated. Both amount to structured versions of what good management has always looked like: listen to your people, involve them in decisions that affect their work, and pay attention to unintended consequences. The fact that these principles need to be formalised into acronyms and published in academic journals suggests how far many organisations have drifted from applying them during AI rollouts.

    The consulting opportunity — and obligation

    For consultancies advising on AI transformation, this data represents both a market opportunity and a professional obligation. The firms that positioned AI adoption as primarily a technical challenge — choose the right model, integrate the right data pipeline, deploy at scale — are leaving the harder, more valuable work on the table.

    The harder work is organisational. It involves job redesign, not just process automation. It requires mapping which tasks benefit from AI augmentation and which need to remain human — not as a one-off exercise, but as an ongoing practice. It means building managerial capability in areas that technology deployments rarely prioritise: emotional intelligence, coaching, and the ability to recognise cognitive overload before it becomes a retention problem.

    Only 44% of business leaders currently involve workers in AI implementation decisions. That figure alone explains much of the resistance and burnout data. People who have no say in how their work changes will inevitably feel that change is happening to them rather than with them — regardless of whether the technology itself is genuinely useful.

    What comes next

    The most thoughtful organisations are beginning to reframe AI not as a productivity multiplier but as a capacity management tool. The question shifts from “how can AI help people do more?” to “how can AI help people think more clearly, recover more effectively, and focus on what actually matters?” It is a subtle but significant distinction — one that moves the conversation from output volume to work quality and sustainability.

    As maddaisy noted in its analysis of Deloitte’s 2026 AI report, the gap between strategic confidence and operational readiness remains one of the defining features of enterprise AI. The burnout and engagement data suggests that this gap is not just a technology integration problem. It is, at its core, a people management problem — and one that will not be solved by deploying more tools faster.

    The organisations that get this right will not be the ones with the most AI models in production. They will be the ones that treated adoption as a human process from the beginning.