Skip to main content

AI Change Management: Getting Your Team to Actually Use It

You built the AI system. It works. Nobody uses it. The change management playbook for enterprise AI adoption - from resistance to routine.
10 September 2025·9 min read
Tim Hatherley-Greene
Tim Hatherley-Greene
Chief Operating Officer
The hardest part of enterprise AI isn't the technology. It's the people. A technically excellent AI system with 20% adoption delivers less value than a mediocre system with 80% adoption. Change management is the multiplier that turns AI capability into AI impact.

The Adoption Problem

We see the same pattern across every enterprise AI deployment:
Month 1: Excitement. Early adopters embrace the new tool. Usage spikes. Month 2: Reality. The tool doesn't handle every edge case. Some users hit friction. Usage plateaus. Month 3: Drift. Users who hit friction quietly return to old methods. Usage drops. Month 6: Shelfware. The tool exists. Some people use it. Most don't. Nobody talks about it.
65%
of enterprise AI tools see declining usage after the first 90 days
Source: Forrester, AI Adoption Analytics 2025
This isn't a technology problem. It's a change management problem. And it has known solutions.

Why People Resist AI (And Why They're Often Right)

Understanding resistance is the first step to addressing it. Most AI resistance falls into four categories:

1. Loss of Expertise

People who've spent years developing domain expertise feel threatened when an AI system appears to devalue that expertise. This isn't irrational. It's a legitimate concern about professional identity and value.
How to address: Position AI as amplifying expertise, not replacing it. The claims assessor with 15 years of experience doesn't lose value when AI handles routine claims, they become more valuable because they can focus on the complex cases where their expertise actually matters.

2. Quality Concerns

"I don't trust the AI's output" is the most common resistance statement we hear. And it's often justified. AI systems make mistakes, and professionals who are accountable for outcomes are right to verify.
How to address: Don't dismiss quality concerns. Validate them. Show accuracy data. Create clear escalation paths for errors. Build review workflows that let users verify before acting on AI output.
The worst thing you can do is tell a domain expert to "just trust the AI." Trust is earned through demonstrated accuracy and transparent error handling, not mandated from above.
Tim Hatherley-Greene
Chief Operating Officer

3. Workflow Disruption

Even when AI improves outcomes, it disrupts established workflows. People have systems, formal and informal, for getting work done. AI that requires a different workflow faces resistance proportional to the disruption.
How to address: Integrate AI into existing workflows rather than requiring new ones. The best AI adoption happens when users barely notice the transition. The AI appears where they already work, not in a separate tool they need to remember to open.

4. Accountability Gaps

"Who's responsible when the AI is wrong?" If this question doesn't have a clear answer, people will avoid using the system to protect themselves from accountability they didn't agree to.
How to address: Define accountability clearly before deployment. The human who reviews and acts on AI output remains accountable for the decision. The AI team is accountable for system performance. Document this. Communicate it. Reinforce it.

The Change Management Playbook

Phase 1: Co-Design (Before Deployment)

The single highest-ROI change management activity: involve users in the design.
What this looks like:
  • Identify 5-8 representative users from the target audience
  • Include them in requirements gathering and prioritisation
  • Demo prototypes early and incorporate feedback
  • Let them shape the workflow integration, not just test the output
The Co-Design Rule
If the people who will use the AI system weren't involved in designing it, adoption will be a struggle. Co-design doesn't mean design-by-committee. It means incorporating real user insight into how the AI fits their work.
Why it works: People adopt tools they helped shape. They become advocates, not resistors. And their input prevents workflow integration mistakes that would kill adoption later.

Phase 2: Champion Network (During Deployment)

Identify and enable champions: users who adopt early, use the tool effectively, and can help their colleagues.
What to look for in champions:
  • Respected by peers (not just management-endorsed)
  • Technically curious but not necessarily technical
  • Patient enough to help others
  • Honest about limitations (not just enthusiastic)
How to support them:
  • Early access and training (2-3 weeks before general availability)
  • Direct channel to the AI team for feedback and bug reports
  • Permission to spend time helping colleagues (this should be explicit)
  • Recognition for their role (but not so much that it creates resentment)
3.4×
higher adoption rates in teams with identified AI champions vs teams without
Source: Harvard Business Review, AI Adoption Study 2025

Phase 3: Structured Onboarding (First 30 Days)

Don't just give people access and hope for the best. Structure the first 30 days:
Week 1: Introduction. What the AI does, what it doesn't do, and how to get help. 30-minute session, not a half-day workshop.
Week 2: Guided practice. Users try the AI on real work with support available. Champions are the first line of help.
Week 3: Independent use. Users work independently. Track usage and identify drop-off points.
Week 4: Review. What's working? What isn't? What needs to change? Feed insights back to the AI team.

Phase 4: Continuous Improvement (Ongoing)

Adoption isn't a one-time event. It requires ongoing attention:
  • Monthly usage reviews: Who's using it? Who stopped? Why?
  • Quarterly feedback sessions: What's the AI doing well? What's frustrating?
  • Regular updates: Communicate improvements. "You told us X was frustrating. We fixed it."
  • New capability announcements: Keep the tool evolving. Stagnation kills adoption.

The Metrics That Matter

Leading Indicators (Track Weekly)

MetricTargetWhat it tells you
Daily active users>60% of eligibleAdoption breadth
Sessions per user per day>2Adoption depth - are they coming back?
Time to first action<5 minutesOnboarding effectiveness
Support ticket volumeDecreasing week over weekLearning curve trajectory

Lagging Indicators (Track Monthly)

MetricTargetWhat it tells you
Process time reduction>30% vs baselineValue delivery
User satisfaction score>7/10Trust and quality perception
Error escalation rate<10% of AI outputsAccuracy in practice
Champion-to-user ratio1:10-15Support network health
The Vanishing Middle
Watch for the vanishing middle: where power users love the tool and resistors hate it, but the 60% in the middle quietly stop using it. This middle group determines overall adoption. They need the most change management attention.

Anti-Patterns

Mandate Without Support

"Everyone must use the AI tool by March 1st." Mandates without onboarding, support, and workflow integration create resentment and compliance-only usage (people open the tool to tick a box, then do real work the old way).

Train and Abandon

A one-time training session followed by no ongoing support. Usage drops within weeks because people forget, hit edge cases, and have nobody to ask.

Ignore Legitimate Concerns

Dismissing quality concerns, workflow friction, or accountability questions as "resistance to change" instead of addressing them substantively. This guarantees adversarial adoption dynamics.

Over-Automate

Removing human oversight too quickly. Users need to build trust through verified AI output before they're comfortable with automated decisions. Trust is built gradually, not mandated.
How long does AI change management take?
Plan for 3-6 months from deployment to routine adoption. The first 30 days establish the trajectory. If usage is declining at day 30, intervene immediately. Stable adoption (>60% of eligible users, consistent daily usage) typically takes 8-12 weeks with structured change management.
Should AI adoption be mandatory?
Not at first. Start with voluntary adoption supported by champions and structured onboarding. Once adoption reaches 50-60% and the value is demonstrated, you can set expectations for the remaining users. Mandating too early creates resistance; waiting too long leaves value on the table.
What's the budget for AI change management?
Budget 15-25% of your AI project cost for change management. This covers co-design workshops, champion enablement, onboarding materials, and ongoing support. Most enterprises budget 0% and then wonder why adoption is low.