Workforce Planning for AI: Roles, Reskilling, and the Human–AI Team
The question enterprise leaders most commonly ask about AI and workforce is: which jobs will be automated? This is the wrong question. The right question is: how does AI change the work — and what does that mean for the people doing it? Here is how to plan your workforce transition correctly.
Key Takeaways
- Task-level analysis reveals more than job-level analysis: different tasks within the same job title are affected by AI in fundamentally different ways.
- Four workforce segments: AI-augmented workers (judgement supported by AI), AI-managed process workers (humans handling AI-structured workflow exceptions), AI-native roles (new roles that govern AI), and unaffected/expanding roles.
- The highest-return reskilling investment for most organisations is AI literacy in the augmented-worker population — critical evaluation of AI outputs, not technical AI development.
- Workforce planning for AI is a design exercise, not a headcount exercise — the question is what the new human–AI team looks like, not how many people AI replaces.
- WEF's 2023 Future of Jobs report estimates 23% of jobs globally will change significantly due to AI within five years — the majority through augmentation, not replacement.
The question enterprise leaders most commonly ask about AI and workforce is: which jobs will be automated? This is the wrong question. It is too blunt to be useful for planning and too binary to reflect how AI actually changes work in practice.
The more useful question is: how does AI change the work, and what does that mean for the people doing it? Answering this requires workforce analysis at a granular level — not by job title, but by task composition. Different tasks within the same job are affected by AI in fundamentally different ways.
The Four Workforce Segments
Framework Reference
The Four AI Workforce Segments
Task-level analysis reveals more than job-level analysis — the same job title spans multiple segments
AI-Augmented Workers
AugmentedCore judgement work supported and accelerated by AI. Human provides interpretation, accountability, and final judgement.
Examples
Legal analysis · Financial modelling · Clinical decision support · Strategic research
Key skill needed
Critical evaluation of AI outputs; directing AI effectively
AI-Managed Process Workers
RestructuredTasks restructured around AI-managed workflows. Humans handle exceptions, escalations, and contextually ambiguous cases.
Examples
Customer service tier 2 · Content moderation · Claims review · Compliance review
Key skill needed
Working within AI-structured processes; identifying when AI categorisation is wrong
AI-Native Roles
New rolesNew roles that did not exist before AI deployment. Exist to oversee, maintain, and govern AI systems.
Examples
AI system operators · Model performance monitors · HITL reviewers · AI governance staff
Key skill needed
Technical literacy in AI systems; monitoring; governance design
Unaffected or Expanding Roles
StableCore value is human presence, physical capability, or relationship. Consistently underrepresented in AI impact analyses.
Examples
Field services · Relationship management · Physical operations · Creative direction
Key skill needed
Role-specific — AI literacy for awareness; not core operational change
WEF (2023): 23% of jobs will change significantly due to AI within five years — the majority through augmentation, not replacement.
AI-Augmented Workers — Professionals whose core judgement work is supported and accelerated by AI. Legal analysis, financial modelling, strategic research, clinical decision support. The AI handles retrieval, summarisation, and pattern recognition; the human provides judgement, contextual interpretation, and accountability. These roles require workers who can critically evaluate AI outputs, direct AI systems effectively, and take responsibility for AI-assisted decisions.
AI-Managed Process Workers — Workers whose tasks have been substantially restructured around AI-managed workflows. AI handles routing, prioritisation, and initial processing; humans handle exceptions, escalations, and cases requiring contextual judgement. The role requires the ability to work within an AI-structured process rather than a human-designed one — including the ability to identify when the AI's categorisation is wrong.
AI-Native Roles — New roles that did not exist before AI deployment: AI system operators, model performance monitors, HITL reviewers, AI governance function staff. These roles exist to oversee, maintain, and govern AI systems. They require technical literacy in AI systems, though not necessarily software development expertise.
Unaffected or Expanding Roles — Roles where core value is human presence, physical capability, or relationship. These exist in every organisation and are consistently underrepresented in AI impact analyses. Understanding which roles are genuinely unaffected prevents misallocation of reskilling investment.
Reskilling Priorities
For most enterprise organisations, the reskilling investment with the highest return is building AI literacy in the augmented-worker population — the professionals who work alongside AI systems but are not expected to build or govern them.
AI literacy at this level means: understanding what AI systems can and cannot do, being able to evaluate AI outputs critically, knowing when to escalate or override an AI recommendation, and understanding the basic governance and regulatory context. The WEF's 2023 Future of Jobs report identified AI and big data literacy as the fastest-growing skill requirement across industries — and the one with the largest gap between demand and current availability.
Deeper technical AI skills — model monitoring, prompt engineering at a sophisticated level, AI system governance — are required for AI-native roles. These typically require either specialised hiring or intensive reskilling of a smaller number of technically capable individuals.
Designing the Human–AI Team
Workforce planning for AI is ultimately a design exercise, not a headcount exercise. The question is not how many people a given AI system replaces — it is what the new human–AI team looks like, what roles it contains, and how those roles interact with AI systems day to day.
The organisations that manage this transition best treat it as an operating model question: designing the new workflow first, identifying the human roles that workflow requires, then managing the transition from current to future state deliberately. The organisations that manage it worst focus on cost reduction and find themselves with AI systems that their workforce does not know how to work with — and does not trust.
Imagine Works designs AI operating models including workforce transition plans for enterprise organisations. Get in touch to discuss your workforce strategy.
Related Service
AI Strategy & Operating Model
Designing the AI strategy, vision, and operating model that aligns your entire organisation — from the boardroom to the workflow layer.
Explore this serviceMore Insights
More on AI Strategy
AI Centre of Excellence: Should You Build One — and What Should It Do?
As enterprise AI programmes mature, a recurring question emerges: who owns AI across the organisation? The AI Centre of Excellence is the most common answer — and one of the most misunderstood and misimplemented structures in enterprise AI. Here is how to get it right.
AI Portfolio Management: How to Prioritise Which Use Cases to Fund
Most enterprise AI investment is made project by project. Each decision is rational in isolation and collectively they produce a fragmented, duplicated, ungoverned portfolio that does not compound into organisational capability. Here is how to manage AI investment as a portfolio instead.
How to Assess Your Organisation's AI Maturity
Most organisations do not know where they actually are on the AI maturity curve. Without an honest assessment, investment decisions are made in the wrong order and operating models are designed for an organisation that does not yet exist. Here is how to assess AI maturity accurately.