You Wouldn’t Hand a New Hire the Keys on Day One

Imagine you just hired a brilliant new analyst. Top of their class. Impressive resume. Clearly talented.

Would you hand them your company’s most critical client account on their first day? Give them signing authority on contracts? Let them make pricing decisions without review?

Of course not. Not because they’re incompetent — but because trust is earned through demonstrated performance. You’d start them on lower-stakes work, review their output, give feedback, and gradually expand their scope as they prove themselves.

This is how every effective manager delegates. It’s intuitive. Nobody teaches a seminar on it because it’s obvious.

And yet — this is exactly the step most organizations skip with AI.


The Two Mistakes

When organizations deploy AI, they almost always make one of two mistakes:

Mistake 1: Too much autonomy, too fast.

This is the Zillow pattern. Give the AI decision-making authority before establishing what it’s good at (and what it’s not). The AI operates confidently — because that’s what AI does — and nobody catches the drift until the losses are visible. The BCG/Harvard “jagged frontier” study showed this clearly: consultants who blindly delegated to AI on tasks outside its competence performed 23% worse than consultants working without AI.

Mistake 2: Too little autonomy, forever.

This is the more common failure and it’s less visible. The organization deploys AI but wraps it in so much process, approval, and oversight that it never delivers meaningful value. Every output gets reviewed. Every recommendation gets second-guessed. The AI becomes an expensive autocomplete that requires more work to manage than it saves.

McKinsey found that 67% of organizations using generative AI hadn’t moved past experimentation in even a single business function. Most of those are stuck in Mistake 2 — not because the technology isn’t ready, but because they have no framework for when to let go.

  MISTAKE 1: Too much autonomy, too fast MISTAKE 2: Too little autonomy, forever
Mindset “Let AI handle everything” “Review everything AI does”
Risks Zillow ($304M loss), Knight Capital ($440M), Amazon recruiting bias AI never delivers ROI, expensive autocomplete, permanent “pilot” mode
Failure mode Headline-grabbing failure Quiet, invisible failure

The middle ground: Graduated delegation based on demonstrated performance. This is where the 10% of companies getting real results from AI actually operate.


The Delegation Model

The management science on this is well-established. Hersey and Blanchard’s Situational Leadership Model — developed in 1969 and refined over decades — defines four leadership styles based on the readiness level of the person you’re delegating to:

Style Approach Applied to AI
S1 — Directing Provide specific instructions, supervise closely AI outputs are fully reviewed. Human makes all decisions.
S2 — Coaching Explain decisions, solicit input AI suggests, human decides — but the human explains reasoning back to refine the approach.
S3 — Supporting Share decision-making, facilitate AI handles routine decisions autonomously. Human handles exceptions and judgment calls.
S4 — Delegating Turn over responsibility, monitor periodically AI operates within defined boundaries. Human reviews periodically, intervenes rarely.

The critical Hersey-Blanchard insight: the right leadership style depends on the follower’s readiness, not the leader’s preference. You don’t delegate based on how much you want to automate. You delegate based on how well the AI has demonstrated it can handle specific tasks.

This means the right autonomy level isn’t one setting across your organization. It’s per-task, per-domain, per-context:

  • AI extracting data from invoices? Maybe S4 — it’s demonstrated high accuracy, let it run.
  • AI drafting customer responses? Maybe S2 — it’s good but tone matters, human reviews.
  • AI making pricing decisions? Maybe S1 — the stakes are too high and the variables too complex for current capability.
  • AI handling Tier 1 support tickets? Maybe S3 — handles routine well, escalates unusual cases.

One organization. Four different autonomy levels. Four different delegation relationships. All running simultaneously.


What This Looks Like at Scale

Accenture studied this pattern across their client base and found that companies using AI to augment human decision-making — what Daugherty and Wilson called “collaborative intelligence” — saw productivity improvements of up to 61% in certain processes. The gains came not from replacing humans or from using AI as a passive tool, but from designing workflows where each side handles what it’s best at.

Here’s what effective AI delegation looks like across a real organization:

Department Task Delegation Level
Customer Support Tier 1 (password resets, order status) AI handles, no review
  Tier 2 (billing disputes, returns) AI drafts, agent reviews
  Tier 3 (escalations, complaints) AI summarizes, human leads
Finance Invoice data extraction AI handles, spot-checked monthly
  Expense categorization AI handles, exceptions flagged
  Payment approvals AI recommends, human approves above $10K
  Financial forecasting AI generates, CFO reviews all
Sales Lead scoring AI handles, sales reviews top 20%
  Email drafts AI writes, rep reviews before send
  Pricing/discounting AI suggests, manager approves >15%
  Contract terms Human only, AI provides research
Legal Contract clause extraction AI handles, attorney spot-checks
  Risk flagging AI flags, attorney reviews all flags
  Drafting responses AI assists, attorney writes
  Strategic decisions Human only

Notice the pattern: it’s not “legal uses AI less than sales.” Every function has a mix of high-autonomy and low-autonomy tasks. The delegation level is set by the task, not the department.


The Trust Equation

Mayer, Davis, and Schoorman published one of the most cited papers in management science — “An Integrative Model of Organizational Trust” (Academy of Management Review, 1995) — defining trust as a function of three factors:

  • Ability: Can they do this specific task competently?
  • Benevolence: Are they acting in my interest?
  • Integrity: Do they behave consistently and transparently?

This model applies directly to AI delegation:

Ability — Does the AI perform this specific task well? Not in general. Not on benchmarks. On your data, in your context, with your edge cases. The BCG/Harvard study showed that AI ability is jagged — brilliant at some tasks, confidently wrong at others. You assess ability per task, not globally.

Benevolence — Is the AI optimizing for your goals? Amazon’s recruiting AI was competent at pattern matching but was optimizing for the wrong pattern (historical bias). The alignment between what you want and what the AI is actually doing requires inspection, especially early on.

Integrity — Does the AI behave consistently? Can you understand why it made a specific decision? Can you predict how it will handle a novel situation? This is where transparency and explainability matter — not as compliance requirements, but as trust-building mechanisms.

When you frame AI adoption through trust rather than technology, the delegation model becomes intuitive:

  1. Assess ability on a specific task (not general capability).
  2. Start supervised — review outputs, catch errors, calibrate expectations.
  3. Expand gradually — reduce oversight on tasks where performance is consistently strong.
  4. Keep checkpoints — periodic review even on high-autonomy tasks.
  5. Adjust in both directions — increase autonomy as trust builds, pull back when conditions change.

This is exactly how you’d manage a new team member. The only difference is that AI can operate at scale — handling thousands of decisions simultaneously across different autonomy levels.


Why Most Platforms Can’t Do This

Here’s the practical problem: most AI and automation platforms give you a switch, not a dial. Zapier runs a workflow or it doesn’t. An AI chatbot responds autonomously or you review every message. There’s no mechanism for “handle Tier 1 tickets autonomously but flag Tier 2 for review” within the same system.

Parasuraman, Sheridan, and Wickens defined 10 levels of automation back in 2000 — from fully manual to fully autonomous — in their seminal paper on human-automation interaction. Their key finding: the optimal level of automation differs by function. You might want high automation for data gathering but low automation for decision-making.

The SAE levels of driving automation (0-5) made this concept mainstream for cars. Everyone understands that Level 2 (hands on wheel) is different from Level 4 (fully autonomous in certain conditions). Nobody argues that all cars should be either fully manual or fully autonomous.

Yet that’s exactly the binary choice most AI platforms offer for business processes.

The infrastructure to match — per-task autonomy levels, approval workflows at specific thresholds, automatic escalation on anomalies, graduated expansion over time — is what’s missing from most of the market. The technology to delegate well exists. The platforms to implement delegation well are rare.


The Practical Framework

For anyone thinking about how to delegate to AI in their organization, here’s the starting point:

Step 1: List the tasks you want AI to handle. Be specific — not “customer support” but “Tier 1 password reset requests” and “billing dispute resolution” separately.

Step 2: For each task, honestly assess the AI’s current ability. Where has it demonstrated competence on your data? Where hasn’t it?

Step 3: Assign a delegation level:

  • Direct: AI suggests, human decides everything
  • Coach: AI handles routine, human reviews before action
  • Support: AI acts, human reviews after action (spot-check)
  • Delegate: AI acts autonomously, human reviews exceptions only

Step 4: Define the trigger to move between levels. What does the AI need to demonstrate before you increase autonomy? What would cause you to pull back?

Step 5: Build the review cadence. Even S4-delegated tasks need periodic human review. How often? What are you checking?

This isn’t a governance exercise. It’s a delegation exercise. The fact that it produces an audit trail, escalation paths, and documented decision criteria is a side effect of delegating well — not the goal.


This is Part 2 of a 4-part series on Working with AI. Next: The AI Dial, Not the Switch — Why AI Autonomy Shouldn’t Be Binary.


References:

  1. Hersey, P. and Blanchard, K. “Management of Organizational Behavior: Utilizing Human Resources.” Various editions, 1969-2012.
  2. Mayer, R.C., Davis, J.H., and Schoorman, F.D. “An Integrative Model of Organizational Trust.” Academy of Management Review, Vol. 20, No. 3, 1995, pp. 709-734.
  3. Dell’Acqua, F., et al. “Navigating the Jagged Technological Frontier.” Harvard Business School Working Paper 24-013, September 2023.
  4. Daugherty, P. and Wilson, H.J. “Human + Machine: Reimagining Work in the Age of AI.” Harvard Business Review Press, 2018.
  5. Parasuraman, R., Sheridan, T.B., and Wickens, C.D. “A Model for Types and Levels of Human Interaction with Automation.” IEEE Transactions on Systems, Man, and Cybernetics, 2000.
  6. McKinsey & Company. “The State of AI in 2023.” August 2023.
  7. SAE International. “J3016: Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles.”