The Approval Click

The manager had 42 seconds.

An internal mobility agent had produced a recommendation for a senior analyst role. It summarized the employee’s last two performance reviews, pulled a skills profile from the HCM system, compared the profile with the role architecture, and marked the employee as a strong match. The workflow asked the manager to approve the move or send it back.

The manager was not lazy. She knew the employee. She also had nine other approvals waiting, a team budget call in seven minutes, and no easy way to inspect why the agent had discounted a recent certification that was missing from the profile. The screen showed a clean recommendation. The policy text said human review was required. The audit log would later show that a human approved the decision.

So she clicked approve.

That is the next failure mode in HR AI governance. It will not look like a fully autonomous machine firing someone in a dark room. It will look more ordinary: an AI system makes a recommendation, a person is placed in the loop, and the organization treats the person’s click as evidence that judgment occurred.

Sometimes judgment did occur.

Often it did not.

Call this human override theater: the use of a human approval step to make an AI-assisted employment decision look controlled, even when the human reviewer lacks the evidence, time, authority, independence, training, or interface needed to challenge the system.

It is not a small compliance detail. It is becoming a product test, a legal exposure, a manager capacity problem, and a buying criterion for HR technology. The old promise was simple: keep a human in the loop and the AI risk becomes manageable. The new question is harsher: what if the loop is real, but the human in it cannot do real work?

Why This Became Urgent in 2026

HR is moving AI from administrative help into decisions that affect people’s working lives.

SHRM’s 2026 State of AI in HR report says 46% of organizations expect to use AI in HR in 2026. SHRM also cites its 2026 CHRO research showing that 92% of CHROs expect AI to be further integrated into the workforce this year, and 87% expect greater adoption of AI within HR processes. AI’s organizational impact, SHRM reports, is 5.7 times more likely to shift job responsibilities and three times more likely to create new roles than to displace jobs.

That is the demand side.

The control side is weaker. In the same SHRM report, HR professionals said legal and compliance functions primarily lead AI governance and oversight in 37% of organizations. SHRM also found that 56% of HR professionals do not formally measure AI investment success. In states with workforce-related AI regulations, 57% of HR professionals said they were not aware of those policies.

This is a bad combination: more AI in HR workflows, more legally sensitive decisions, and uneven ownership of the rules.

The enterprise-wide signal is even sharper. Grant Thornton’s 2026 AI Impact Survey, based on 950 C-suite and senior business leaders, says 78% of executives lack strong confidence that their organization could pass an independent AI governance audit within 90 days. The report frames the gap plainly: many organizations are scaling AI they cannot explain, measure, or defend. Fully integrated AI organizations were nearly four times more likely to report AI-driven revenue growth than companies still piloting, 58% versus 15%, but the dividing line was not just model capability. It was accountability infrastructure.

That phrase sounds abstract until it reaches an HR screen.

A recruiter sees a candidate ranking. A payroll leader sees a variance recommendation. A manager sees a promotion workflow. A talent partner sees a retention-risk profile. A scheduler sees an optimized shift plan. Each system can say a human remains in control. The harder question is whether the human is equipped to contest the output before it shapes a person’s opportunity, pay, time, reputation, or employment record.

Vendors are already moving closer to those moments. ADP introduced ADP Assist agents in January 2026, describing agents that can think, plan, and take action with human oversight across payroll and HR. ADP says its agents can identify payroll variances, suggest and facilitate remediation under human oversight, generate HR insights, analyze employee-level data, and initiate talent actions such as promotion through a natural-language request.

The words “under human oversight” carry the weight.

Workday’s 2025 global research found that 75% of workers were comfortable teaming with AI agents, but only 30% were comfortable being managed by one. Only 24% were comfortable with agents operating in the background without human knowledge. The same release said 82% of organizations were expanding their use of agents.

Employees are not rejecting AI. They are drawing a boundary around power.

They can accept an agent that helps. They are less comfortable with an agent that evaluates, ranks, schedules, promotes, disciplines, or quietly shapes what managers see. That boundary is where human override theater becomes dangerous. The organization claims a person remains responsible. The employee assumes a person understood the decision. The reviewer may have had less than a minute and a weak interface.

The paper trail still says “approved.”

Regulators Are Defining a Real Job

The phrase “human in the loop” used to be enough for a vendor slide.

It is no longer enough for a serious governance program.

The European Union’s AI Act is not an HR-specific law, but its treatment of high-risk AI is already reshaping employment AI procurement. The European Commission’s AI Act overview lists AI tools for employment, worker management, and access to self-employment as high-risk use cases, including CV-sorting software for recruitment. High-risk systems are subject to obligations that include risk assessment, high-quality datasets, logging for traceability, documentation, information to deployers, appropriate human oversight, robustness, cybersecurity, and accuracy. The Commission says the rules for high-risk AI come into effect in August 2026 and August 2027, with some timeline details still being discussed under simplification proposals.

Article 14 is the center of the human oversight problem. The AI Act text explorer maintained by the Future of Life Institute presents Article 14 with an August 2, 2026 entry-into-application date for the provision. It says high-risk AI systems must be designed so natural persons can effectively oversee them. It specifies that the overseer should be enabled to understand system capacities and limitations, monitor operation, detect anomalies, avoid over-reliance, interpret outputs, decide not to use the system, disregard or reverse the output, and interrupt the system through a stop procedure.

That is not a checkbox.

It is a job description.

The UK’s Information Commissioner’s Office makes the practical version even clearer. Its AI human review guidance says meaningful human review requires reviewers with appropriate knowledge, experience, authority, and independence to challenge decisions. It also calls for manageable caseloads, adequate training, documented review methods, logs of overrides and reasons, reporting to senior management, fallback options, standardized review procedures, and re-review or overturning processes.

Those details matter because they are the difference between oversight and ceremony.

If the reviewer lacks authority, the loop is symbolic. If the reviewer lacks evidence, the loop is blind. If the reviewer lacks time, the loop becomes queue processing. If the reviewer cannot override, the loop becomes notification. If the organization does not log overrides and reasons, it cannot prove that review changed anything.

Other rules add pressure from different angles. New York City’s Local Law 144 page says employers and employment agencies using automated employment decision tools must satisfy bias audit and notice requirements; the city clarified that notice must be provided 10 business days before use of an AEDT. California’s Civil Rights Department announced that employment automated-decision system regulations were approved on June 27, 2025 and set to take effect on October 1, 2025. The CRD summary says the rules clarify how existing anti-discrimination law applies to AI and automated-decision systems in employment, and require employment records, including automated-decision data, to be maintained for at least four years.

These are not identical regimes. They do not all define the same systems, duties, rights, or timelines. But they point in the same direction.

The organization must be able to show what happened.

Human review is part of that showing. A bare approval record will not be enough when the relevant question is whether the reviewer could meaningfully assess the AI output. The record must show the system version, the data, the recommendation, the uncertainty, the reviewer, the evidence shown to the reviewer, the available alternatives, the action taken, the reason for the action, and any later challenge or correction.

HR teams used to ask whether a vendor had AI. Then they asked whether the AI was explainable. The next buyer question will be more operational:

Can your workflow prove that our humans were able to disagree?

The Reviewer Is Part of the System

The weakest assumption in many AI governance programs is that adding a human automatically adds judgment.

Research says otherwise.

In March 2026, Harvard Data Science Review published a just-accepted article, Bias in the Loop: How Humans Evaluate AI-Generated Suggestions. Jacob Beck, Stephanie Eckman, Christoph Kern, and Frauke Kreuter ran an experiment with 2,784 participants. Participants reviewed corporate greenhouse gas emissions tables and checked whether AI-extracted values were accurate. The study manipulated early AI reliability, the effort required to correct an AI error, and bonus incentives.

The results should worry anyone designing HR review queues.

When flagging an AI error required participants to also type the corrected value, they made fewer corrections and accepted more incorrect suggestions. Participants who were skeptical of AI detected errors more reliably and performed better. Participants favorable toward automation more often accepted incorrect AI suggestions. The authors concluded that successful human-AI collaboration depends not only on model performance, but on who reviews AI outputs and how the review process is structured.

That is a direct warning for HR.

An AI-assisted promotion queue is not only a model problem. It is an interface problem. It is a workload problem. It is a training problem. It is a reviewer-selection problem. It is also an incentive problem: when rejecting the AI requires more work than accepting it, the system has already nudged the human.

The accountability evidence is moving in the same direction. An April 2026 arXiv paper, AI-Induced Human Responsibility in AI-Human teams, by Greg Nyilasy, Brock Bastian, Jennifer Overbeck, and Abraham Ryan Ade Putra Hito, reported four experiments with 1,801 participants in AI-assisted lending contexts. Participants attributed more responsibility to a human decision maker paired with AI than to a human paired with another human, by an average of 10 points on a 0 to 100 scale. The authors argue that people saw AI as a constrained implementer, making the human the default locus of discretionary responsibility.

That finding is not employment law. It is not an HR case. It is still highly relevant.

Companies may believe that AI spreads responsibility across the machine, the vendor, the workflow owner, and the reviewer. Employees may see something else. If a human approved the recommendation, the human becomes the place where discretion entered the decision.

This is the trap. A company adds a human review step to reduce legal and reputational risk. The design of the step makes meaningful review unlikely. The audit trail then records the human as the decision owner. If the decision is challenged, the company points to the reviewer, and the reviewer points to a system they could not inspect.

Oversight has become responsibility transfer.

The result is not just unfair to workers affected by the decision. It is unfair to the managers and HR professionals placed in the loop. They are asked to absorb accountability without receiving the tools of accountability.

That is what makes human override theater different from ordinary bad UX. Bad UX frustrates people. Override theater creates a false legal and organizational story about judgment.

Where HR Workflows Break

Human override theater is most likely where three conditions meet: the decision is high volume, the output looks plausible, and the reviewer is under time pressure.

Recruiting is the obvious first case. AI can rank, match, summarize, score, flag, and route candidates. A recruiter may technically be the final decision maker, but the order of candidates, the summary text, and the fraud or fit labels shape attention before any formal rejection. If the recruiter sees 300 candidates and the AI only surfaces 30, the approval record answers the wrong question. Meaningful review depends on whether the recruiter had a practical way to detect who was hidden, why they were hidden, and whether the criteria were lawful and job-related.

Performance management is more sensitive.

An agent can summarize peer feedback, goals, skills, manager notes, project artifacts, productivity signals, learning history, and sentiment data. It can draft the performance narrative that the manager edits. It can recommend coaching, promotion readiness, redeployment, or pay-band movement. The manager remains the reviewer, but the review starts from a machine-shaped record. If the agent misses invisible work, overweights measurable output, or inherits biased historical feedback, the manager may not know what is missing.

Payroll and compensation create a different problem. ADP’s examples show why agents are useful: variance detection, tax registration guidance, employee-level analytics, and promotion initiation can remove real administrative friction. But the stakes are concrete. Pay errors are not abstract model failures. They hit a paycheck. Compensation recommendations can affect equity. A human reviewer needs more than a green check. They need local rules, source data, exception history, confidence, and a clear escalation route.

Scheduling and frontline workforce management add volume pressure. A scheduling agent may optimize coverage, labor cost, skills, availability, and legal constraints. But real workers have child care, transportation, second jobs, health constraints, and informal agreements with managers. A human override step is meaningful only if the reviewer can see the human constraint the optimizer missed, change the result, and leave a reason that improves the system rather than punishing the exception.

The pattern is easy to see in a table.

HR workflowWhat AI can doWhere theater appearsWhat meaningful override requires
Recruiting screenRank, match, summarize, flag riskRecruiter only sees top-ranked candidatesVisibility into ranking logic, hidden candidates, criteria, and adverse impact checks
Interview evaluationSummarize notes, compare answers, suggest scoresHiring manager accepts a polished summaryOriginal evidence, scoring rubric, uncertainty, and ability to change the record
Performance reviewDraft narrative, identify strengths and gapsManager edits language but not assumptionsSource-level evidence, missing-data flags, employee response, and escalation
Promotion workflowPre-fill case, recommend readinessApproval click becomes proof of judgmentSkills, job architecture, pay equity signal, alternatives, and reason logging
Payroll varianceDetect anomalies and suggest remediationReviewer approves bulk fixesLocal rule context, variance source, audit trail, and rollback path
SchedulingOptimize shifts and coverageManager approves schedule without seeing tradeoffsConstraint visibility, worker appeal path, and exception documentation
Internal mobilityRecommend roles and learning pathsEmployee is nudged away from opportunitiesExplainable match logic, profile completeness, and human correction rights

The risky workflows are not always the most futuristic ones. A simple AI-generated summary can be more influential than a visible AI score because it feels less like a decision tool. The reviewer may trust it as administrative support while it quietly defines the evidence frame.

This is why “AI assistant” can be a misleading category in HR.

An assistant that drafts a performance review is not merely assisting text production. It is deciding what evidence becomes salient. An assistant that summarizes a candidate is not merely saving time. It is deciding what the recruiter sees first. An assistant that flags payroll variance is not merely finding anomalies. It is deciding which anomalies feel urgent.

The power sits upstream of the approval click.

The Metrics That Expose Theater

Human review needs instrumentation.

Without metrics, organizations will default to the easiest evidence: number of AI decisions reviewed, number of approvals, number of policy documents, and existence of training. Those are weak controls. They prove that a process exists, not that it works.

The stronger evidence starts with override rate.

If a human review process almost never changes AI output, two explanations are possible. The system may be extremely accurate. Or the reviewers may be rubber-stamping. Without additional metrics, no one can tell the difference.

Override rate alone is not enough. A high override rate may mean the model is poor, the policy is unclear, or reviewers are overcorrecting. A low override rate may mean the model is good, the interface discourages disagreement, or reviewers lack time. The rate matters because it forces the organization to ask better questions.

The next metrics are more revealing:

Governance metricWhat it can reveal
Review time per decisionWhether the reviewer had enough time for the decision’s risk level
Evidence-open rateWhether reviewers actually inspect source evidence before approving
Override rate by workflowWhere human review changes AI output and where it does not
Override reason distributionWhether reviewers are correcting data, policy, model, fairness, or context errors
Escalation rateWhether reviewers use a higher-level path for uncertain or sensitive cases
Batch approval rateWhether high-volume queues are being cleared without real inspection
Appeal rate after approvalWhether affected employees or candidates challenge AI-assisted outcomes
Reversal rate after appealWhether the review process missed correctable errors
Reviewer loadWhether caseload makes meaningful review plausible
Reviewer independenceWhether the reviewer is pressured by the team that benefits from the AI output

This is where HR AI governance will start to look like operational risk management.

The system should not treat every approval as success. It should treat approvals, overrides, escalations, and appeals as signals. A healthy review process creates friction in the right places. It finds mistakes. It captures reasons. It teaches the organization where the model, data, policy, workflow, or training is weak.

That requires product design.

A reviewer console should show the AI recommendation, the source evidence, the policy basis, the confidence or uncertainty, missing-data warnings, comparison cases, potential protected-class or pay-equity flags where lawful and appropriate, and the available actions. The reviewer should not have only two buttons. Approve and reject are not enough. The real options are approve, approve with modification, request more evidence, escalate, override, pause the workflow, notify an affected person, and trigger an incident review.

The system must make disagreement as easy as agreement.

That sentence may become one of the most important design principles in HR technology.

If challenging the AI requires typing a long justification, searching source systems, messaging another team, and waiting for legal approval, while accepting the AI requires one click, the organization has already designed for acceptance. It may call that human oversight. An auditor may call it weak control.

The Vendor Test Is Changing

The old HR AI demo showed speed.

The next demo will have to show control.

A serious buyer will ask a vendor to replay a decision. Not a marketing example. A real workflow. Show the candidate screen, the prompt or instruction set, the model version, the data sources, the ranking logic at the appropriate level, the summary, the reviewer view, the missing data, the available actions, the approval, the override log, the appeal path, and the retention policy.

Some vendors will resist because not every component is under their control. HR workflows span HCM, ATS, payroll, employee service, identity, data warehouses, productivity suites, assessment platforms, background checks, and scheduling tools. A vendor may provide one part of the chain and not the full record.

That is exactly why the buying surface is moving toward governance layers.

ServiceNow positions its AI Control Tower as a centralized command center to govern, manage, secure, and realize value from agents, models, and workflows. Microsoft says its Copilot Control System can identify and mitigate risk, monitor and manage agent usage, and track ROI. Workday has described an Agent System of Record to manage digital workers alongside people and money. These are not only IT conveniences. They are attempts to own the evidence layer.

HR buyers should treat that evidence layer as a product category.

The minimum bar is an AI decision evidence packet. For any sensitive employment workflow, the packet should capture:

  • The AI system, model, agent, or automation involved
  • The system version and relevant configuration
  • The data sources used and the material data excluded or unavailable
  • The policy, rubric, workflow rule, or instruction that framed the output
  • The output shown to the reviewer
  • The evidence and uncertainty shown with the output
  • The reviewer identity, role, authority, and training status
  • The reviewer action and reason
  • Any override, escalation, appeal, correction, or rollback
  • The retention period and access controls for the record

This is not glamorous software. It will not look as exciting as an agent writing a job description or answering a manager’s question in natural language. But it will decide which products survive procurement.

The vendor that cannot produce evidence will be forced to sell on trust. The vendor that can produce evidence can sell into legal, compliance, security, audit, HR operations, and the CHRO’s office at the same time.

That changes the economics. A tool that saves a recruiter 20 minutes but creates an unreviewable employment decision may become expensive. A slower tool that produces defensible evidence may win the renewal.

The Reviewer Capacity Tax

Human oversight is often described as a safeguard. It is also labor.

Every meaningful review takes time. Someone has to read the evidence, understand the recommendation, check for missing context, decide whether to accept or change the output, write a reason, and handle follow-up. If the decision is challenged, someone has to reconstruct the path.

That work does not disappear because the AI system is fast.

It moves.

This is the reviewer capacity tax: the human effort required to make AI-assisted decisions legitimate enough to use in high-stakes HR workflows. It includes review time, training time, escalation time, documentation time, appeal time, audit time, and emotional labor when the affected person wants a real explanation.

Most AI ROI cases undercount it.

A vendor may show that an agent reduces average case-handling time by 60%. An HR leader may estimate that managers can approve twice as many workflows. A finance leader may expect administrative headcount to fall. But if the organization adds meaningful review, quality sampling, appeal handling, bias monitoring, incident response, and evidence retention, the savings change.

This does not mean the AI is not worth deploying.

It means the deployment math must be honest.

Review capacity should be planned the way companies plan support capacity. How many high-risk AI-assisted decisions can one trained reviewer handle per hour? Which decisions require two-person review? Which decisions can be sampled? Which decisions need full review? What review load causes override rates to collapse? How much training is needed before a reviewer can detect model error, data gaps, policy conflicts, or automation bias? Who pays for the time?

Those questions belong in workforce planning, not only compliance.

They also belong in manager training. The manager of a human-agent team needs to know when to trust a system, when to slow it down, when to ask for more evidence, and when to refuse the workflow. That is not the same skill as being good at the underlying HR process. It is a new layer of supervisory competence.

NIST’s AI Risk Management Framework core gives a useful baseline: processes for operator and practitioner proficiency should be defined, assessed, and documented; processes for human oversight should be defined, assessed, and documented according to organizational policies. That sounds dry. It is actually the foundation of the reviewer capacity model.

If reviewers are not trained, measured, resourced, and protected, they should not be used as proof that the system is controlled.

HR Cannot Delegate the Loop Away

The ownership problem is awkward.

Legal may write the policy. IT may own the identity and access controls. Security may monitor model and data exposure. Procurement may negotiate vendor obligations. Internal audit may test the controls. The business line may own the decision. HR may own the employee relationship.

Human override theater appears when all of them can point to someone else.

The vendor says the client configures the workflow. IT says HR owns the policy. Legal says the manager made the decision. The manager says the system recommended it. HR says the business approved it. Audit says the evidence is incomplete.

This is why HR cannot treat AI governance as a committee it occasionally attends.

HR owns the contexts where employment decisions are meaningful. It knows which moments require empathy, discretion, appeal, and explanation. SHRM’s 2026 report captures that instinct: HR professionals repeatedly said AI should support rather than replace human judgment, especially in areas requiring empathy, nuanced judgment, sensitive salary conversations, final candidate evaluations, employee relations, and ethical reasoning.

But a value statement is not an operating model.

The operating model needs named owners for five layers:

LayerOwner question
PolicyWho defines which HR decisions can use AI and at what autonomy level?
EvidenceWho decides what data, explanation, and uncertainty a reviewer must see?
ReviewWho is trained, authorized, and resourced to override AI output?
AppealWho handles employee or candidate challenges after an AI-assisted decision?
ImprovementWho turns overrides, appeals, incidents, and audit findings into system changes?

If those owners are not named, the human in the loop becomes a person standing in a gap the organization refused to close.

The strongest HR teams will stop using “human in the loop” as a comfort phrase. They will classify review types. They will distinguish low-risk routing from high-risk employment decision support. They will define where sampling is acceptable and where full review is required. They will measure reviewer behavior. They will give reviewers time and authority. They will require vendors to expose evidence. They will create appeal paths that do not punish employees for challenging a machine-shaped record.

That is slower than a demo.

It is faster than a lawsuit, regulatory inquiry, failed audit, or broken employee trust.

The Last Human in the Room

The manager at the approval screen is not the whole system.

She is the visible end of it.

Behind her sits the vendor that designed the workflow, the HR team that selected the product, the IT team that connected the data, the legal team that interpreted the rules, the executive team that demanded efficiency, and the audit team that will later ask for proof. If the only evidence is that she clicked approve, the organization has not preserved judgment. It has preserved a gesture.

The next phase of HR AI will not be judged only by how many workflows it automates. It will be judged by how honestly it treats the humans it leaves inside those workflows.

A real reviewer needs a real view of the decision.

A real override needs a real path.

A real audit needs a real record.

Without those, the human in the loop is not a safeguard. It is the last human in the room being asked to carry the weight of a machine-shaped decision.


This article provides a deep analysis of human override theater in HR AI governance. Published April 26, 2026.