Field notes from the operation.
Working papers on Transfer of Experience and AI agents — shipped by teams running agents in production.
AI agent approval gates: where humans stay in the loop
A practical guide to human approval gates for AI agents that touch clients, sales, finance, or firm memory.
The fastest way to lose trust in an AI agent is to let it act before the organization knows where the approval gates are.
Human-in-the-loop should not be a vague reassurance. It should be designed into the workflow: what the agent can do alone, what it can draft, what it must escalate, and what it can never do without a named human.
Start with the action, not the model
Approval gates depend on the action.
A meeting-summary agent may safely draft an internal summary. A business-development agent should not send a sensitive follow-up without review. A finance agent can flag a late invoice, but payment instructions need a stricter gate. A knowledge agent can suggest a reusable paragraph, but final client advice needs human judgment.
The same model can be safe or unsafe depending on the workflow around it.
Four practical gates
Most early agents need four gates.
Draft gate
The agent can prepare output, but the output is clearly a draft. This works for follow-up emails, meeting notes, status updates, proposal fragments, and research briefs.
Approval gate
A named human approves before the action leaves the firm or changes a source system. This is the default for client-facing output.
Escalation gate
The agent must stop when it sees uncertainty, missing context, sensitive topics, or low confidence. It should say why it stopped.
Audit gate
A second process samples output and checks quality. For client-affecting agents, this should not be optional. The evaluator can be another agent, but a human owner needs to review the rubric and exceptions.
What the agent needs to show
Approval is easier when the agent exposes its reasoning trail.
For each draft or recommendation, show:
- source material used
- key assumptions
- confidence level
- missing context
- suggested action
- risk flags
- owner of approval
This is not about making the page look sophisticated. It is about letting a busy operator approve quickly without guessing.
Where firms get it wrong
The common mistake is treating human-in-the-loop as a button at the end. The agent does a bunch of unclear work, then asks for approval. The human has to reverse-engineer whether the draft is safe.
A better pattern is to keep the approval context beside the draft. The reviewer should know what changed, what source was used, and what the agent is asking permission to do.
Approval gates by role
For Bob, the business-development agent, the gate sits before outreach. Bob can surface opportunities and draft messages. A human approves the relationship move.
For EVA, the operating rhythm agent, the gate sits before delegation changes or external commitments. EVA can draft tasks and status updates. A human confirms ownership and promises.
For INO, the knowledge agent, the gate sits before firm memory becomes reusable guidance. INO can propose distilled knowledge. A senior human validates the take.
For Mo-Ni, the finance agent, the gate sits before payment-sensitive communication or data changes. Mo-Ni can remind, reconcile, and flag. A human approves anything that affects money movement.
The default rule
Draft autonomously. Ship with permission.
That rule is boring. It is also what lets teams adopt agents without feeling like the floor has disappeared.
If you are designing the first managed agent, write the approval contract before writing the prompt. Then test the workflow through the AI Jungle Assessment before giving it access to sensitive systems.