Exoskeleton, Not Colleague

The AI governance conversation is stuck in the wrong frame. Everyone is trying to govern autonomous agents — limit what they can do, audit what they did, assign accountability when things go wrong. The assumption baked into this is that the end state is a system that acts on its own, and governance is what we bolt on to keep it safe.

That assumption is wrong about what actually works in production.

After building AI systems in financial services, the pattern that survives contact with reality isn’t autonomous agents that governance tries to contain. It’s what I’d call the exoskeleton model: micro-agents handling narrow, well-scoped tasks, with human judgment gates at every decision point that actually matters. The agent does the cognitive lifting. The human provides the proprietary context — the institutional knowledge, the relationship history, the political read — that no model has and no amount of RAG retrieval fully substitutes for.

A compliance analyst at a bank using an AI workflow to triage suspicious transaction alerts isn’t being replaced by the system. They’re wearing it. The agent surfaces patterns, pre-classifies risk signals, drafts the SAR narrative. The analyst makes the call. They go from reviewing 30 cases a day to 150, and the quality of each decision goes up because they’re spending cognitive effort on the cases that need it, not on the ones that don’t.

This isn’t the exciting vision of autonomous agents doing things while the human sleeps. It’s more interesting. The analyst’s judgment — the thing that accumulated over years and doesn’t live in any document — is now the bottleneck in a good way. The system is calibrated around extracting and amplifying that judgment, not replacing it.

Here’s the governance implication that the current frameworks miss: certify the amplification pattern, not the autonomy level.

The question for a bank deploying AI in a regulated workflow shouldn’t be “how autonomous is this system?” It should be “does this system amplify the right human judgments, at the right points, in a way that’s auditable?” Those are different questions with different answers. A highly automated workflow with well-designed human judgment gates is safer and more governable than a lightly automated workflow where humans are nominally in the loop but the gate is a rubber stamp.

Regulators trying to govern AI autonomy are chasing an architectural assumption that practitioners have already mostly abandoned. The real frontier is human amplification patterns — what they look like, how to spec them, how to verify they’re working, and how to certify that the human judgment gate is real rather than performative.

The firms that get this right will have an edge that isn’t easily copied. Not because the technology is hard, but because the connective tissue — the institutional knowledge that makes the amplification pattern valuable — is proprietary by definition.

The governance question isn’t “how much can this agent do on its own?” It’s “whose judgment is this system multiplying, and is that person actually in the loop or just in the diagram?”


Related: [Human-in-the-Loop Is an Architecture Decision](Human-in-the-Loop Is an Architecture Decision) · [AI Governance Routing vs Compliance](AI Governance Routing vs Compliance) · [The Maker-Checker Trap](The Maker-Checker Trap) · ai-assisted-credit-due-diligence · agent-authority-delegation · agentic-workflow-audit-framework