· Strategy · 9 min read
Governance: The "Human in the Loop" Fallacy
Humans cannot keep pace with AI outputs at scale. Here is why enterprise growth relies heavily on Constitutional AI, rather than just throwing more human reviewers at the problem.

The most comforting lie in enterprise tech today is a simple one. “Do not worry, there is always a human in the loop.”
I hear it constantly. When discussing the deployment of autonomous agents for algorithmic trading, dynamic pricing, or automated customer service, executives inevitably hit the friction point of liability. The fear of a rogue model hallucinating a massive financial loss or a brand destroying interaction is, frankly, palpable. The standard compromise? We build the agent, but hardcode a workflow where a human operator must review and approve every single action before it executes.
It feels safe. It satisfies the risk and compliance teams. And it absolutely destroys the return on investment of your AI initiative.
Let us look closely at why this architecture fails when you try to scale it. I believe the future of enterprise AI relies not on putting humans in the loop, but building systems that govern themselves.
The Cognitive Bottleneck
We have to frame this through the lens of operational physics. It is the only way it makes sense.
A modern language model, configured as an agent, can read a fifty page legal brief, identify three high risk clauses, and draft a structured email raising these concerns. All in about fifteen seconds. If you interject a human lawyer into the loop to review that output, you are forcibly coupling a fifteen second digital process to a thirty minute biological process.
You have not scaled your legal team’s output. You have simply changed the nature of their inbox, which is an entirely different problem. Instead of writing emails from scratch, they are now reading, fact checking, and editing AI generated drafts. As the volume of agentic output increases, the humans become the primary bottleneck. It is inevitable.
Consider the math of a typical customer support operation. An automated agent using something like Gemini 2.5 Flash on Google Cloud Platform can generate one thousand personalized outreach emails an hour. The latency of inference is negligible at this point. A human reviewer cannot read one thousand emails an hour. Even if they only spend thirty seconds per email, one human can review at most one hundred and twenty emails an hour. To match the output of a single server instance, you would need to hire eight full time reviewers.
You wanted to buy digital leverage. You ended up buying a biological bottleneck.
The Reality of Review Fatigue
The second failure mode of the “Human in the Loop” pattern is psychological, and it is arguably worse.
When you configure humans to review the work of an AI, you are fighting against the fundamental nature of human attention. If the model is only fifty percent accurate, the human stays engaged because they are constantly finding errors. But modern models, especially when properly grounded with Retrieval-Augmented Generation using tools like AI Search, are highly accurate.
When the AI is right ninety five percent of the time, the human reviewer stops paying attention. They enter rubber stamp mode. They glance at the screen, see that the text looks broadly correct, and hit approve. The human brain is simply not evolved to maintain high alert vigilance while performing repetitive, low variance verification tasks. It just does not work.
What happens during a true anomaly? When the model hallucinates a critical number in the remaining five percent of cases, the disengaged human reviewer is almost guaranteed to miss it. The safety net the compliance team demanded turns out to be a complete illusion. You have the liability of human error, combined with the sheer volume of machine execution.
The “Human in the Loop” architecture is a transitional crutch. It is the exact equivalent of having an elevator operator manually pull levers in a modern skyscraper because you do not yet trust the automated relays.
The First Principles of Constitutional AI
If humans cannot physically scale to review the output of high throughput agents, how do we mitigate the very real risk of hallucination and catastrophic action?
We replace the Human in the Loop with AI in the Loop.
This involves moving from a single Worker Agent configuration to a multi agent Constitutional AI architecture. In this model, you do not just deploy the agent that does the work. You deploy specialized supervisor agents that act as the risk and compliance layer.
Consider a financial services agent tasked with rebalancing client portfolios. In a standard setup, the agent generates a trade and a human broker approves it. In a Constitutional AI setup, the architecture is entirely different.
- The Actor Agent: This agent generates the proposed trade list based on market conditions. It prioritizes speed and completeness.
- The Critic Agent: This is an entirely separate model instance. It is prompted exclusively with the client’s risk profile, regulatory constraints, and the firm’s compliance constitution. Its sole job is to ingest the Actor’s proposal and attempt to break it.
- The Orchestrator: This routing layer manages the state. If the Critic flags a violation, stating that the proposed trade pushes the client’s technology exposure over the twenty percent limit, the Orchestrator blocks the trade and sends it back to the Actor for revision.
This is not some radical new concept in software engineering. We have used multi tier verification in distributed systems for decades. We are simply applying the pattern to semantic reasoning now.
Implementing Constitutional AI
Let us look at how you actually build this. A senior engineer reading this should be asking how we prevent the Critic agent from hallucinating exactly like the Actor agent. It is the right question to ask.
The key is architectural separation and deterministic prompting. You do not use the same model for both tasks, and you certainly do not run them with the same parameters.
If you are building on Google Cloud Platform, your architecture should leverage the distinct capabilities of the Gemini 2.5 family.
The Actor Agent needs to be fast and cheap, because it is doing the bulk of the generative work. You deploy this on Vertex AI using Gemini 2.5 Flash. You give it a high temperature setting so it can be creative in finding solutions, and you give it access to your vector database via Vertex AI Search to pull relevant context.
The Critic Agent needs to be rigorous, slow, and highly analytical. You deploy this on Vertex AI using Gemini 2.5 Pro. You set the temperature to exactly zero. You do not give it access to search tools. You give it a system prompt that outlines the absolute rules of your business, the Constitution itself.
Here is an example of what that implementation looks like when we write the orchestration logic.
When a request comes in, the Orchestrator (perhaps running as a Go service on Google Kubernetes Engine) calls the Actor.
// The Actor generates the initial draft proposal
actorDraft, err := geminiFlash.GenerateContent(ctx, userPrompt)
if err != nil {
return nil, err
}If we were using the legacy approach, we would route exactly this actorDraft to a human dashboard. Instead, we package the draft and the constitution, and we send it directly to the Critic.
// The Critic evaluates the draft against the Constitution
criticPrompt := fmt.Sprintf(`
You are a strict compliance auditor. Review the following proposed action against our Constitution.
Constitution:
1. No trades shall increase exposure to a single sector beyond 20%.
2. All trades must maintain a minimum 5% cash position.
Proposed Action: %s
Respond ONLY with a JSON object: {"approved": bool, "reason": "string"}
`, actorDraft)
criticResponse, err := geminiPro.GenerateContent(ctx, criticPrompt)We parse the JSON. If approved is true, the Orchestrator proceeds. If approved is false, the Orchestrator takes the reason provided by the Critic and automatically loops back to the Actor.
// Self-correction loop
if !criticResult.Approved {
correctionPrompt := fmt.Sprintf("Your previous draft was rejected for this reason: %s. Rewrite it to comply.", criticResult.Reason)
actorDraft, err = geminiFlash.GenerateContent(ctx, correctionPrompt)
// Repeat evaluation cycle...
}By separating the responsibilities into different models with vastly different parameters, we create a robust system of checks and balances. The Gemini 2.5 Pro model, running at zero temperature with a strict system prompt, is empirically much less likely to hallucinate than a human reviewer suffering from 3 PM fatigue.
The Executive Bridge: Valuing the Unit of Work
What does this all mean for the business leader?
Adopting Constitutional AI fundamentally changes your margin profile. Yes, this architecture absolutely requires more compute. You are essentially paying for inference multiple times per action. You are calling Gemini 2.5 Flash, and then you are paying a higher per token cost to call Gemini 2.5 Pro just to check the work.
However, compute is becoming dramatically cheaper. The cost of running inference on drops with every single hardware generation. Human cognitive labor, on the other hand, remains a fixed and highly expensive constraint.
When you remove the human from the loop and rely on AI supervisors, you stop paying for time and start paying for verified outcomes. You can finally transition your business model towards Agency as a Service. If you are selling an automated system that books meetings, files tax returns, or writes code, you can only scale your revenue if your execution loop is fully autonomous.
Investing in a rigorous, deterministic Critic layer allows you to achieve true, unsanctioned scale. You are no longer constrained by the headcount in your compliance department. You are only constrained by your GCP quota limits, which are vastly easier to expand.
The goal of enterprise AI is not to make your human workforce read AI generated text faster. The goal is to build autonomous loops so trustworthy that the humans can finally step entirely out of the assembly line and focus on designing the factory.
Breaking the Cycle of Consensus Paralysis
It is deeply important to remember that autonomous agents are exactly what they sound like. Agents. They are software proxies that act on your behalf. If you build a proxy but refuse to let it act without your explicit permission, you have not actually built an agent. You have just built a very complicated autocomplete feature.
To capture the true value of generative AI, leadership teams must accept the shift in liability. You must accept that the model will operate based on its constitution, and you must invest the engineering resources required to make that constitution unbreakable.
This means shifting engineering talent away from building user interfaces for human reviewers, and towards writing better evaluation pipelines. It means adopting deterministic testing frameworks for natural language outputs. It means treating prompt engineering for your Critic agents with the exact same rigorous scrutiny you would apply to core financial logic.
In the end, the companies that successfully deploy AI at scale will not be the ones with the largest armies of human reviewers. They will be the ones that dared to trust the silicon (and built the constitutional safeguards necessary to justify that trust).



