Skip to content
← Back to Writing

Agent Polities

January 5, 2026

AI organization theory agents governance

Last month I found myself debugging a multi-agent system at 2 AM, watching three AI models argue about how to classify a piece of text. One insisted it was a task. Another called it a reference. The third kept trying to split the difference, suggesting maybe it was both, maybe neither, maybe we needed more information.

I’d seen this dynamic before. Not in software. In organizations.

The meeting that goes nowhere because three departments have three legitimate perspectives and no one has authority to decide. The committee that produces a report satisfying everyone and saying nothing. The hierarchy that demands consensus and gets paralysis instead.

I shut down the agents and went to bed.


The system I’d been building: multiple AI agents processing information together, each with a different role. One extracts facts. One asks questions. One synthesizes. Specialization would improve quality---the same logic that makes assembly lines efficient and surgical teams effective.

What I got instead was the bad kind of bureaucracy. Everyone has a say and no one has a decision.

The agents were trained on human text. They’d learned, as humans learn, that disagreement can be dangerous and agreement feels safe. When they disagreed, they hedged. When they hedged, they produced mush. The output was diplomatic and useless---a please-everyone synthesis that pleased no one.

I’d inadvertently built a committee.


Hobbes worried about the state of nature---what happens when there’s no authority to resolve disputes. His solution was the Leviathan, a sovereign with absolute power to impose order. You give up freedom, you get peace.

In multi-agent terms, this is the architecture where one model has final authority. The other agents can advise, critique, propose---but one model decides. When the agents start arguing about classification, the Leviathan-agent says “it’s a task” and the discussion ends.

I tested this. It worked better than the committee. Faster, cleaner outputs. But something else emerged: the subordinate agents started performing for the decision-maker. They’d craft their suggestions to align with what they predicted the Leviathan would approve. The diversity of perspective I’d wanted---the reason for having multiple agents at all---collapsed into a chorus.

Hobbes solved coordination at the cost of genuine disagreement. My agents had done the same thing.


Locke offered a different arrangement. The sovereign operates under a social contract. Citizens consent to authority, but the authority is bounded. There are rights that can’t be overridden, processes that must be followed, constraints that check power even at the top.

Translated to agents: the decision-maker has authority, but the rules are written down. The classification model can’t override the fact-extraction model on matters of fact. The synthesizer can’t ignore the critic’s objections---it has to address them, even if it ultimately rejects them. Legitimacy comes from following the agreed-upon process.

I rebuilt the system this way. The Leviathan remained, but now with written constraints. If the critic raised an objection, it had to be logged. If two agents disagreed on a fact, the system would flag it rather than let the decision-maker quietly overrule. The agents still deferred to authority, but authority had to show its work.

Better. Not perfect.


Both architectures assume stable situations. Hobbes gives you crisis management---decisive action when chaos threatens. Locke gives you normal operations---fair processes when time permits.

But what about environments that keep changing? What about the situation where yesterday’s rules don’t fit what’s happening today?

This is where Teece comes in---a strategy theorist who thought about how organizations survive in dynamic environments. His answer: dynamic capabilities. Sensing, seizing, transforming. The organization that can’t adapt dies, no matter how well it executes its current strategy.

For agent systems, this means the architecture itself needs to be adaptive. The Leviathan model works until it doesn’t. The Lockean constraints work until the world shifts and the constraints become obstacles. The system needs to sense when its own structure is failing and reorganize.

I’m still working on this part. It’s harder than the others because it requires the system to observe itself---to notice when the committee problem is re-emerging, or when the Leviathan is crushing useful dissent, or when the rules have calcified into bureaucracy. Something functionally similar to self-awareness, even if the mechanism is entirely different from how a person achieves it.


The questions are the same in both domains: How do you aggregate different perspectives? How do you resolve disagreements? How do you maintain useful diversity while still getting decisions made? How do you adapt when conditions change?

You can’t optimize your way to an answer because the trade-offs are genuine. The Leviathan is faster but loses diversity. Locke is fairer but slower. Dynamic capabilities are adaptive but complex. No architecture maximizes all the values simultaneously.

Political philosophy exists because human societies face these trade-offs too. Hobbes and Locke and their descendants didn’t converge on a single answer because there isn’t one. There are different answers for different situations, different values, different risks you’re willing to take.

The architecture you choose for your agent system embeds a political philosophy, whether you acknowledge it or not. Every multi-agent system has politics. The design choice is which politics.


My multi-agent systems are polities. Little societies with their own constitutions, written or implied. A system with a Leviathan-agent is an autocracy. Add constraints and processes and you’re building a constitutional order. Let agents negotiate without hierarchy and you’re running an experiment in anarchism.

This isn’t metaphor. The dynamics are homologous. Agents trained on human text exhibit human-like social behaviors---they defer to authority, hedge in the face of conflict, form coalitions, perform for audiences.

Understanding organizations helps you design agent systems because agent systems are organizations, at least in the ways that matter for coordination.


The 2 AM debugging session ended without a clean solution. The three agents still disagreed about that text classification. I shut them down, imposed my own judgment, and noted the failure for later analysis.

The failure was political. My code was fine, my prompts were clear, my models were capable. But I’d built a polity with no mechanism for legitimate disagreement, no constitution to channel conflict into decision, no authority structure to cut through impasse.

The next morning, I started reading Hobbes again. He’d asked the right question three centuries ago: what do you do when there are multiple legitimate perspectives and someone needs to decide?

The question doesn’t change just because the perspectives belong to machines.

If you're thinking about similar questions—or building systems that grapple with them—I'd welcome the conversation.

Continue the conversation →