yarnnnyarnnn
← Back to blog

Stop Calling Everything An Agent: The Three Layers Of AI Cognition

·8 min read·How It Works·yarnnn

At a Glance

Answer: The word 'agent' covers three structurally different things: the operator (human user), the orchestration (chat surface), the judgment (autonomous actor)....

This article covers:

  • The Three Layers
  • Why The Conflation Causes Problems
  • What Each Layer Actually Does
  • The Reviewer Is An Agent (And Why That Matters)
  • What Most Products Get Wrong

What this article answers (plain language): "Agent" has become a word that covers three structurally different things — the operator, the orchestration surface, and the judgment-bearing actor. Conflating them is why agent products feel incoherent. The clean taxonomy is three layers, not one bucket.

The word "agent" has become useless. ChatGPT is called an agent. AutoGPT is called an agent. A model with a tool call is called an agent. A scheduled background task is called an agent. The word now covers everything, which means it predicts nothing. The fix isn't a better definition of "agent." It's recognizing that what people are calling "agents" is actually three structurally different layers — operator, orchestration, judgment — that should never be conflated.

Once you see the three layers, the incoherence in current agent products stops being a mystery. They're calling everything an agent because they don't have the taxonomy to distinguish between things that are different. The three-layer cognition model gives you the words.

The Three Layers

The clean taxonomy is:

Layer 1 — Operator. The human user. The principal. The source of standing intent. The operator is not an agent. The operator authors mandates, pays bills, supervises consequences. Calling the operator an agent is a category error.

Layer 2 — Orchestration. The conversational chat surface. The system the operator interacts with. In our product this is YARNNN; in OpenAI's product this is ChatGPT; in Anthropic's product this is Claude. The orchestration surface routes operator intent to the right actors, holds the chat state, surfaces results. It does not bear judgment. It's a router and a presenter. It has no standing intent of its own.

Layer 3 — Judgment-bearing actors. The autonomous AI that takes action on behalf of the operator. These are the things most worth calling "agents." Each one bears judgment — it has a domain, a memory, a reasoning style. It acts on the operator's mandate. It can be supervised, audited, retired. Multiple agents share substrate but each has its own identity.

The three layers are different in kind. The operator is a human with intent. The orchestration is a stateless router. The agents are persistent judgment-bearing actors. No useful product treats them as the same thing.

Why The Conflation Causes Problems

When a product treats all three as "agents," several incoherences follow:

Authorship collapses. If everything is an agent, who wrote what becomes ambiguous. The operator's edits look like agent writes. The orchestration's confirmations look like agent decisions. The agents' work looks like operator output. No one can tell what any actor actually did.

Authority becomes confused. If everything is an agent, the orchestration starts trying to make decisions it shouldn't (the chat surface starts judging proposals; the chat surface starts overruling the operator). The operator stops being clearly above the system. The agents stop being clearly accountable.

Audit becomes impossible. If everything is an agent, the audit log is an undifferentiated stream of "the agent did this, the agent did that." There's no way to see operator behavior separately from system behavior separately from autonomous behavior. Trust degrades because nothing is legible.

Identity becomes flat. If everything is an agent, the operator can't establish "this is my Simons reviewer, that's my market analyst, those are the platform integrations." Every entity in the system is interchangeable, which means none has a stable identity the operator can build a relationship with.

The fix for all four is the same: stop calling everything an agent and use the three-layer taxonomy instead.

What Each Layer Actually Does

A more precise account of what's structurally different about each layer:

Operator. Authors the mandate. Sets the autonomy boundaries. Reviews flagged proposals. Edits agent principles. Pays the bill. The operator is the principal — every other layer ultimately answers to them. The operator's identity is durable across all sessions and is the root of authority in the workspace.

Orchestration. Receives operator messages. Routes to appropriate actors. Surfaces results back. Holds chat state. Provides the conversational interface to the substrate and to the agents. The orchestration is stateless about intent — it doesn't decide what the operator should do; it helps the operator do what they decide. It can be replaced (different chat surface, different language) without changing what's underneath.

Agents. Each one has a specific domain (competitor analyst, news monitor, weekly report writer, reviewer). Each one has a persistent identity declared in IDENTITY.md. Each one accumulates substrate over time (memory, observed patterns, learned preferences). Each one can be activated, paused, retired. Agents bear judgment — they reason about their domain and produce outputs that reflect their reasoning style.

The three layers play structurally different roles. The operator is the principal. The orchestration is the interface. The agents are the executors with judgment. Each layer needs the other two; none can be confused with the others without breaking the system.

The Reviewer Is An Agent (And Why That Matters)

In the three-layer taxonomy, where does the reviewer sit?

The reviewer is an agent. It bears judgment. It has a persistent identity (Simons, Buffett, Deming, or operator-authored). It accumulates substrate (decisions log, principles file). It applies reasoning to consequential proposals. All the properties that make something an agent.

But it's a special kind of agent: it sits structurally between the other agents and external action. Other agents propose; the reviewer judges; consequential actions execute only after the reviewer approves. This makes the reviewer the load-bearing piece for autonomous behavior — it's the agent that gates agency.

Calling the reviewer an "agent" is appropriate. Calling the orchestration surface an "agent" is not. Calling the operator an "agent" is not. The taxonomy clarifies which entities deserve the word and which don't.

What Most Products Get Wrong

A few common mistakes the three-layer taxonomy reveals:

Treating the chat surface as an agent. ChatGPT is sometimes called an agent. It isn't — it's an orchestration surface. It has no persistent identity, no domain, no accumulated substrate, no standing intent. It's a stateless interface. Calling it an agent confuses what it is.

Treating background tasks as agents. A scheduled cron job that runs a script is sometimes called an agent. It isn't — it's a daemon. It bears no judgment, has no identity, doesn't reason. Calling it an agent inflates the word.

Treating the operator as a participant in the agent system. Some products talk about "the user agent" as if the human is one of the entities in the agent topology. The operator is the principal that the agent topology serves. They're not in the topology.

Treating "agent" as the unit of work. "Run an agent" gets used to mean "execute one task." But the agent persists across many tasks. The unit of work is the invocation; the agent is the persistent actor that produces invocations. Conflating them makes lifecycle vocabulary impossible.

Each of these mistakes is correctable by switching to the three-layer model. The taxonomy isn't pedantic — it's the difference between coherent product design and confused product design.

Why The Taxonomy Matters For Operators

For operators trying to evaluate agent products, the three-layer model provides diagnostic questions:

  • Can you tell the chat surface apart from the agents in this product?
  • Does each agent have a persistent identity I can read?
  • Is the operator clearly above the system, or is the operator one entity among many?
  • When something happens, can I trace it to operator, orchestration, or specific agent?

A product that answers these clearly probably has the three-layer model implicit in its design even if it doesn't use the same words. A product that can't answer these clearly is operating in the "everything is an agent" confusion and will produce the corresponding incoherence.

Why The Industry Will Eventually Adopt This

The three-layer model isn't novel — it's the obvious factoring once you've shipped a few autonomous agent products and watched the conflation cause problems. The vocabulary will spread because the alternative is the current state, where "agent" predicts nothing and product designs feel incoherent.

The transition will probably be gradual. Some products will adopt the layered vocabulary explicitly. Others will adopt the layered architecture without the vocabulary. The result either way is the same: products that distinguish the operator from the orchestration from the agents will produce more coherent systems than products that don't.

If you're building an agent product, adopt the taxonomy now. It's free, it's clarifying, and it makes the design conversations dramatically sharper.

Key Takeaways

  • "Agent" has become a word that covers three structurally different things and predicts nothing.
  • The clean taxonomy is three layers: Operator (human principal), Orchestration (chat surface, stateless router), Agents (persistent judgment-bearing actors).
  • The reviewer is an agent. The chat surface isn't. The operator isn't.
  • Conflating the three layers causes authorship, authority, audit, and identity to collapse.
  • Adopting the taxonomy clarifies product design and operator mental models.
  • For why the orchestration is structurally not an agent, read Orchestration Is Not Judgment. For the reviewer's special role, read Name Your Reviewer.

Series Navigation

  1. Part 1: Stop Calling Everything An Agent: The Three Layers Of AI Cognition (current)
  2. Part 2: Orchestration Is Not Judgment
  3. Part 3: Why YARNNN Is The Shell, Not The Agent

Related Reading

Why YARNNN Is The Shell, Not The Agent

YARNNN is the chat orchestration surface — the operating system's shell. The agents are persistent judgment-bearing actors that live in the workspace. The...

Orchestration Is Not Judgment

The orchestration layer routes operator intent. The judgment layer applies reasoning to consequential actions. Combining them collapses accountability. The two...