yarnnnyarnnn
← Back to blog

What Should An AI Cockpit Actually Show?

·8 min read·How It Works·yarnnn

At a Glance

Answer: An AI product cockpit should show the state of the operation, not the history of conversations. Four faces: mandate (standing intent), money truth (where you...

This article covers:

  • Why "Chat Plus History" Isn't A Cockpit
  • The Four Faces
  • What Each Face Has To Have
  • Why Programs Configure Faces, Not Layout
  • What This Replaces

What this article answers (plain language): An AI cockpit for autonomous operations should show four faces — mandate (the operator's standing intent), money truth (where the account stands right now), performance (how it's trending), tracking (what's pending and what's running). Most current products show chat history instead, which is not a cockpit.

Most AI products call themselves "AI workspaces" but show you a chat transcript and a history sidebar. That's not a cockpit. That's a transcript viewer. A real cockpit for autonomous AI shows the state of the operation — what the operator is trying to accomplish, where the operation actually stands right now, how it's performing over time, what's pending. Four faces, all live, all answering questions the operator actually has when they sit down at the workspace.

This is the design question that quietly shapes the operator's experience more than any other feature decision. The cockpit is what the operator sees first, what they return to between conversations, what they trust or distrust. Getting it right is more leverage than improving any individual agent's reasoning quality.

Why "Chat Plus History" Isn't A Cockpit

Open most AI products. The default view: a chat with the AI, a sidebar of past chats, maybe a settings page. This UI shape is borrowed from messaging apps (Slack, iMessage, ChatGPT) where the primary action is conversation and the primary state is "what was said."

For autonomous AI, this shape is wrong. The operator's primary question when they return isn't "what did I say last time" — it's "what's the state of my operation." Past conversation is data; current state is what the operator needs to act on.

A few specific failures of the chat-shaped UI for autonomous use cases:

It doesn't surface autonomous activity. The AI did something while the operator was asleep. Where does that show up? In a chat message they have to scroll to find. The cockpit should show "the news monitor flagged 3 things since you were last here" prominently, not buried in a transcript.

It doesn't show standing state. What's the current mandate? What's the autonomy level? What's at risk? These don't have a place in a chat-shaped UI. They live in settings pages the operator forgets about.

It doesn't show performance. How has the operation been going? Better than last week? Worse? Chat history doesn't answer this. A performance face does.

It doesn't show what's pending. What proposals is the reviewer waiting on? What's queued for execution? What needs operator decision? The chat shape buries these in unread messages.

The operator returns to the workspace and has to reconstruct the state of their operation from a chat transcript. This is a UI failure dressed up as conversational design.

The Four Faces

A real cockpit for autonomous AI shows the state of the operation through four faces, in order:

Mandate. What's the operator's standing intent? What's the autonomy posture? The operator should see this on every visit because the mandate governs everything downstream. If the mandate is empty, the cockpit should say so loudly — nothing autonomous runs without it.

Money truth. Where does the operation stand right now? For trading: account value, open positions, recent P&L. For commerce: revenue, subscribers, recent transactions. For marketing: spend, conversions, ROI. The face shows live external truth from the platforms the operation interacts with. Substrate fallback when the platform is between syncs.

Performance. How is the operation trending over time? Realized P&L over rolling windows. Win rate. Reviewer calibration metrics (approval rate, recent verdict patterns). The performance face is where the operator sees whether the operation is on track or drifting.

Tracking. What's pending? What's running? What's blocked? The proposal queue (waiting on reviewer or operator decision). The schedule of recurring activities. Recent outcomes that need attention. Tracking is the face the operator scans for "what do I need to do today."

Together these four faces answer the operator's actual return-to-workspace questions: what am I trying to do (mandate), where do I stand (money truth), how am I trending (performance), what's pending (tracking). Chat exists separately, as one tab among others, not as the primary view.

What Each Face Has To Have

The faces aren't decorative. Each has specific properties that make it operational:

Mandate face has to be live and writable through chat. The operator sees the current mandate state and can edit it through conversation with the chat surface (not through a form). The face also surfaces autonomy posture (manual review, principles-gated, full autonomous) so the operator knows what their AI can do right now without asking.

Money truth face has to be platform-live where applicable. When the broker is connected, the face shows account state from the broker, refreshed continuously. When the broker is unavailable, the face shows the substrate version (last known state) with a clear "stale" indicator. Operator should never wonder if the number on screen is current.

Performance face has to compose the substrate, not synthesize new content. The face reads _performance.md from the substrate and renders it. The substrate is the source of truth; the face is the rendering. Same data the reviewer reads when judging proposals, same data the operator sees on the cockpit.

Tracking face has to make pending decisions actionable. Each pending proposal has approve/reject affordances. Each recurring activity has pause/resume. Each blocked item has a clear "why blocked, what to do." The face isn't read-only; it's the operator's decision queue.

These properties together turn the cockpit from a status display into a place the operator actually operates from. The cockpit becomes the operation's interface, not a summary of it.

Why Programs Configure Faces, Not Layout

Different operations have different shapes. A trading operation cares about positions and P&L; a commerce operation cares about subscribers and revenue; a marketing operation cares about spend and conversions. The cockpit has to flex without becoming arbitrarily configurable.

The pattern that works: the four faces are kernel-level (every operation has them); what fills each face is program-level (the active program supplies the bindings).

Activate a trading program, and the money truth face binds to broker positions and the performance face binds to realized P&L. Activate a commerce program, and the money truth face binds to subscriber state and the performance face binds to MRR trajectory. The face structure is constant; the bindings are program-specific.

This gives operators a recognizable cockpit shape across programs (you always know where to look for "current state" and "performance trend") while letting each program supply the right specific content for its domain. Programs don't reshape the cockpit; they fill it.

What This Replaces

Adopting the four-face cockpit replaces several common UI patterns:

Replaces dashboards. A dashboard shows "all the metrics." The four-face cockpit shows the four questions the operator actually has, with the metrics organized to answer them. More focused, less visual noise.

Replaces task boards. A task board shows work items in columns. The tracking face shows what the operator actually needs to decide on, organized by urgency and type. Operators don't think in kanban columns; they think in "what's pending."

Replaces chat-as-primary-UI. Chat moves to its own tab as one channel among others. The cockpit becomes the primary view; chat becomes the conversational channel into the operation, not the operation itself.

Replaces settings pages for mandate-shaped content. The mandate, autonomy posture, principles aren't tucked into settings — they're surfaced on the mandate face because they're core operational state, not configuration.

The result is a workspace that reads like an operations dashboard, with chat available when the operator wants conversational interaction. The shape matches the operator's mental model of running an operation, not their mental model of having a conversation.

Why This Matters For The Industry

AI products that organize the UI around chat will keep producing operators who can't tell what their AI is actually doing between sessions. AI products that organize the UI around the operation's four faces will produce operators who can run autonomous operations confidently because they always know where they stand.

The shift will happen as autonomous use cases become common enough that "what was said last time" stops being the operator's primary question. The products that adopt the four-face cockpit early will have a UX that fits the use case; the products that stay chat-shaped will keep retrofitting status displays into the conversation.

If you're designing an AI product for autonomous use cases, design the cockpit before you design the chat. The cockpit is what the operator sees first and returns to most. It's the primary surface, not the secondary one.

Key Takeaways

  • "Chat plus history" is a transcript viewer, not a cockpit.
  • A real cockpit shows four faces: mandate (standing intent), money truth (live state), performance (trend), tracking (pending decisions).
  • Each face has specific operational properties — live data, actionable affordances, substrate-backed.
  • The face structure is kernel-level; what fills each face is program-specific.
  • The cockpit shape matches the operator's mental model of running an operation, not having a conversation.
  • For why this matters for autonomous AI specifically, read Mandate-Driven AI. For the schedule face specifically, read The Schedule Is Not A Calendar.

Series Navigation

  1. Part 1: What Should An AI Cockpit Actually Show? (current)
  2. Part 2: The Schedule Is Not A Calendar: Cadence Framing For Recurring AI Work

Related Reading

Money-Truth As A File, Not A Dashboard

Autonomous AI needs performance feedback to learn. Putting that feedback in a dashboard makes it visible to humans and invisible to the AI. Putting it in a file...