yarnnnyarnnn
← Back to blog

The SaaS Homepage Is Dying. What Comes After Chat?

·5 min read·What We're Seeing·YARNNN

At a Glance

Answer: Linear, PostHog, and Attio replaced dashboards with chat bars. That's step two. The end state is autonomous agents that do the work without being asked.

This article covers:

  • What are these companies actually admitting?
  • Where does this trajectory actually end?
  • Why is everyone building for stage two?
  • What does stage four actually require?
  • What does this mean for the products people use at work?

Linear, PostHog, and Attio all shipped the same change in the same month: the homepage is now a chat bar, not a dashboard. This isn't a design trend. It's the SaaS industry admitting that one static interface can't serve every user — and the logical end of that admission isn't chat. It's no interface at all.

What are these companies actually admitting?

The dashboard era assumed something that was never true: that every user of a product needs the same view of the same data at the same time. Dashboards are opinionated summaries designed by product teams who had to pick something to show on login.

Chat dissolves that assumption. Instead of the product deciding what matters, the user asks for what they need. It's a meaningful architectural shift — from declarative UI (here's your data, arranged our way) to imperative UI (tell us what you want to see).

Rabi Shanker Guha framed the emerging playbook concisely: expose your core APIs, connect an agentic layer, let users use software the way they want. He's right. But the playbook has more steps than people are acknowledging.

Where does this trajectory actually end?

The shift isn't dashboard-to-chat. It's a four-stage transition, and chat is stage two.

Stage 1: Dashboard. Static, opinionated, one-size-fits-all. The homepage shows what the product team thinks matters. Every user sees approximately the same thing. This is where most SaaS still lives.

Stage 2: Chat. Dynamic, user-initiated, conversational. The user asks questions and the product responds. Linear, PostHog, and Attio are here now. The user still does the asking. The system still waits.

Stage 3: Generative UI. The agent doesn't just reply in text — it composes the interface itself. Custom views, assembled in real time, based on who you are and what you're doing. This is what Guha's company Thesys is building toward. The user still initiates, but the response is richer than text.

Stage 4: Autonomous agents. No chat bar. No homepage. No waiting for the user to ask. The system observes your work context, decides what needs to happen, does it, and surfaces the result. The interface becomes the output, not the input mechanism.

Each stage removes a layer of human effort. Dashboards require interpretation. Chat requires articulation. Generative UI requires initiation. Autonomous agents require only oversight.

Why is everyone building for stage two?

Because it's the easiest step from where they already are. If you have a product with APIs, wrapping a chat interface around those APIs is a well-understood engineering problem. The LLM translates natural language to API calls. The existing product does the actual work.

Stage two also preserves the existing business model. The user still logs in. They still interact with the product. Usage is still measurable in sessions and queries. The product is still the center of the workflow.

Stages three and four break that model. If the agent composes the interface, the product becomes invisible infrastructure. If the agent acts autonomously, the user doesn't log in at all — they receive outputs. That's a harder product to sell, a harder product to meter, and a harder product to build.

But it's where the value concentrates. The reason someone opens Linear isn't to look at a dashboard or type in a chat bar. It's to know what to work on next. The closer you get to just telling them — or better, just doing it — the more value the product delivers.

What does stage four actually require?

Autonomous agents need three things that chat interfaces don't: persistent context, temporal awareness, and judgment about when to act.

A chat bar is stateless by default. Each conversation starts fresh unless the product explicitly maintains memory. That works for queries — "show me last week's conversion rate" — but it fails for recurring work. Recurring work requires the system to know what happened last time, what changed since, and what that change means for what should happen next.

Temporal awareness is the hard part. A chat bar responds when prompted. An autonomous agent has to decide, unprompted, that something is worth doing. That requires a model of time — schedules, cadences, freshness thresholds — that's fundamentally different from request-response architecture.

And judgment about when to act is what separates an autonomous agent from a cron job. The agent needs to assess whether there's enough new information to justify a run, whether the output would be meaningfully different from last time, and whether the user actually needs it right now.

What does this mean for the products people use at work?

The chat-bar wave is real and it's a genuine improvement over dashboards. But it's a waypoint, not a destination. The companies building chat interfaces today will face a choice within two years: evolve toward autonomous operation, or watch a new generation of products skip the chat phase entirely.

YARNNN is built for stage four. Persistent agents with accumulated context that run on schedule, assess whether action is needed, generate outputs that improve over time, and deliver results without being asked. No dashboard. No chat bar. The work just gets done.

The SaaS homepage is dying. But the replacement isn't a better input mechanism — it's no input at all.

Related Reading

The Speciation Is Already Happening

LLMs are monolithic — one model for every task. But AI's jagged frontier means models fail hardest in domains nobody evaluates. The future isn't better...