Stratum← Stratum Journal
← Stratum Journal
OperationsFebruary 28, 20267 min read

The Fleet Problem


Every serious AI deployment eventually becomes a fleet problem. You start with one agent. The agent works. You add a second to handle a second domain — different data, same logic. Then a third for a different team. Before long you have a dozen agents operating in parallel, each with different memory requirements, different cadences, different failure modes. The single-agent mental model that carried you through the prototype phase breaks down completely.

The tooling breaks down too. Almost everything built for AI agents in the past two years was designed with a single agent, a single session, and a human in the loop. That is a fine assumption for demos and experiments. It is not a viable assumption for production deployments in organizations that have accountability requirements, regulatory exposure, and teams that don't spend their days babysitting AI processes.

The fleet problem is not a corner case. It is where every AI deployment that actually works eventually ends up. And the infrastructure built to solve it is almost entirely absent from the current market.

The Single-Agent Fallacy

Most AI tooling is built around an implicit mental model: one agent, one task, one session, one human watching the output. The success criteria are simple — did the agent produce something useful? — and the failure mode is equally simple: the human notices, intervenes, tries again.

This model is appropriate for the experimentation phase. It is how you validate that the agent is capable of the task at all. It is also the model that shapes the demos, the developer tutorials, the benchmarks, and the GitHub repos that define the category.

The problem is that organizations deploying AI in production don't operate this way. They have SLAs. They have compliance teams. They have stakeholders who need audit trails. They have infrastructure budgets that require predictable costs. They have on-call rotation and they cannot add "watch the AI agent" as a 24/7 task. The single-agent, human-in-the-loop mental model does not survive contact with an actual enterprise deployment.

"The tooling was built for the demo. The enterprise needs the 3am version — what happens when no one is watching, the agent hits an error, and the output feeds a decision at 9am."

What a Fleet Actually Is

A fleet is not just multiple agents. It is multiple agents operating continuously, with defined responsibilities, persistent state, and accountability chains that trace back to humans. The defining characteristics of fleet deployment are worth naming precisely:

Continuity. Fleet agents run on schedules, not on demand. A research monitoring agent runs every morning. A financial analysis agent runs every trading day. A compliance agent runs whenever a regulatory update is detected. The session model — invoke, receive output, close — gives way to a process model: persistent execution, accumulating state, scheduled actions.

Interdependence. Fleet agents are not isolated. They share data, hand off work, and depend on each other's outputs. An agent that monitors news feeds the agent that synthesizes market intelligence that feeds the agent that drafts the weekly briefing. Each agent in the chain has dependencies; each dependency is a potential failure point. Understanding and managing those dependencies requires coordination infrastructure that single-agent frameworks were not designed to provide.

Accountability. Every output from a fleet agent traces back to a human decision — the decision to deploy this agent, configure it this way, give it access to this data. That accountability chain must be preserved and auditable. When a fleet agent makes an error that propagates into a business decision, the organization needs to reconstruct exactly what the agent did, what data it used, and why it produced that output. This is not a nice-to-have. Under emerging AI regulations, it is becoming a compliance requirement.

Unattended operation. This is the defining characteristic. Fleet agents must operate reliably without human supervision. Not occasionally unattended — continuously. The system must detect failures, handle errors, recover gracefully, and alert humans only when the situation genuinely requires human judgment. An agent that requires a human check-in every few hours is not a fleet agent. It is an expensive copilot.

The Four Infrastructure Requirements

When you actually build for fleet deployment, four infrastructure requirements emerge that standard agent tooling was not designed to meet.

Persistent, accumulating memory. Fleet agents need memory that persists across sessions and accumulates over time. Not a vector store for similarity search — a structured memory system where each agent maintains a private working context, can access a shared fleet knowledge base, and can write findings that downstream agents read. This memory must be durable (survives restarts), versioned (you can reconstruct state at any point in time), and scoped (agents can only access what they're supposed to access).

Coordination primitives. Agents in a fleet communicate. They hand off tasks, share context, and trigger each other. This requires message passing infrastructure with delivery guarantees — an inbox/outbox model where messages are persisted, retried on failure, and acknowledged on receipt. Without this, inter-agent communication relies on shared databases and polling, which is fragile and produces the kind of coordination bugs that only appear at 3am on a Sunday.

Fleet-level observability. You cannot manage what you cannot observe. Fleet operators need a control plane: which agents are active, which are failing, what each agent has done in the last 24 hours, which dependencies are degraded. This is qualitatively different from application monitoring. Fleet observability is about agent health, task progress, memory consumption, and output quality — not just CPU and latency. The existing monitoring stack was not built for this.

Auditability by default. Fleet operators need to reconstruct any agent's decision at any point in time. What prompt was used. What memory was retrieved. What tools were called. What output was produced. This is an append-only log requirement — the kind that financial services and legal teams have enforced on human decision-making for decades, now extending to the AI agents making or informing those decisions. Building audit trails as an afterthought is always expensive. Building them as a core infrastructure primitive is the only approach that survives regulatory scrutiny.

"The existing monitoring stack was not built for agents. Fleet observability is about agent health, task progress, memory consumption, and output quality. It requires a different instrument entirely."

Why Standard Tooling Fails

LangChain, LlamaIndex, and the orchestration libraries that dominate the current agent landscape are genuinely useful. They solve real problems — connecting retrieval to generation, chaining LLM calls, managing prompt templates. They are the right tools for the single-agent, human-supervised prototype.

They were not designed for fleet operation. The memory model is session-scoped. The state is ephemeral between invocations. The error handling assumes a human can intervene. The observability is whatever you add yourself. These are not implementation details; they are architectural choices that reflect the use case the tools were built for.

Retrofitting fleet requirements onto single-agent tooling is possible — many enterprise teams are doing exactly that — but the cost is high. You end up building bespoke persistence layers, custom message buses, homegrown monitoring dashboards, and audit logging systems. All of this is undifferentiated infrastructure: the same problems solved the same way by every organization that reaches fleet scale.

The right answer is not to build it yourself. It is to have a fleet-first infrastructure layer that makes these requirements the default, not the exception.

What Fleet-First Infrastructure Looks Like

Fleet-first infrastructure starts from the fleet operation use case and works backward to the primitives. The design principles differ from single-agent tooling in ways that matter at the architecture level.

Memory is a first-class entity, not a retrieval plugin. Each agent has a defined memory scope, with explicit rules about what can be written, what can be read, and how memory persists across restarts. The infrastructure enforces these rules — it is not a convention left to the application layer.

Communication is message-passing with guarantees. Agents do not share databases or call each other's APIs directly. They send structured messages to inboxes, and those messages are delivered exactly once, in order, with acknowledgment. This is the pattern that made distributed systems reliable; it is the right pattern for agent coordination.

Observability is part of the runtime, not a layer on top. Every agent action — memory write, message send, tool call, output generated — is captured in an immutable log. The fleet operator's control plane reads from this log. Audit queries answer from this log. You cannot turn it off.

Health management is automatic. The infrastructure detects stalled agents, failed tasks, and degraded dependencies. It restarts what can be restarted, pages humans for what cannot, and maintains a queue of backlogged work so that transient failures do not cause permanent data loss.

The Regulatory Tailwind

There is a second force making fleet infrastructure urgent that has nothing to do with engineering preferences: regulation. The Colorado AI Act takes effect June 30, 2026. The EU AI Act's high-risk obligations apply from August 2, 2026. Both impose requirements on organizations deploying AI in consequential decision-making: documentation of system design, disclosure of AI involvement, impact assessments, and the ability to audit AI outputs after the fact.

NIST's AI Risk Management Framework, increasingly cited as the safe harbor under state AI laws, asks organizations to govern, map, measure, and manage AI risk. Doing this seriously for a fleet of agents — not a single chatbot, but a fleet of agents making continuous decisions — requires infrastructure that captures what those agents did and why.

The organizations that will have the easiest compliance path are those that built auditability into their infrastructure from the start. The organizations scrambling to retrofit audit logging onto stateless agents in June 2026 will have a harder time. The infrastructure layer is where compliance posture is established — not in the application, and not in the policy document.

The Window

Enterprise AI deployments are moving from prototype to production. The teams that got to interesting single-agent results in 2024 are now asking whether they can deploy fleets, what that requires, and who has built the infrastructure to support it.

The answer, right now, is that almost no one has built it as a coherent product. The components exist in fragments — various persistence libraries, monitoring tools, message queues assembled by hand. But the purpose-built fleet infrastructure layer, designed from the ground up for continuous multi-agent operation with enterprise accountability requirements, is what Stratum is building.

The single-agent moment is ending. Every AI deployment that actually works eventually becomes a fleet. The question is what infrastructure it runs on.


Stratum builds domain-specialist AI products on fleet infrastructure — persistent memory, agent coordination, observability, and audit logging as first-class primitives. The same infrastructure that runs our products is what we offer to organizations building their own fleets.

Follow along at onstratum.com →
Sean / Stratum
© 2026 Stratum · hello@onstratum.com · onstratum.com