From 12 Agents to 1: AI Agent Architecture Decision Guide
Published on 27.03.2026
From 12 Agents to 1: AI Agent Architecture Decision Guide
TLDR: Most teams are overengineering their AI agent architectures. This decision framework from Decoding AI lays out a complexity spectrum from simple workflows to multi-agent systems, and argues you should stay as far left (simple) as possible. A real-world marketing platform that initially spec'd 12+ agents was solved with just one.
Summary:
Alright, this one hits close to home because I have watched this exact mistake happen over and over again. A team gets excited about AI agents, reads a few papers on multi-agent orchestration, and suddenly their architecture diagram looks like a conspiracy theorist's corkboard. Louis-Francois Bouchard from Towards AI laid out a framework, and the Decoding AI newsletter breaks it down into something you can actually use to make decisions. The core argument is deceptively simple: there is a complexity spectrum with three levels, and you should stay as far left as you possibly can while still solving the problem.
Level 1 is workflows, where you define the steps and the order. Level 2 is a single agent with tools, where the model decides what to do next. Level 3 is multi-agent systems, where multiple models coordinate. The critical distinction that most people miss is this: in a workflow, YOU control the flow. In an agent, the MODEL controls the flow. That is a massive difference in terms of predictability, debuggability, and operational risk. And here is the thing that really needs to sink in: a single model calling ten different APIs is not ten agents. It is one agent with ten tools. Tools are not agents. I cannot stress this enough because the confusion between tools and agents is responsible for a huge amount of unnecessary architectural complexity in the wild right now.
The article shares a real example that is worth its weight in gold. A marketing platform team initially designed a system with twelve-plus agents. After applying this framework, they solved the entire problem with a single agent. Twelve to one. That is not a minor optimization, that is an entirely different system with dramatically lower operational complexity, fewer failure modes, and easier debugging. The reason this works is that most tasks are tightly coupled, sequential, and need shared global context. When all your "agents" need to see the same information and operate in sequence, you do not have a multi-agent problem. You have a single-agent-with-tools problem.
Now, the article does acknowledge when multi-agent systems are genuinely justified: true parallelism where tasks can run independently, context overload where a single context window degrades performance due to the "lost in the middle" problem, modularity requirements with third-party integrations, and security boundaries where different components need different access levels. The recommended pattern for when you do need multiple agents is plan-and-execute combined with orchestrator-worker, essentially one brain that delegates to specialized workers. That is a solid pattern, but I wish the article had pushed harder on the operational cost side. Running multiple agents means multiple LLM calls, multiple failure points, multiple retry strategies, and multiple sets of logs to correlate. The cost is not just architectural complexity, it is operational complexity that compounds every single day your system is in production.
What I think is missing from this framework is a discussion about evaluation and observability. How do you know if your single agent is actually performing well enough, or if you have hit the point where context rot is degrading your results? The article mentions the "lost in the middle" problem but does not give you concrete signals to watch for. That would be the difference between a decision guide and a decision-plus-monitoring guide.
Key takeaways:
- Stay as far left on the complexity spectrum as possible: workflows first, then single agent with tools, then multi-agent only when truly justified
- Tools are not agents — a single model calling multiple APIs is one agent with many tools, not many agents
- Workflows win for predictable, repeatable processes where you want deterministic control flow
- Multi-agent is only justified for true parallelism, context overload, third-party modularity, or security boundaries
- The plan-and-execute plus orchestrator-worker pattern is the recommended approach when you genuinely need multi-agent coordination
- Context rot is real — LLM performance degrades as context grows, so monitor for the "lost in the middle" problem
Why do I care: If you are building anything with LLMs right now, this framework should be taped to your monitor. The industry is in the "peak of inflated expectations" phase for multi-agent architectures, and the gravitational pull toward overengineering is enormous. Every vendor wants to sell you an agent framework. Every conference talk features elaborate multi-agent demos. But the teams actually shipping reliable AI systems in production are, more often than not, using the simplest architecture that solves the problem. This decision guide gives you the vocabulary and the framework to push back when someone on your team says "we need twelve agents" and the answer is actually one.