Caylent Catalysts™
Generative AI Strategy
Accelerate your generative AI initiatives with ideation sessions for use case prioritization, foundation model selection, and an assessment of your data landscape and organizational readiness.
Explore the importance of AI orchestration and learn how to design and implement it effectively using Amazon Bedrock.
Generative AI has been advancing at a rapid pace over the past year, and even in just the last few months, we’ve seen some remarkable breakthroughs. While the direct use of LLM models can still be tremendously effective, especially when combined with fine-tuning and Retrieval-Augmented Generation (RAG), the use of agents represents a seismic change. When using agents, orchestration is a requirement. This blog explains the importance of AI orchestration and how to use it with Amazon Bedrock.
The evolution of GenAI systems typically follows this progression:
Agents are components that models can call to request additional information or perform actions. The capabilities of an agent are defined by the instructions given to it when it is created. For example, a simple agent instruction might be: “Search for a patient’s ID using their full name.”
If a large language model (LLM) with access to this agent is asked to perform a task that requires an ID lookup, it can delegate that part of the task to the agent.
When agents were first being developed, a common pattern was to give the agent multiple optional instructions, allowing it to perform a variety of tasks. However, it was soon noticed that this led to confusion by the LLM if the distinct instructions were not clearly separated.
In this example, we show an agent with overlapping instructions:
Agent Instructions:
Resulting Agent Confusion:
If a customer asks about a feature that isn’t available, the agent must decide whether to:
In some cases, these instructions can conflict—for example, recommending a new product (Instruction 3) vs. suggesting a workaround (Instruction 4) might lead to contradictory actions.
To avoid this confusion, we replace one complex agent with multiple simpler agents. However, this now means that we must determine how the agents can work together. For agents, this is referred to as AI orchestration. It also means that we must have a deeper understanding of the system's underlying business logic. In other words, you can’t automate a system you don’t understand (more on that later).
The type of AI orchestration required depends heavily on the workflow it is supporting. Let’s examine two examples that require quite different types of orchestration. The first example has a highly structured orchestration because the underlying business process is also highly structured. The second example is non-deterministic because, again, the underlying business process is non-deterministic. When we say 'non-deterministic,' we mean that the conversation can go in multiple directions based on the input.
Workflow:
A "Know Your Customer" (KYC) process for onboarding new banking clients.
Steps:
Why Deterministic Orchestration Works:
Compare this structured workflow with a non-deterministic one.
Workflow:
Resolving a complex customer complaint (e.g., a technical issue with a smart home device).
Steps:
Why Non-Deterministic Orchestration Works:
If you’ve ever worked in a call center, you know from experience how every call can be different and every customer interaction can be unpredictable.
Amazon Bedrock’s approach to AI agent orchestration is a significant step forward in making complex, multi-agent AI systems accessible and manageable, but it comes with both strengths and trade-offs that practitioners should carefully consider. Amazon Bedrock introduces a supervisor-based architecture in which a central “supervisor agent” coordinates a set of specialized agents, each handling distinct subtasks within a larger workflow. This design streamlines the orchestration of multi-step, domain-specific processes such as financial due diligence or customer service automation by automatically decomposing tasks, routing them to the right agents, and consolidating outputs.
However, while Amazon Bedrock reduces much of the technical overhead traditionally required for agent orchestration, such as manual coordination logic and state management, it also introduces a layer of abstraction that may limit granular control for advanced users. For example, the default ReAct orchestration method provides transparency and step-by-step reasoning but can become a bottleneck in scenarios demanding high throughput or parallelism. AWS addresses this with a custom orchestrator option, allowing users to implement bespoke orchestration strategies via AWS Lambda, thereby regaining fine-grained control over agent workflows, real-time adjustments, and state transitions. Yet, this flexibility comes at the cost of increased complexity and responsibility for the developer; bespoke strategies are almost always more complex than standard ones.
It’s important to note that Amazon Bedrock’s multi-agent pipelines are primarily designed for workflows that follow a clear, linear path. If your use case involves frequent back-and-forth or unpredictable interactions between agents, Amazon Bedrock may not be the ideal fit on its own. In such situations, combining Amazon Bedrock with open-source orchestration tools like LangGraph or CrewAI can provide the flexibility and collaborative reasoning you need, as they enable more customization. However, this comes at the cost of increased integration work and operational overhead. It is worth noting that new orchestration tools and frameworks are being introduced every month (see the new Strands Agent SDK, for example).
Amazon Bedrock offers several distinct approaches to creating orchestrations, enabling flexible ways to design and manage complex agentic workflows. One of these approaches is using Flows, which are created via the Amazon Bedrock Flow Editor, an interface that has a similar feel to the Step Function Editor. You drag and drop one or more items from the Nodes list and then draw connections between the nodes. The following flow is from an AWS video tutorial:
To use Amazon Bedrock agent orchestration, you first create one or more agents, then create an agent with “Multi-agent collaboration” enabled. This agent acts as the Orchestrator, allowing you to associate additional collaborator agents with it.
Note that the process of setting up collaborator agents can be somewhat unintuitive, but being aware of them can help you navigate the process more smoothly. You’ll first create the agent, prepare it, and then create an alias (which creates a version). Only after performing these steps will the new agent be available to the supervisor. Also, be aware that when naming collaborator agents, the names are supposed to follow a specific pattern, but the interface will still allow you to enter an invalid name. To fix it, simply replace spaces with an underscore or dash. Knowing this in advance can help you avoid unnecessary troubleshooting.
Just as you give any agent a set of instructions, the supervisor agent requires instructions on how it should delegate requests to and between its collaborator agents. This is where understanding and clearly defining your business logic becomes critical.
Just as one needs clean and well-understood data as input to an LLM, one needs clear and well-understood business rules to create an agentic orchestration. Referring back to our earlier example of an agent with overlapping instructions, imagine that agent being split into four specific agents. The problem now becomes “What should the supervisor agent’s instructions be?” The business rules for responding to requests that may appear ambiguous need to be clarified before a supervisor can encode them. Put simply: you cannot automate a process you don’t understand.
As an example, the author created a system for his local town that lets residents ask questions about town bylaws, building codes, and various committee meetings. The system has three agents, one for each type of information to be requested. The instructions for each of those agents are very specific. For example, the instructions to review committee meetings state that the agent can find information about what was discussed in various committees on various topics. It also says that if the user request mentions “meetings” or “minutes," the agent can handle that as well. The word “minutes” is interesting because it has a specific meaning when applied to meeting minutes, but it's also a common word. In this case, however, it is unlikely that someone would use the word minutes when asking a question about bylaws or building codes. This type of analysis is necessary when creating instructions for a multi-agent system.
Multi-agent systems can be thought of as similar to distributed systems or microservice systems. All of these have large attack surfaces that require protection. Distributed systems made calls to various APIs, some of which were local and some remote. All of these remote calls had a myriad of failure cases; we need to remember the lessons learned and apply them to multi-agent systems.
For agent orchestration, many of these errors will manifest themselves as decreased responsiveness:
The supervisor agent must also contain instructions for how it should handle cases where agents fail to return results. In the example above, we should add a statement along the lines of “if the agent you select does not return a result, you should apologize.”
Every line in the diagram below is a potential attack vector.
As with all AWS systems, we rely on the Shared Responsibility Model for protecting our systems. This means we need to utilize all the standard tools, such as AWS IAM principle of least privilege, Amazon CloudWatch, and AWS CloudTrail, among others. It also means that we need to use Amazon Bedrock’s specialized defence system: Amazon Bedrock Guardrails.
Amazon Bedrock Guardrails can defend against a multitude of attacks by detecting and preventing discussions of certain topics or keywords, redacting various types of PII, and reducing hallucinations through contextual grounding. Refer to the diagrams below for examples of guardrails that can be implemented.
As generative AI systems grow increasingly sophisticated, the need for robust orchestration, particularly in multi-agentic systems, cannot be overstated. Amazon Bedrock offers a powerful and accessible solution for managing complex workflows. However, its effectiveness ultimately depends on well-defined business logic and a thoughtful approach to agent design.
By leveraging Amazon Bedrock’s supervisor-based architecture, organizations can streamline multi-step processes and maintain compliance, while custom orchestration options and integration with open-source tools offer flexibility for more dynamic or unpredictable scenarios. However, as with any distributed system, security and operational resilience must remain top priorities, with AWS’s Shared Responsibility Model and Amazon Bedrock’s specialized guardrails serving as essential safeguards. Ultimately, mastering agent orchestration is not just a technical challenge but a strategic imperative for any organization looking to harness the full potential of generative AI.
At Caylent, we help organizations build scalable, cost-efficient GenAI solutions that turn innovation into impact. Whether you're looking to accelerate your generative AI initiatives or future-proof your AI initiatives with a well-defined LLMOps strategy, our experts are here to guide you at every step. Get in touch to find out how we can help!
Brian is an AWS Community Hero, Alexa Champion, runs the Boston AWS User Group, has ten US patents and a bunch of certifications. He's also part of the New Voices mentorship program where Heros teach traditionally underrepresented engineers how to give presentations. He is a private pilot, a rescue scuba diver and got his Masters in Cognitive Psychology working with bottlenosed dolphins.
View Brian's articlesCaylent Catalysts™
Accelerate your generative AI initiatives with ideation sessions for use case prioritization, foundation model selection, and an assessment of your data landscape and organizational readiness.
Caylent Catalysts™
Accelerate investment and mitigate risk when developing generative AI solutions.
Leveraging our accelerators and technical experience
Browse GenAI OfferingsAI-powered solution automates up to 70% of the migration process to help enterprises modernize faster—without the risk; breaks vendor lock-in and eliminates licensing costs.
AWS's new Nova models, in conjunction with Amazon Bedrock Flows, bring new capabilities to AI-enabled applications.