| Specialist | Handles a narrow domain with focused tools | SQL agent, code-review agent |
| Orchestrator | Manages multi-step pipelines and state | Project planner that sequences sub-agents |
| Guardrail | Validates outputs before they reach the user | Safety classifier, PII redactor |
+
+- Note: recent community patterns (e.g., Anthropic's Claude Code implementations and LangChain's "Deep Agents") show practical benefits of combining planning tools, subagents, detailed system prompts, and file-backed workspaces for long-running research and coding tasks. See LangChain Deep Agents (blog) and Anthropic Claude Code docs for examples.
### Model selection & granularity
- Match model capability to the agent role. Use smaller, cheaper models for high-volume routing/triage and larger, more capable models for specialist reasoning or multimodal work.
- Prefer the latest provider-recommended model classes for specialist reasoning when available (example: OpenAI GPT-5.4 as the current high-capability class at the time of writing). Benchmark latency and cost for both router and specialist roles on representative workloads.
−- Recent community signals and vendor updates show managed agent hosting and model-native harnesses are now common; test managed offerings for latency, cost, and sandboxing trade-offs before committing to a deployment architecture (see provider docs linked in Tooling & libraries).
+- Recent vendor updates add model-native harnesses and first-class sandbox primitives; test managed agent hosting and model-native harnesses for latency, cost, and sandboxing trade-offs before committing to a deployment architecture (see Tooling & libraries).
- For routers, optimize for classification accuracy and latency; for specialists, optimize for reasoning depth and tool integration (code execution, file system access, multimodal inputs).
### Agent authorization and credentials
@@ −56 +58 @@
- Keep the schema vendor-agnostic (target, contextPointer, instructions, metadata, auth, model_hint) so routing logic can span multiple providers and platforms (OpenAI, Vercel AI Gateway, Anthropic/Claude deployments).
- Prefer pointers (contextPointer) for long-lived context to avoid token bloat; fetch and rehydrate only the slices the specialist needs.
−### State management patterns
+## State management patterns
Shared state approaches:
- Conversation history passthrough — Simple, stateless agents; token bloat on long chains
@@ −71 +73 @@
## Async & long-running tasks
- Adopt task lifecycle semantics: start_async_task, check_async_task, update_async_task, cancel_async_task, list_async_tasks. Persist task IDs, status, provenance, expected duration, and owner in your external store.
+- LangChain's Deep Agents work formalizes async subagent lifecycle semantics for background work; adopt similar lifecycle semantics when implementing supervisors and UIs (see LangChain Deep Agents: https://blog.langchain.com/deep-agents).
−- LangChain's Deep Agents work and related community implementations formalize async subagent lifecycle semantics for background work (start/check/update/cancel/list) — adopt similar lifecycle semantics when implementing supervisors and UIs (see LangChain Deep Agents: https://blog.langchain.com/deep-agents).
+- OpenAI's Agents SDK now includes native sandbox execution primitives suitable for long-running, file-based workflows where an agent must inspect or modify a controlled workspace. Use sandboxed agents to isolate untrusted execution and restrict filesystem/tool access; consult the OpenAI sandboxes guide for primitives and runtime configuration (OpenAI docs: https://developers.openai.com/api/docs/guides/agents/sandboxes).
−- OpenAI's Agents SDK now includes native sandbox execution primitives suitable for long-running, file-based workflows where an agent must inspect or modify a controlled workspace. Use sandboxed agents to isolate untrusted execution and restrict filesystem/tool access; see the OpenAI sandboxes guide for primitives such as SandboxAgent and SandboxRunConfig (OpenAI docs: https://developers.openai.com/api/docs/guides/agents/sandboxes).
- Design supervisors to avoid blocking on long-running subagents: allow the supervisor to continue interacting with the user, poll or subscribe to task completion events, and provide follow-up instructions to running tasks when necessary.
- Ensure idempotency and cancellation semantics for background tasks. Provide meaningful timeouts and resource limits; surfaced task metadata should include expected duration and resource class.
@@ −109 +111 @@
## Tooling & libraries
−- OpenAI Agents SDK (Apr 2026): model-native harness + native sandbox execution. See the SDK docs for SandboxAgent, SandboxRunConfig, and client examples (e.g., UnixLocalSandboxClient) for file-based workflows and isolated execution (OpenAI docs: https://developers.openai.com/api/docs/guides/agents/sandboxes).
+- OpenAI Agents SDK (Apr 2026): model-native harness + native sandbox execution. See the SDK docs for SandboxAgent, SandboxRunConfig, and example clients for file-based workflows and isolated execution (OpenAI docs: https://developers.openai.com/api/docs/guides/agents/sandboxes).
- LangChain Deep Agents (blog + docs): async subagents, planning tools, subagent lifecycle patterns, and Agent Protocol concepts for cross-deployment interoperability (LangChain blog: https://blog.langchain.com/deep-agents; docs: https://docs.langchain.com/oss/python/deepagents/overview).