configs/examples/sdk_quickstart.yaml
, then graduate to the richer examples in configs/examples/openai_agent.yaml
, http_agent.yaml
, and python_agent.yaml
.
All examples reference files in the Atlas SDK repo (
configs/examples/
). If you cloned the repo for the quickstart, you already have them.Layout at a Glance
Agent Configuration (agent
)
This block defines how the orchestrator communicates with your agent.
Adapter | Best for | Key fields |
---|---|---|
openai | OpenAI or Azure OpenAI chat models | llm block (model, key env var, temperature, token limits) |
http_api | Remote services behind an HTTP interface | transport (base URL, retry policy), payload_template , result_path |
python | Local Python functions or callables | import_path , attribute , working_directory , optional llm metadata |
OpenAI Adapter Example
Any non-OpenAI-compatible API (like Anthropic Claude) should use the
http_api
adapter until a dedicated integration is available.Bring Your Own Agent
for full walkthroughs of all three adapters.
Student Configuration (student
)
The Student block shapes planning, execution, and synthesis.
- Prompts: Templates that receive
{base_prompt}
from the agent’s system prompt and any prompt rewriting. - Token limits: Guardrails for LLM calls; increase them when steps are truncated.
tool_choice
:auto
lets the Student call registered tools. Userequired
to force a tool call on every step.
Teacher Configuration (teacher
)
The Teacher reviews plans, validates outputs, and provides guidance.
- LLM block: Often mirrors the Student’s provider but can be a more powerful model for complex reviews.
plan_cache_seconds
: Caches an approved plan for a given task ID to avoid re-running reviews.- Guidance/validation caps: Keep feedback concise; increase when you expect lengthy traces.
Orchestration (orchestration
)
These policies govern the runtime loop.
max_retries
: Default is 1; increase to allow for more revision loops.step_timeout_seconds
: Per-step timeout; extend when calling slow tools or external APIs.rim_guidance_tag
: The tag used to inject feedback from the Reward System back into prompts.emit_intermediate_steps
: Keeptrue
to stream events for logging and telemetry. (Note: Telemetry dashboards are not yet documented).
Reward System (the rim
block)
The Reward System scores each attempt so the orchestrator knows whether to accept an answer or trigger a retry. The YAML block is named rim
for compatibility with the core training engine.
- Judges: At least one is required. Mix
process
,helpfulness
, or custom judges to fit your domain. - Arbiter: A separate LLM that resolves disagreements between judges.
- Thresholds: If a score is below
retry_threshold
, the Teacher is asked for guidance. Abovesuccess_threshold
, the answer is accepted.
Reward Design
.
Storage (storage
)
Optional Postgres persistence for traces and session metadata.
storage: null
(as in the quickstart) when you don’t have a database. Enable it when you’re ready to collect sessions.
Prompt Rewrite (prompt_rewrite
)
This block lets an additional LLM tighten or extend Student/Teacher prompts before a run starts. The quickstart keeps this disabled (prompt_rewrite: null
).
Cheat Sheet
Goal | Section to edit | Quick pointer |
---|---|---|
Swap to Anthropic via HTTP | agent | Use the http_api adapter and point transport.base_url to your service. |
Increase reasoning depth | student and teacher | Raise the max_..._tokens limits for the planner, executor, or guidance. |
Persist sessions | storage | Provide a reachable Postgres URL and credentials. |
Tighten the quality bar | rim | Increase the success_threshold in the Reward System or add a new judge. |
Personalize prompts | prompt_rewrite | Enable the block and configure the rewrite LLM. |
Next Steps
- Walk through the runtime loop in
How Orchestration Works
. - Choose the right connector in
Bring Your Own Agent
. - See how runtime roles compare to training in
Student & Teacher Roles
.