Adaptive Flow Overview
- Triage dossier – Every run invokes a triage adapter (default:
atlas.utils.triage.default_build_dossier) that normalises session metadata into risks, signals, and persona hints. - Capability probe – The probe LLM inspects the dossier plus recent history, returning
{mode, confidence, evidence}. Ifcertify_first_runis enabled and the fingerprint is unseen, the runtime forces a one-time certification (paired) before probing. - Adaptive mode – The orchestrator records the decision, stores probe evidence, and chooses between three lanes:
auto– single-shot execution without validation to keep latency low.paired– single-shot execution with a single validation pass (ideal for certifications).coach– converts the reviewed plan into a single step but always validates and allows a retry.
- Execution loop – Depending on the lane, the Student either executes a single combined step or walks through the reviewed plan. Teacher interventions (validation, guidance, retries) are lane-aware.
- Reward & learning – The Reward System aggregates judges, emits
session_reward, and captures learning notes. Certification verdicts are reused as the reward signal when possible. - Memory & telemetry – Persona memories are refreshed, adaptive summaries are stored, and exporters/streamers consume the structured metadata.
Lane Cheatsheet
| Lane | When it triggers | Supervision profile | Persistence highlights |
|---|---|---|---|
auto | High confidence history | Student executes once, no validation | Telemetry records lane + confidence, reward may be skipped |
paired | Certification required or medium confidence | Student executes once, Teacher validates final answer | Certification flag stored, reward reused from validation |
coach | Low confidence or probe confidence below paired threshold | Plan collapses to single step with validation + optional retry | Guidance compact, adaptive summary logs probe evidence |
adaptive_teaching.probe in your config (see SDK Configuration).
Retries, Guidance, and Certification
- Step retries are only attempted in the
coachlane and are capped byorchestration.max_retries. - The Teacher’s guidance is appended to the execution context and streamed to the console so you can see why a retry happened.
- Certification runs (
pairedon a new fingerprint) mark the session ascertification_runinsideadaptive_summaryand store the verdict for future routing.
Event Stream & Telemetry
Every significant action is published to theExecutionContext event stream. Subscribe to it to power CLI streams, dashboards, or custom logging:
adaptive_summary– active mode, probe payload, confidence, and recent history (surfaces in the console streamer and JSONL exports).steps– per-step attempts, timings, retry status, and guidance messages.session_reward/reward_summary– aggregated reward score plus individual judge breakdowns.triage_dossier,personas_used,persona_updates– the context and outcomes that feed persona learning.
TelemetryPublisher if you need to forward events elsewhere; otherwise the default console streamer handles everything automatically.
Anatomy of atlas.core.run
At a high level the public API (atlas.core.run / atlas.core.arun) performs the following:
- Load and validate your config (
atlas/config/loader.py) and reset theExecutionContext. - Build the agent adapter (
create_from_atlas_config) plus Student and Teacher prompts (atlas.prompts.build_student_prompts/build_teacher_prompts). - Instantiate Student, Teacher, and the session-level Reward
Evaluator. - Load the triage adapter and capability probe client defined in
adaptive_teaching, collecting fingerprint hints for persona memory. - Connect to optional storage (Postgres) and set up telemetry publishers or console streaming.
- Run the
Orchestrator, which performs triage, probes for a lane, executes the plan (single-shot or stepwise), and records results. - Persist session metadata—including
adaptive_summary, reward payloads, and persona updates—before returning anatlas.types.Result.
When to Customize
| Goal | Consider tweaking |
|---|---|
| Force a specific lane | adaptive_teaching.mode_override |
| Bias routing thresholds | adaptive_teaching.probe.thresholds and fallback_mode |
| Tighten or relax retries | orchestration.max_retries and Teacher guidance prompts |
| Adjust reward escalation | rim.variance_threshold, rim.uncertainty_threshold, or custom reward objectives |
| Stream custom telemetry | Attach a TelemetryPublisher or subscribe directly to the event stream |
Next Steps
- Configure each YAML block in detail with the
SDK Configuration Reference. - Bring your own agent with the
Bring Your Own Agentguide. - Understand the dual-agent reasoning concept in
Adaptive Dual-Agent Reasoning.