Skip to content

Latest commit

 

History

History
266 lines (194 loc) · 9.05 KB

File metadata and controls

266 lines (194 loc) · 9.05 KB

Token Optimization

Introduced in v2.17.0

GSD 2.17 introduces a coordinated token optimization system that can reduce token usage by 40-60% without sacrificing output quality for most workloads. The system has three pillars: token profiles, context compression, and complexity-based task routing.

Token Profiles

A token profile is a single preference that coordinates model selection, phase skipping, and context compression level. Set it in your preferences:

---
version: 1
token_profile: balanced
---

Three profiles are available:

budget — Maximum Savings (40-60% reduction)

Optimized for cost-sensitive workflows. Uses cheaper models, skips optional phases, and compresses dispatch context to the minimum needed.

Dimension Setting
Planning model Sonnet
Execution model Sonnet
Simple task model Haiku
Completion model Haiku
Subagent model Haiku
Milestone research Skipped
Slice research Skipped
Roadmap reassessment Skipped
Context inline level Minimal — drops decisions, requirements, extra templates

Best for: prototyping, small projects, well-understood codebases, cost-conscious iteration.

balanced — Smart Defaults (default)

The default profile. Keeps the important phases, skips the ones with diminishing returns for most projects, and uses standard context compression.

Dimension Setting
Planning model User's default
Execution model User's default
Simple task model User's default
Completion model User's default
Subagent model Sonnet
Milestone research Runs
Slice research Skipped
Roadmap reassessment Runs
Context inline level Standard — includes key context, drops low-signal extras

Best for: most projects, day-to-day development.

quality — Full Context (no compression)

Every phase runs. Every context artifact is inlined. No shortcuts.

Dimension Setting
All models User's configured defaults
All phases Run
Context inline level Full — everything inlined

Best for: complex architectures, greenfield projects requiring deep research, critical production work.

Context Compression

Each token profile maps to an inline level that controls how much context is pre-loaded into dispatch prompts:

Profile Inline Level What's Included
budget minimal Task plan, essential prior summaries (truncated). Drops decisions register, requirements, UAT template, secrets manifest.
balanced standard Task plan, prior summaries, slice plan, roadmap excerpt. Drops some supplementary templates.
quality full Everything — all plans, summaries, decisions, requirements, templates, and root files.

How Compression Works

Dispatch prompt builders accept an inlineLevel parameter. At each level, specific artifacts are gated:

Minimal level reductions:

  • buildExecuteTaskPrompt — drops the decisions template, truncates prior summaries to the most recent one
  • buildPlanMilestonePrompt — drops PROJECT.md, REQUIREMENTS.md, decisions, and supplementary templates like secrets-manifest
  • buildCompleteSlicePrompt — drops requirements and UAT template inlining
  • buildCompleteMilestonePrompt — drops root GSD file inlining
  • buildReassessRoadmapPrompt — drops project, requirements, and decisions files

These are cumulative — standard drops a subset, minimal drops more. The full level preserves all context (the pre-2.17 behavior).

Overriding Inline Level

The inline level is derived from your token_profile. To control phases independently of the profile, use the phases preference:

---
version: 1
token_profile: budget
phases:
  skip_research: false    # override: run research even on budget
---

Explicit phases settings always override the profile defaults.

Complexity-Based Task Routing

GSD automatically classifies each task by complexity and routes it to an appropriate model tier. This means simple documentation fixes don't burn expensive Opus tokens, while complex architectural work gets the reasoning power it needs.

How Classification Works

Tasks are classified by analyzing the task plan:

Signal Simple Standard Complex
Step count ≤ 3 4-7 ≥ 8
File count ≤ 3 4-7 ≥ 8
Description length < 500 chars 500-2000 > 2000 chars
Code blocks ≥ 5
Signal words None Any present

Signal words that prevent simple classification: research, investigate, refactor, migrate, integrate, complex, architect, redesign, security, performance, concurrent, parallel, distributed, backward compat, migration, architecture, concurrency, compatibility.

Empty or malformed plans default to standard (conservative).

Unit Type Defaults

Non-task units have built-in tier assignments:

Unit Type Default Tier
complete-slice, run-uat Light
research-*, plan-*, execute-task, complete-milestone Standard
replan-slice, reassess-roadmap Heavy
hook/* Light

Model Routing

Each tier maps to a model configuration:

Tier Model Phase Key Typical Model
Light completion Haiku (budget) / user default
Standard execution Sonnet / user default
Heavy execution Opus / user default

Simple tasks use the execution_simple model key when configured. This is set automatically by the budget profile to Haiku.

Budget Pressure

When approaching your budget ceiling, the classifier automatically downgrades tiers:

Budget Used Effect
< 50% No adjustment
50-75% Standard → Light
75-90% Standard → Light
> 90% Everything except Heavy → Light; Heavy → Standard

This graduated approach preserves model quality for the most complex work while progressively reducing cost as the ceiling approaches.

Adaptive Learning (Routing History)

GSD tracks the success and failure of each tier assignment over time and adjusts future classifications accordingly. This is opt-in — it happens automatically and persists in .gsd/routing-history.json.

How It Works

  1. After each unit completes, the outcome (success/failure) is recorded against the unit type and tier used
  2. Outcomes are tracked per-pattern (e.g., execute-task, execute-task:docs) with a rolling window of the last 50 entries
  3. If a tier's failure rate exceeds 20% for a given pattern, future classifications for that pattern are bumped up one tier
  4. The system also accepts tag-specific patterns (e.g., execute-task:test vs execute-task:frontend) for more granular routing

User Feedback

GSD accepts manual feedback to accelerate learning:

  • "over" — the model was overpowered for this task (encourages downgrading)
  • "under" — the model wasn't capable enough (encourages upgrading)
  • "ok" — correct assignment (no adjustment)

Feedback signals are weighted 2× compared to automatic outcomes.

Data Management

# Routing history is stored per-project
.gsd/routing-history.json

# Clear history to reset adaptive learning
# (happens via the routing-history module API)

The feedback array is capped at 200 entries. Per-pattern outcome counts use a rolling window of 50 to prevent stale data from dominating.

Configuration Examples

Cost-Optimized Setup

---
version: 1
token_profile: budget
budget_ceiling: 25.00
models:
  execution_simple: claude-haiku-4-5-20250414
---

Balanced with Custom Models

---
version: 1
token_profile: balanced
models:
  planning:
    model: claude-opus-4-6
    fallbacks:
      - openrouter/z-ai/glm-5
  execution: claude-sonnet-4-6
---

Full Quality for Critical Work

---
version: 1
token_profile: quality
models:
  planning: claude-opus-4-6
  execution: claude-opus-4-6
---

Per-Phase Overrides

The token_profile sets defaults, but explicit preferences always win:

---
version: 1
token_profile: budget
phases:
  skip_research: false     # override: keep milestone research
models:
  planning: claude-opus-4-6  # override: use Opus for planning despite budget profile
---

How the Pieces Fit Together

preferences.md
  └─ token_profile: balanced
       ├─ resolveProfileDefaults() → model defaults + phase skip defaults
       ├─ resolveInlineLevel() → standard
       │    └─ prompt builders gate context inclusion by level
       └─ classifyUnitComplexity() → routes to execution/execution_simple model
            ├─ task plan analysis (steps, files, signals)
            ├─ unit type defaults
            ├─ budget pressure adjustment
            └─ adaptive learning from routing-history.json

The profile is resolved once and flows through the entire dispatch pipeline. Explicit preferences override profile defaults at every layer.