Intelligent context compression that preserves conversation fidelity while preventing LLM context window overflow — with unlimited follow-up questions.
💡 Click any node to see details, code snippets, and implementation notes
When compaction triggers, the buffer is segmented into three zones based on recency. Each zone receives different treatment to balance token savings vs. fidelity.
Beyond the core spec — these significantly increase robustness, cost-efficiency, and quality.
pinned. System prompts, key decisions, and errors survive even Zone C full compaction. Prevents silent loss of critical context.Source of truth for context limits, thresholds, and compaction model assignments. Checked on every trigger.