Skillshub common-context-optimization
Techniques to maximize context window efficiency, reduce latency, and prevent 'lost in middle' issues through strategic masking and compaction. (triggers: *.log, chat-history.json, reduce tokens, optimize context, summarize history, clear output)
install
source · Clone the upstream repo
git clone https://github.com/ComeOnOliver/skillshub
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/ComeOnOliver/skillshub "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/HoangNguyen0403/agent-skills-standard/common-context-optimization" ~/.claude/skills/comeonoliver-skillshub-common-context-optimization && rm -rf "$T"
manifest:
skills/HoangNguyen0403/agent-skills-standard/common-context-optimization/SKILL.mdsource content
Priority: P1 (OPTIMIZATION)
Manage the Attention Budget. Treat context as a scarce resource.
1. Observation Masking (Noise Reduction)
Problem: Large tool outputs (logs, JSON lists) flood context and degrade reasoning. Solution: Replace raw output with semantic summaries after consumption.
- Identify: outputs > 50 lines or > 1kb.
- Extract: Read critical data points immediately.
- Mask: Rewrite history to replace raw data with
.[Reference: <summary_of_findings>] - See:
for patterns.references/masking.md
2. Context Compaction (State Preservation)
Problem: Long conversations drift from original intent. Solution: Recursive summarization that preserves State over Dialogue.
- Trigger: Every 10 turns or 8k tokens.
- Compact:
- Keep: User Goal, Active Task, Current Errors, Key Decisions.
- Drop: Chat chit-chat, intermediate tool calls, corrected assumptions.
- Format: Update
orSystem Prompt
with compacted state.Memory File - See:
for algorithms.references/compaction.md
3. KV-Cache Awareness (Latency)
Goal: Maximize pre-fill cache hits.
- Static Prefix: strict ordering: System -> Tools -> RAG -> User.
- Append-Only: Avoid inserting into the middle of history if possible.
References
Anti-Patterns
- No raw tool dumps: Mask large outputs immediately after extracting data.
- No append-only growth: Compact every 10 turns to preserve intent over dialogue.
- No middle insertions: Append-only history maximizes KV cache hits.