Autosearch autosearch:model-routing
Advisory skill — three-tier (Fast / Standard / Best) model routing catalog for autosearch skills. Tells the runtime AI which tier each leaf skill needs, and how to escalate or de-escalate. Autosearch does not switch models itself; the runtime AI is the decision-maker.
install
source · Clone the upstream repo
git clone https://github.com/0xmariowu/Autosearch
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/0xmariowu/Autosearch "$T" && mkdir -p ~/.claude/skills && cp -r "$T/autosearch/skills/meta/model-routing" ~/.claude/skills/0xmariowu-autosearch-autosearch-model-routing && rm -rf "$T"
manifest:
autosearch/skills/meta/model-routing/SKILL.mdsource content
Model Tier Routing — Advisory
Routing principle: most steps use the runtime's cheapest model; only the critical 1–2 steps use the best model.
Autosearch stamps every leaf skill with a
model_tier suggestion. This skill tells the runtime AI what the three tiers mean, which skills default to which tier, and when to escalate or de-escalate.
Three Tiers
| Tier | Typical runtime pick | When used | Share of skills |
|---|---|---|---|
| Fast | Claude Haiku / GPT-5-mini / Gemini 2.5 Flash / Qwen local | Retrieval, normalization, schema checks, URL reading, metadata | ~60% |
| Standard | Claude Sonnet / GPT-5.4 / Gemini 2.5 Pro | Semantic ranking, evidence extraction, mid-complexity planning | ~25% |
| Best | Claude Opus / GPT-5 / Gemini 2.5 Ultra | Clarify, decompose, synthesize, evaluate delivery, skill evolution — the 1-2 steps that shape everything | ~15% |
Tier Assignments
Each autosearch skill carries
model_tier: Fast|Standard|Best in its frontmatter. The runtime AI reads that field before choosing which provider/model to call.
Best (~13 skills — the critical 1-2 steps per session)
— disambiguate intent (wrong clarification cascades)clarify
— global recall planning (missed angles compound)systematic-recall
— breaking a multi-part problemdecompose-task
— produce frameworks, not link listssynthesize-knowledge
— quality gate on final outputevaluate-delivery
— cross-evidence relation graphknowledge-map
/check-rubrics
— rubric-driven evaluationgenerate-rubrics
/auto-evolve
— anything that changes future behaviorcreate-skill
— multi-round goal convergencegoal-loop
(when present) — research plan as graphgraph-search-plan
(when present) — multi-persona question generationperspective-questioning
(when present) — explicit gaps / visited / bad-URLs loopreflective-search-loop
Standard (~20 skills — semantic judgment, structurable)
— pick 5-10 channels from 41select-channels
— combinatorial query generationgene-query
— prior art lookupconsult-reference
— semantic ranking of resultsrerank-evidence
— per-item relevance scorellm-evaluate
— spam / score-gaming detectionanti-cheat
— token-budgeted context assemblyassemble-context
— structured extraction from textextract-knowledge
/fetch-crawl4ai
/fetch-playwright
/fetch-firecrawlfollow-links
— rule promotionexperience-compact
— user preference inferenceobserve-user
— speed vs. deep choiceresearch-mode
(when present) /delegate-subtask
/trace-harvest
/citation-indexrecent-signal-fusion
/interact-user
/pipeline-flowoutcome-tracker
Fast (~60 skills — bulk of every session)
- All 41 channel skills (
) — retrieval, not reasoningsearch-*
,fetch-jina
— URL → Markdownfetch-webpage
+ three video-to-text transcription skills — mechanical extractionyt-dlp
routing skillmcporter
,discover-environment
— env probingprovider-health
,normalize-results
— schema + dedupeextract-dates
— routing decision, no deep reasoningautosearch:router
— append-only event write, often no LLMexperience-capture
(when present) — keep-last-k rulescontext-retention-policy
Escalation Rules
Start a step at its default tier. Escalate only when:
- Conflicting evidence from multiple sources needs semantic reconciliation →
fromStandard
.Fast - Final synthesis or skill-evolution output depends on this step →
fromBest
.Standard - User explicitly asks for deeper analysis / higher quality on the topic.
De-escalate when:
- A Standard step is running on highly structured / deterministic input (e.g.
on 3 items with clear metadata) → drop to Fast.rerank-evidence - Exploratory / draft iteration loop — first passes can be Fast, final pass Best.
Runtime Advisory (non-binding)
Autosearch cannot force the runtime to change models. The
model_tier field is a suggestion to help the runtime AI choose a provider/model route that matches the quality bar of the step. The runtime AI may ignore the advisory if it has better information (e.g. user specified a fixed model).
Boss Rules
- Cost: run the cheapest model that clears the quality bar; reserve best model for the 1-2 steps that shape the whole session outcome.
- Judge: keep LLM-as-judge (pairwise A/B preference, open_deep_research pattern). Do not invent N-dim × 0/3/5 rubrics.
- AVO: any
-tier skill that modifies SKILL.md (likeBest
) must commit viaauto-evolve
and be reversible viascripts/committer
.git revert
Quality Bar
- Evidence items have non-empty title and url.
- No crash on empty or malformed API response.
- Source channel field matches the channel name.