Nexus-agents gemini-delegator

install
source · Clone the upstream repo
git clone https://github.com/williamzujkowski/nexus-agents
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/williamzujkowski/nexus-agents "$T" && mkdir -p ~/.claude/skills && cp -r "$T/skills/gemini-delegator" ~/.claude/skills/williamzujkowski-nexus-agents-gemini-delegator && rm -rf "$T"
manifest: skills/gemini-delegator/SKILL.md
source content

Gemini Delegator Skill

<!-- CANONICAL SOURCES: - docs/architecture/ROUTING_SYSTEM.md - packages/nexus-agents/src/mcp/tools/delegate-to-model.ts -->

Full documentation:

Real-World Performance (Tested 2026-01-18)

MetricValueNotes
Max files analyzed978 files in single query~280k LOC processed successfully
Latency (simple)10-15 secondsSingle directory analysis
Latency (complex)58-73 secondsFull codebase analysis
Cache efficiency94k tokens cachedSignificant cost savings
Tool calls/sessionUp to 20File reads auto-executed
Model routingflash-lite → proTwo-model approach for optimization

When to Delegate to Gemini

ConditionThreshold/CriteriaReason
Context size> 100K tokensGemini: 1M context window
Large codebase analysisMultiple files, > 50 filesFast bulk processing
Image/screenshot analysisAny visual contentNative multimodal support
Video/audio processingAny A/V contentNative multimodal support
Speed-critical tasksLatency sensitiveFlash models optimized
Cost-sensitive operationsBudget constraintsGenerous free tier (1K/day)
Google Cloud integrationBigQuery, Cloud FunctionsNative integration

When NOT to Delegate

  • Complex reasoning requiring deep analysis
  • Security-critical implementations
  • Production code generation needing high quality
  • Tasks requiring careful planning and architecture

Methods

Method 1: delegate_to_model MCP Tool

Use the nexus-agents MCP tool for intelligent routing:

# The tool analyzes task and recommends optimal model
nexus-agents delegate_to_model --task "Analyze this 500K token codebase"

Method 2: Direct Gemini CLI

# Basic delegation (human-readable output)
gemini -p "Analyze this codebase"

# JSON output for parsing (note: response may contain markdown fences)
gemini -p "Analyze this codebase" --output-format json

# With auto-approve for autonomous operation
gemini --yolo -p "Review all files in src/" --output-format json

# Specify model
gemini -m gemini-2.5-flash -p "Quick analysis task"
gemini -m gemini-3-pro-preview -p "Complex reasoning task"

JSON Output Parsing Note: The

response
field in JSON output may contain markdown code fences. Parse accordingly:

const result = JSON.parse(output);
const response = result.response.replace(/```\w*\n?/g, '').trim();

Multimodal Examples

Image Analysis

# Analyze screenshot
gemini -p "Describe this UI and identify usability issues" < screenshot.png

# Analyze architecture diagram
gemini -p "Extract components and relationships from this diagram" < diagram.png

# Code screenshot OCR
gemini -p "Extract the code from this screenshot" < code-image.png

Video Analysis

# Analyze video content (uses File API for > 1 minute)
gemini -p "Summarize the key points from this demo video" < demo.mp4

Batch Processing

# Process multiple images
for img in screenshots/*.png; do
  gemini -p "Analyze this UI screenshot" < "$img" --output-format json >> results.json
done

Context Advantage

ModelContext WindowBest For
Claude~200K tokensQuality reasoning
Gemini Pro1M tokensLarge context, multimodal
Gemini Flash200K-1M tokensSpeed, high volume

The 1M token context window enables:

  • Entire codebase analysis in single context
  • Full documentation sets without chunking
  • Complete git history review
  • Large data file processing

Output Formats

# Human readable (default)
gemini -p "task"

# JSON for parsing
gemini -p "task" --output-format json

# Streaming JSON for real-time
gemini -p "task" --output-format stream-json

Process

  1. Evaluate task requirements:

    • Estimate context size
    • Check for multimodal content
    • Assess speed/cost sensitivity
  2. Choose delegation method:

    • Use
      delegate_to_model
      for intelligent routing
    • Use direct Gemini CLI for explicit control
  3. Configure execution:

    • Select appropriate model (Pro vs Flash)
    • Set output format for parsing
    • Use
      --yolo
      for autonomous tasks
  4. Process results:

    • Parse JSON output
    • Integrate findings into workflow

Quick Reference

# Large context analysis
gemini -p "Analyze this entire codebase structure" --yolo

# Screenshot analysis
gemini -p "What does this UI show?" < screenshot.png

# Fast iteration
gemini -m gemini-2.5-flash -p "Quick review" --output-format json

# Cost-sensitive batch
for f in *.ts; do gemini -p "Review: $(cat $f)" >> reviews.txt; done