Compound-engineering-plugin ce-agent-native-audit
Run comprehensive agent-native architecture review with scored principles
git clone https://github.com/EveryInc/compound-engineering-plugin
T=$(mktemp -d) && git clone --depth=1 https://github.com/EveryInc/compound-engineering-plugin "$T" && mkdir -p ~/.claude/skills && cp -r "$T/plugins/compound-engineering/skills/ce-agent-native-audit" ~/.claude/skills/everyinc-compound-engineering-plugin-ce-agent-native-audit && rm -rf "$T"
plugins/compound-engineering/skills/ce-agent-native-audit/SKILL.mdAgent-Native Architecture Audit
Conduct a comprehensive review of the codebase against agent-native architecture principles, launching parallel sub-agents for each principle and producing a scored report.
Core Principles to Audit
- Action Parity - "Whatever the user can do, the agent can do"
- Tools as Primitives - "Tools provide capability, not behavior"
- Context Injection - "System prompt includes dynamic context about app state"
- Shared Workspace - "Agent and user work in the same data space"
- CRUD Completeness - "Every entity has full CRUD (Create, Read, Update, Delete)"
- UI Integration - "Agent actions immediately reflected in UI"
- Capability Discovery - "Users can discover what the agent can do"
- Prompt-Native Features - "Features are prompts defining outcomes, not code"
Workflow
Step 1: Load the Agent-Native Skill
First, invoke the agent-native-architecture skill to understand all principles:
/ce-agent-native-architecture
Select option 7 (action parity) to load the full reference material.
Step 2: Launch Parallel Sub-Agents
Launch 8 parallel sub-agents using the Task tool with
subagent_type: Explore, one for each principle. Each agent should:
- Enumerate ALL instances in the codebase (user actions, tools, contexts, data stores, etc.)
- Check compliance against the principle
- Provide a SPECIFIC SCORE like "X out of Y (percentage%)"
- List specific gaps and recommendations
Agent 1: Action Parity
Audit for ACTION PARITY - "Whatever the user can do, the agent can do." Tasks: 1. Enumerate ALL user actions in frontend (API calls, button clicks, form submissions) - Search for API service files, fetch calls, form handlers - Check routes and components for user interactions 2. Check which have corresponding agent tools - Search for agent tool definitions - Map user actions to agent capabilities 3. Score: "Agent can do X out of Y user actions" Format: ## Action Parity Audit ### User Actions Found | Action | Location | Agent Tool | Status | ### Score: X/Y (percentage%) ### Missing Agent Tools ### Recommendations
Agent 2: Tools as Primitives
Audit for TOOLS AS PRIMITIVES - "Tools provide capability, not behavior." Tasks: 1. Find and read ALL agent tool files 2. Classify each as: - PRIMITIVE (good): read, write, store, list - enables capability without business logic - WORKFLOW (bad): encodes business logic, makes decisions, orchestrates steps 3. Score: "X out of Y tools are proper primitives" Format: ## Tools as Primitives Audit ### Tool Analysis | Tool | File | Type | Reasoning | ### Score: X/Y (percentage%) ### Problematic Tools (workflows that should be primitives) ### Recommendations
Agent 3: Context Injection
Audit for CONTEXT INJECTION - "System prompt includes dynamic context about app state" Tasks: 1. Find context injection code (search for "context", "system prompt", "inject") 2. Read agent prompts and system messages 3. Enumerate what IS injected vs what SHOULD be: - Available resources (files, drafts, documents) - User preferences/settings - Recent activity - Available capabilities listed - Session history - Workspace state Format: ## Context Injection Audit ### Context Types Analysis | Context Type | Injected? | Location | Notes | ### Score: X/Y (percentage%) ### Missing Context ### Recommendations
Agent 4: Shared Workspace
Audit for SHARED WORKSPACE - "Agent and user work in the same data space" Tasks: 1. Identify all data stores/tables/models 2. Check if agents read/write to SAME tables or separate ones 3. Look for sandbox isolation anti-pattern (agent has separate data space) Format: ## Shared Workspace Audit ### Data Store Analysis | Data Store | User Access | Agent Access | Shared? | ### Score: X/Y (percentage%) ### Isolated Data (anti-pattern) ### Recommendations
Agent 5: CRUD Completeness
Audit for CRUD COMPLETENESS - "Every entity has full CRUD" Tasks: 1. Identify all entities/models in the codebase 2. For each entity, check if agent tools exist for: - Create - Read - Update - Delete 3. Score per entity and overall Format: ## CRUD Completeness Audit ### Entity CRUD Analysis | Entity | Create | Read | Update | Delete | Score | ### Overall Score: X/Y entities with full CRUD (percentage%) ### Incomplete Entities (list missing operations) ### Recommendations
Agent 6: UI Integration
Audit for UI INTEGRATION - "Agent actions immediately reflected in UI" Tasks: 1. Check how agent writes/changes propagate to frontend 2. Look for: - Streaming updates (SSE, WebSocket) - Polling mechanisms - Shared state/services - Event buses - File watching 3. Identify "silent actions" anti-pattern (agent changes state but UI doesn't update) Format: ## UI Integration Audit ### Agent Action → UI Update Analysis | Agent Action | UI Mechanism | Immediate? | Notes | ### Score: X/Y (percentage%) ### Silent Actions (anti-pattern) ### Recommendations
Agent 7: Capability Discovery
Audit for CAPABILITY DISCOVERY - "Users can discover what the agent can do" Tasks: 1. Check for these 7 discovery mechanisms: - Onboarding flow showing agent capabilities - Help documentation - Capability hints in UI - Agent self-describes in responses - Suggested prompts/actions - Empty state guidance - Slash commands (/help, /tools) 2. Score against 7 mechanisms Format: ## Capability Discovery Audit ### Discovery Mechanism Analysis | Mechanism | Exists? | Location | Quality | ### Score: X/7 (percentage%) ### Missing Discovery ### Recommendations
Agent 8: Prompt-Native Features
</sub-agents>Audit for PROMPT-NATIVE FEATURES - "Features are prompts defining outcomes, not code" Tasks: 1. Read all agent prompts 2. Classify each feature/behavior as defined in: - PROMPT (good): outcomes defined in natural language - CODE (bad): business logic hardcoded 3. Check if behavior changes require prompt edit vs code change Format: ## Prompt-Native Features Audit ### Feature Definition Analysis | Feature | Defined In | Type | Notes | ### Score: X/Y (percentage%) ### Code-Defined Features (anti-pattern) ### Recommendations
Step 3: Compile Summary Report
After all agents complete, compile a summary with:
## Agent-Native Architecture Review: [Project Name] ### Overall Score Summary | Core Principle | Score | Percentage | Status | |----------------|-------|------------|--------| | Action Parity | X/Y | Z% | ✅/⚠️/❌ | | Tools as Primitives | X/Y | Z% | ✅/⚠️/❌ | | Context Injection | X/Y | Z% | ✅/⚠️/❌ | | Shared Workspace | X/Y | Z% | ✅/⚠️/❌ | | CRUD Completeness | X/Y | Z% | ✅/⚠️/❌ | | UI Integration | X/Y | Z% | ✅/⚠️/❌ | | Capability Discovery | X/Y | Z% | ✅/⚠️/❌ | | Prompt-Native Features | X/Y | Z% | ✅/⚠️/❌ | **Overall Agent-Native Score: X%** ### Status Legend - ✅ Excellent (80%+) - ⚠️ Partial (50-79%) - ❌ Needs Work (<50%) ### Top 10 Recommendations by Impact | Priority | Action | Principle | Effort | |----------|--------|-----------|--------| ### What's Working Excellently [List top 5 strengths]
Success Criteria
- All 8 sub-agents complete their audits
- Each principle has a specific numeric score (X/Y format)
- Summary table shows all scores and status indicators
- Top 10 recommendations are prioritized by impact
- Report identifies both strengths and gaps
Optional: Single Principle Audit
If $ARGUMENTS specifies a single principle (e.g., "action parity"), only run that sub-agent and provide detailed findings for that principle alone.
Valid arguments:
oraction parity1
ortools
orprimitives2
orcontext
orinjection3
orshared
orworkspace4
orcrud5
orui
orintegration6
ordiscovery7
orprompt
orfeatures8