Claude-skill-registry-data moai-alfred-practices
Enterprise practical workflows, context engineering strategies, JIT (Just-In-Time) retrieval optimization, real-world execution examples, debugging patterns, and moai-adk workflow mastery; activates for workflow pattern learning, context optimization, debugging issue resolution, feature implementation end-to-end, and team knowledge transfer
install
source · Clone the upstream repo
git clone https://github.com/majiayu000/claude-skill-registry-data
Claude Code · Install into ~/.claude/skills/
T=$(mktemp -d) && git clone --depth=1 https://github.com/majiayu000/claude-skill-registry-data "$T" && mkdir -p ~/.claude/skills && cp -r "$T/data/moai-alfred-practices" ~/.claude/skills/majiayu000-claude-skill-registry-data-moai-alfred-practices && rm -rf "$T"
manifest:
data/moai-alfred-practices/SKILL.mdsource content
Enterprise Practical Workflows & Context Engineering v4.0.0
Skill Metadata
| Field | Value |
|---|---|
| Skill Name | moai-alfred-practices |
| Version | 4.0.0 Enterprise (2025-11-12) |
| Focus | Practical execution patterns, real-world scenarios |
| Auto-load | When workflow guidance or debugging help needed |
| Included Patterns | 15+ real-world scenarios |
| Lines of Content | 950+ with 20+ production examples |
| Progressive Disclosure | 3-level (quick-patterns, scenarios, advanced) |
What It Does
Provides practical workflows, context engineering strategies, real-world execution examples, and debugging solutions for moai-adk. Covers JIT context management, efficient agent usage, SPEC→TDD→Sync execution, and common problem resolution.
JIT (Just-In-Time) Context Strategy
Principle: Load Only What's Needed Now
Traditional (overload): Load entire codebase → Context window fills immediately → Limited reasoning capacity → Slow, inefficient JIT (optimized): Load core entry points → Identify specific function/module → Load only that section → Cache in thread context → Reuse for related tasks → Minimal context waste
Practice 1: Core Module Mapping
# 1. Get high-level structure find src/ -type f -name "*.py" | wc -l # Output: 145 files total # 2. Identify entry points (only 3-5 files) find src/ -name "__main__.py" -o -name "main.py" -o -name "run.py" # 3. Load entry point + immediate dependencies Glob("src/{**/}*.py") # Load only files referenced by entry point # 4. Cache in Task() context for reuse Task(prompt="Task 1 using mapped modules") Task(prompt="Task 2 reuses cached context")
Practice 2: Dependency Tree Navigation
Project Root ├─ src/ │ ├─ __init__.py ← Entry point #1 │ ├─ main.py ← Entry point #2 │ ├─ core/ │ │ ├─ domain.py ← Core models │ │ ├─ repository.py ← Data access │ │ └─ service.py ← Business logic │ └─ api/ │ ├─ routes.py ← API endpoints │ └─ handlers.py ← Request handlers Load strategy: 1. Load main.py + __init__.py (entry points) 2. When modifying API → Load api/ subtree 3. When fixing business logic → Load core/service.py 4. Cache all loaded files in context 5. Share context between related tasks
Practice 3: Context Reuse Across Tasks
# Task 1: Understand module structure analysis = Task({ prompt="Map src/ directory structure, identify entry points, list dependencies" }) # Task 2: Reuse analysis for implementation implementation = Task({ prompt=f"""Using this structure: {analysis} Now implement feature X... """ }) # Task 3: Reuse analysis for testing testing = Task({ prompt=f"""Using this structure: {analysis} Write tests for feature X... """ }) # Result: No re-mapping, efficient context reuse
SPEC → TDD → Sync Execution Pattern
Step 1: Create SPEC with /alfred:1-plan
/alfred:1-plan/alfred:1-plan "Add user authentication with JWT" # This creates: # .moai/specs/SPEC-042/spec.md (full requirements) # feature/SPEC-042 (git branch) # Track with TodoWrite
Step 2: Implement with /alfred:2-run SPEC-042
/alfred:2-run SPEC-042RED: Test agent writes failing tests ↓ GREEN: Implementer agent creates minimal code ↓ REFACTOR: Quality agent improves code ↓ Repeat TDD cycle for each feature component ↓ All tests passing, coverage ≥85%
Step 3: Sync with /alfred:3-sync auto SPEC-042
/alfred:3-sync auto SPEC-042Updates: ✓ Documentation ✓ Test coverage metrics ✓ Creates PR to develop ✓ Auto-validation of quality gates
Debugging Pattern: Issue → Root Cause → Fix
Step 1: Triage & Understand
Error message: "Cannot read property 'user_id' of undefined" Questions: - When does it occur? (always, intermittently, specific scenario) - Which code path? (which endpoint/function) - What's the state? (what data led to this) - What changed recently? (revert to narrow down)
Step 2: Isolate Root Cause
# Method 1: Binary search # Is it in API layer? → Yes # Is it in route handler? → No # Is it in service layer? → Yes # Is it in this function? → Narrow down # Method 2: Add logging logger.debug(f"user_id = {user_id}") # Check where it becomes undefined # Method 3: Test locally # Reproduce with minimal example # Add breakpoint in debugger # Step through execution
Step 3: Fix with Tests
# RED: Write failing test def test_handles_missing_user_id(): """Should handle case when user_id is undefined.""" assert get_user(None) raises ValueError # GREEN: Minimal fix def get_user(user_id): if not user_id: raise ValueError("user_id required") return fetch_user(user_id) # REFACTOR: Improve def get_user(user_id: int) -> User: """Get user by ID. Args: user_id: User identifier Raises: ValueError: If user_id is None or invalid """ if not user_id or user_id <= 0: raise ValueError(f"Invalid user_id: {user_id}") return self.user_repo.find(user_id)
5 Real-World Scenarios
Scenario 1: Feature Implementation (2-3 hours)
1. Create SPEC: /alfred:1-plan "Add user dashboard" 2. Clarify details: AskUserQuestion (which data to show?) 3. Implement: /alfred:2-run SPEC-XXX (TDD cycle) 4. Document: /alfred:3-sync auto SPEC-XXX 5. Result: Production-ready feature
Scenario 2: Bug Investigation (1-2 hours)
1. Reproduce: Create minimal test case 2. Isolate: Narrow down affected code 3. Debug: Add logging, trace execution 4. Fix: TDD RED→GREEN→REFACTOR 5. Validate: Ensure tests pass, regression tests
Scenario 3: Large Refactoring (4-8 hours)
1. Analyze: Map current code structure 2. Plan: Design new structure with trade-offs 3. Clone pattern: Create autonomous agents for parallel refactoring 4. Integrate: Verify all pieces work together 5. Test: Comprehensive test coverage
Scenario 4: Performance Optimization (2-4 hours)
1. Profile: Identify bottleneck with profiler 2. Analyze: Understand performance characteristics 3. Design: Plan optimization approach 4. Implement: TDD RED→GREEN→REFACTOR 5. Validate: Benchmark before/after
Scenario 5: Multi-Team Coordination (ongoing)
1. SPEC clarity: AskUserQuestion for ambiguous requirements 2. Agent routing: Delegate to specialist teams 3. Progress tracking: TodoWrite for coordination 4. Integration: Verify components work together 5. Documentation: Central SPEC as source of truth
Context Budget Optimization
Typical project context: - Config files: ~50 tokens - .moai/ structure: ~100 tokens - Entry points (3-5 files): ~500 tokens - SPEC document: ~200 tokens → Total: ~850 tokens per session Reusable context: - Load once per session - Share across 5-10 tasks - Saves: 3,500-8,500 tokens per session - Result: More reasoning capacity
Best Practices
DO
- ✅ Load entry points first (3-5 files)
- ✅ Identify dependencies before deep dive
- ✅ Reuse analyzed context across tasks
- ✅ Cache intermediate results in Task context
- ✅ Follow SPEC → TDD → Sync workflow
- ✅ Track progress with TodoWrite
- ✅ Ask for clarification (AskUserQuestion)
- ✅ Test before declaring done
DON'T
- ❌ Load entire codebase at once
- ❌ Reanalyze same code multiple times
- ❌ Skip SPEC clarification (causes rework)
- ❌ Write code without tests
- ❌ Ignore error messages
- ❌ Assume context understanding
- ❌ Skip documentation updates
- ❌ Commit without running tests
Related Skills
(Agent orchestration patterns)moai-alfred-agent-guide
(Complex task delegation)moai-alfred-clone-pattern
(Debugging techniques)moai-essentials-debug
For detailed workflow examples: reference.md
For real-world scenarios: examples.md
Last Updated: 2025-11-12
Status: Production Ready (Enterprise v4.0.0)