moai-alfred-practices
About
This skill provides enterprise workflow patterns and context engineering strategies for practical Claude development. It helps developers optimize JIT retrieval, implement features end-to-end, and resolve debugging issues using real-world examples. Use it for mastering workflow patterns, context optimization, and team knowledge transfer.
Quick Install
Claude Code
Recommended/plugin add https://github.com/modu-ai/moai-adkgit clone https://github.com/modu-ai/moai-adk.git ~/.claude/skills/moai-alfred-practicesCopy and paste this command in Claude Code to install this skill
Documentation
Enterprise Practical Workflows & Context Engineering v4.0.0
Skill Metadata
| Field | Value |
|---|---|
| Skill Name | moai-alfred-practices |
| Version | 4.0.0 Enterprise (2025-11-12) |
| Focus | Practical execution patterns, real-world scenarios |
| Auto-load | When workflow guidance or debugging help needed |
| Included Patterns | 15+ real-world scenarios |
| Lines of Content | 950+ with 20+ production examples |
| Progressive Disclosure | 3-level (quick-patterns, scenarios, advanced) |
What It Does
Provides practical workflows, context engineering strategies, real-world execution examples, and debugging solutions for moai-adk. Covers JIT context management, efficient agent usage, SPEC→TDD→Sync execution, and common problem resolution.
JIT (Just-In-Time) Context Strategy
Principle: Load Only What's Needed Now
Traditional (overload):
Load entire codebase
→ Context window fills immediately
→ Limited reasoning capacity
→ Slow, inefficient
JIT (optimized):
Load core entry points
→ Identify specific function/module
→ Load only that section
→ Cache in thread context
→ Reuse for related tasks
→ Minimal context waste
Practice 1: Core Module Mapping
# 1. Get high-level structure
find src/ -type f -name "*.py" | wc -l
# Output: 145 files total
# 2. Identify entry points (only 3-5 files)
find src/ -name "__main__.py" -o -name "main.py" -o -name "run.py"
# 3. Load entry point + immediate dependencies
Glob("src/{**/}*.py")
# Load only files referenced by entry point
# 4. Cache in Task() context for reuse
Task(prompt="Task 1 using mapped modules")
Task(prompt="Task 2 reuses cached context")
Practice 2: Dependency Tree Navigation
Project Root
├─ src/
│ ├─ __init__.py ← Entry point #1
│ ├─ main.py ← Entry point #2
│ ├─ core/
│ │ ├─ domain.py ← Core models
│ │ ├─ repository.py ← Data access
│ │ └─ service.py ← Business logic
│ └─ api/
│ ├─ routes.py ← API endpoints
│ └─ handlers.py ← Request handlers
Load strategy:
1. Load main.py + __init__.py (entry points)
2. When modifying API → Load api/ subtree
3. When fixing business logic → Load core/service.py
4. Cache all loaded files in context
5. Share context between related tasks
Practice 3: Context Reuse Across Tasks
# Task 1: Understand module structure
analysis = Task({
prompt="Map src/ directory structure, identify entry points, list dependencies"
})
# Task 2: Reuse analysis for implementation
implementation = Task({
prompt=f"""Using this structure:
{analysis}
Now implement feature X...
"""
})
# Task 3: Reuse analysis for testing
testing = Task({
prompt=f"""Using this structure:
{analysis}
Write tests for feature X...
"""
})
# Result: No re-mapping, efficient context reuse
SPEC → TDD → Sync Execution Pattern
Step 1: Create SPEC with /alfred:1-plan
/alfred:1-plan "Add user authentication with JWT"
# This creates:
# .moai/specs/SPEC-042/spec.md (full requirements)
# feature/SPEC-042 (git branch)
# Track with TodoWrite
Step 2: Implement with /alfred:2-run SPEC-042
RED: Test agent writes failing tests
↓
GREEN: Implementer agent creates minimal code
↓
REFACTOR: Quality agent improves code
↓
Repeat TDD cycle for each feature component
↓
All tests passing, coverage ≥85%
Step 3: Sync with /alfred:3-sync auto SPEC-042
Updates:
✓ Documentation
✓ Test coverage metrics
✓ Creates PR to develop
✓ Auto-validation of quality gates
Debugging Pattern: Issue → Root Cause → Fix
Step 1: Triage & Understand
Error message: "Cannot read property 'user_id' of undefined"
Questions:
- When does it occur? (always, intermittently, specific scenario)
- Which code path? (which endpoint/function)
- What's the state? (what data led to this)
- What changed recently? (revert to narrow down)
Step 2: Isolate Root Cause
# Method 1: Binary search
# Is it in API layer? → Yes
# Is it in route handler? → No
# Is it in service layer? → Yes
# Is it in this function? → Narrow down
# Method 2: Add logging
logger.debug(f"user_id = {user_id}") # Check where it becomes undefined
# Method 3: Test locally
# Reproduce with minimal example
# Add breakpoint in debugger
# Step through execution
Step 3: Fix with Tests
# RED: Write failing test
def test_handles_missing_user_id():
"""Should handle case when user_id is undefined."""
assert get_user(None) raises ValueError
# GREEN: Minimal fix
def get_user(user_id):
if not user_id:
raise ValueError("user_id required")
return fetch_user(user_id)
# REFACTOR: Improve
def get_user(user_id: int) -> User:
"""Get user by ID.
Args:
user_id: User identifier
Raises:
ValueError: If user_id is None or invalid
"""
if not user_id or user_id <= 0:
raise ValueError(f"Invalid user_id: {user_id}")
return self.user_repo.find(user_id)
5 Real-World Scenarios
Scenario 1: Feature Implementation (2-3 hours)
1. Create SPEC: /alfred:1-plan "Add user dashboard"
2. Clarify details: AskUserQuestion (which data to show?)
3. Implement: /alfred:2-run SPEC-XXX (TDD cycle)
4. Document: /alfred:3-sync auto SPEC-XXX
5. Result: Production-ready feature
Scenario 2: Bug Investigation (1-2 hours)
1. Reproduce: Create minimal test case
2. Isolate: Narrow down affected code
3. Debug: Add logging, trace execution
4. Fix: TDD RED→GREEN→REFACTOR
5. Validate: Ensure tests pass, regression tests
Scenario 3: Large Refactoring (4-8 hours)
1. Analyze: Map current code structure
2. Plan: Design new structure with trade-offs
3. Clone pattern: Create autonomous agents for parallel refactoring
4. Integrate: Verify all pieces work together
5. Test: Comprehensive test coverage
Scenario 4: Performance Optimization (2-4 hours)
1. Profile: Identify bottleneck with profiler
2. Analyze: Understand performance characteristics
3. Design: Plan optimization approach
4. Implement: TDD RED→GREEN→REFACTOR
5. Validate: Benchmark before/after
Scenario 5: Multi-Team Coordination (ongoing)
1. SPEC clarity: AskUserQuestion for ambiguous requirements
2. Agent routing: Delegate to specialist teams
3. Progress tracking: TodoWrite for coordination
4. Integration: Verify components work together
5. Documentation: Central SPEC as source of truth
Context Budget Optimization
Typical project context:
- Config files: ~50 tokens
- .moai/ structure: ~100 tokens
- Entry points (3-5 files): ~500 tokens
- SPEC document: ~200 tokens
→ Total: ~850 tokens per session
Reusable context:
- Load once per session
- Share across 5-10 tasks
- Saves: 3,500-8,500 tokens per session
- Result: More reasoning capacity
Best Practices
DO
- ✅ Load entry points first (3-5 files)
- ✅ Identify dependencies before deep dive
- ✅ Reuse analyzed context across tasks
- ✅ Cache intermediate results in Task context
- ✅ Follow SPEC → TDD → Sync workflow
- ✅ Track progress with TodoWrite
- ✅ Ask for clarification (AskUserQuestion)
- ✅ Test before declaring done
DON'T
- ❌ Load entire codebase at once
- ❌ Reanalyze same code multiple times
- ❌ Skip SPEC clarification (causes rework)
- ❌ Write code without tests
- ❌ Ignore error messages
- ❌ Assume context understanding
- ❌ Skip documentation updates
- ❌ Commit without running tests
Related Skills
moai-alfred-agent-guide(Agent orchestration patterns)moai-alfred-clone-pattern(Complex task delegation)moai-essentials-debug(Debugging techniques)
For detailed workflow examples: reference.md
For real-world scenarios: examples.md
Last Updated: 2025-11-12
Status: Production Ready (Enterprise v4.0.0)
GitHub Repository
Related Skills
sglang
MetaSGLang is a high-performance LLM serving framework that specializes in fast, structured generation for JSON, regex, and agentic workflows using its RadixAttention prefix caching. It delivers significantly faster inference, especially for tasks with repeated prefixes, making it ideal for complex, structured outputs and multi-turn conversations. Choose SGLang over alternatives like vLLM when you need constrained decoding or are building applications with extensive prefix sharing.
evaluating-llms-harness
TestingThis Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.
content-collections
MetaThis skill provides a production-tested setup for Content Collections, a TypeScript-first tool that transforms Markdown/MDX files into type-safe data collections with Zod validation. Use it when building blogs, documentation sites, or content-heavy Vite + React applications to ensure type safety and automatic content validation. It covers everything from Vite plugin configuration and MDX compilation to deployment optimization and schema validation.
llamaguard
OtherLlamaGuard is Meta's 7-8B parameter model for moderating LLM inputs and outputs across six safety categories like violence and hate speech. It offers 94-95% accuracy and can be deployed using vLLM, Hugging Face, or Amazon SageMaker. Use this skill to easily integrate content filtering and safety guardrails into your AI applications.
