Back to Skills

parallel-agent-dispatch

romiluz13
Updated Today
23 views
15
3
15
View on GitHub
Otherai

About

This skill enables parallel dispatch of bug-investigator subagents to handle multiple independent failures concurrently. Use it when facing 3+ unrelated bugs that can be investigated without shared state or dependencies. The orchestrator manages multiple subagents working simultaneously to resolve independent problems faster than sequential processing.

Documentation

Parallel Agent Dispatch - Orchestrator-Driven

Overview

When you have multiple unrelated failures (different test files, different subsystems, different bugs), investigating them sequentially wastes time. Each investigation is independent and can happen in parallel.

Core principle: Orchestrator dispatches one bug-investigator subagent per independent problem domain. Let them work concurrently.

Integration with cc10x: This skill is used by the DEBUG workflow orchestrator to coordinate parallel subagent dispatch for independent bugs.

Quick Start

Dispatch multiple subagents in parallel when facing independent failures.

Example:

  1. Detect 3+ failures: Test files A, B, C failing with different errors
  2. Check independence: Different files, no shared state → Independent
  3. Dispatch subagents: bug-investigator-A, bug-investigator-B, bug-investigator-C (parallel)
  4. Collect results: All subagents complete independently, fixes applied

Result: Multiple independent bugs fixed concurrently, faster resolution.

When to Use

Use when:

  • 3+ test files failing with different root causes
  • Multiple subsystems broken independently
  • Each problem can be understood without context from others
  • No shared state between investigations
  • No dependencies between bugs

Don't use when:

  • Failures are related (fix one might fix others)
  • Need to understand full system state
  • Bugs share files or resources
  • Bugs have dependencies (fixing one requires fixing another first)

Independence Detection

Orchestrator analyzes bugs for independence:

  1. File Overlap Check: Do bugs touch different files?

    • Different files → Likely independent
    • Same files → Check further
  2. Resource Overlap Check: Do bugs use different resources?

    • Different databases, APIs, services → Likely independent
    • Same resources → Check further
  3. Dependency Check: Does fixing one bug require fixing another?

    • No dependencies → Independent
    • Has dependencies → Sequential required
  4. State Check: Do bugs modify shared state?

    • No shared state → Independent
    • Shared state → Sequential required

Decision Logic:

  • All checks pass → Independent → Parallel dispatch
  • Any check fails → Dependent → Sequential execution

The Pattern

1. Identify Independent Domains

Group failures by what's broken:

  • File A tests: Tool approval flow
  • File B tests: Batch completion behavior
  • File C tests: Abort functionality

Each domain is independent - fixing tool approval doesn't affect abort tests.

2. Create Focused Subagent Tasks

Each bug-investigator subagent gets:

  • Specific scope: One test file or subsystem
  • Clear goal: Make these tests pass
  • Constraints: Don't change other code
  • Expected output: Summary of what you found and fixed

3. Orchestrator Dispatches in Parallel

Orchestrator coordinates:

  • Dispatch bug-investigator subagent for each independent bug
  • All subagents run concurrently
  • Orchestrator tracks progress for each subagent
  • Orchestrator waits for all to complete

Example:

Orchestrator detects 3 independent bugs:
- Bug 1: agent-tool-abort.test.ts failures
- Bug 2: batch-completion-behavior.test.ts failures
- Bug 3: tool-approval-race-conditions.test.ts failures

Orchestrator dispatches:
- bug-investigator → Bug 1 (parallel)
- bug-investigator → Bug 2 (parallel)
- bug-investigator → Bug 3 (parallel)

All three run concurrently, orchestrator coordinates.

4. Review and Integrate

When subagents return:

  • Orchestrator collects summaries from all subagents
  • Verify fixes don't conflict
  • Run full test suite
  • Integrate all changes

Subagent Prompt Structure

Good subagent prompts are:

  1. Focused - One clear problem domain
  2. Self-contained - All context needed to understand the problem
  3. Specific about output - What should the subagent return?

Example:

Fix the 3 failing tests in src/agents/agent-tool-abort.test.ts:

1. "should abort tool with partial output capture" - expects 'interrupted at' in message
2. "should handle mixed completed and aborted tools" - fast tool aborted instead of completed
3. "should properly track pendingToolCount" - expects 3 results but gets 0

These are timing/race condition issues. Your task:

1. Read the test file and understand what each test verifies
2. Identify root cause - timing issues or actual bugs?
3. Fix by:
   - Replacing arbitrary timeouts with event-based waiting
   - Fixing bugs in abort implementation if found
   - Adjusting test expectations if testing changed behavior

Do NOT just increase timeouts - find the real issue.

Return: Summary of what you found and what you fixed.

Common Mistakes

❌ Too broad: "Fix all the tests" - subagent gets lost ✅ Specific: "Fix agent-tool-abort.test.ts" - focused scope

❌ No context: "Fix the race condition" - subagent doesn't know where ✅ Context: Paste the error messages and test names

❌ No constraints: Subagent might refactor everything ✅ Constraints: "Do NOT change production code" or "Fix tests only"

❌ Vague output: "Fix it" - you don't know what changed ✅ Specific: "Return summary of root cause and changes"

When NOT to Use

Related failures: Fixing one might fix others - investigate together first Need full context: Understanding requires seeing entire system Exploratory debugging: You don't know what's broken yet Shared state: Subagents would interfere (editing same files, using same resources) Dependencies: Fixing one bug requires fixing another first

Integration with DEBUG Workflow

Orchestrator Integration:

  1. Bug Classification: DEBUG workflow classifies bugs (reproducible/intermittent/independent)
  2. Independence Analysis: Orchestrator analyzes bugs for independence using this skill
  3. Parallel Dispatch: If independent → Orchestrator dispatches bug-investigator subagents in parallel
  4. Sequential Fallback: If dependent → Orchestrator executes sequentially (existing behavior)
  5. Coordination: Orchestrator tracks all subagents, collects results, verifies no conflicts

Troubleshooting

Common Issues:

  1. Bugs not independent but dispatched in parallel

    • Symptom: Subagents interfere with each other or conflicts occur
    • Cause: Independence check failed or bugs share files/resources
    • Fix: Re-check independence, dispatch sequentially if dependent
    • Prevention: Always verify independence before parallel dispatch
  2. Too few bugs for parallel dispatch

    • Symptom: Only 1-2 bugs but parallel dispatch attempted
    • Cause: Didn't check bug count (needs 3+)
    • Fix: Use sequential dispatch for <3 bugs
    • Prevention: Always check bug count before parallel dispatch
  3. Subagents not coordinated

    • Symptom: Results not collected or conflicts not detected
    • Cause: Orchestrator didn't coordinate properly
    • Fix: Ensure orchestrator tracks all subagents, collects results
    • Prevention: Always coordinate subagents through orchestrator

If issues persist:

  • Verify bugs are independent (different files, no shared state)
  • Check that bug count is 3+ for parallel dispatch
  • Ensure orchestrator coordinates all subagents
  • Review independence detection section

Workflow:

DEBUG Workflow:
├─ Phase 1: Bug Classification
├─ Phase 2: Independence Analysis (uses this skill)
│   ├─ Independent? → Parallel dispatch
│   └─ Dependent? → Sequential execution
├─ Phase 3: Subagent Coordination
│   ├─ Parallel: Orchestrator dispatches multiple bug-investigator subagents
│   └─ Sequential: Orchestrator dispatches bug-investigator subagent one at a time
└─ Phase 4: Integration and Verification

Key Benefits

  1. Parallelization - Multiple investigations happen simultaneously
  2. Focus - Each subagent has narrow scope, less context to track
  3. Independence - Subagents don't interfere with each other
  4. Speed - 3 problems solved in time of 1
  5. Orchestrator Coordination - Centralized management of parallel execution

Verification

After subagents return:

  1. Review each summary - Understand what changed
  2. Check for conflicts - Did subagents edit same code?
  3. Run full suite - Verify all fixes work together
  4. Spot check - Subagents can make systematic errors

Real-World Impact

From debugging sessions:

  • 6 failures across 3 files
  • 3 bug-investigator subagents dispatched in parallel
  • All investigations completed concurrently
  • All fixes integrated successfully
  • Zero conflicts between subagent changes

Summary

For Orchestrator:

  • Analyze bugs for independence
  • If independent → Dispatch bug-investigator subagents in parallel
  • If dependent → Execute sequentially
  • Coordinate all subagents, collect results, verify no conflicts

For Subagents:

  • Focus on one problem domain
  • Work independently
  • Return summary of findings and fixes

Result: Faster debugging for multiple independent bugs, orchestrator manages coordination.

Quick Install

/plugin add https://github.com/romiluz13/cc10x/tree/main/parallel-agent-dispatch

Copy and paste this command in Claude Code to install this skill

GitHub 仓库

romiluz13/cc10x
Path: plugins/cc10x/skills/parallel-agent-dispatch

Related Skills

sglang

Meta

SGLang is a high-performance LLM serving framework that specializes in fast, structured generation for JSON, regex, and agentic workflows using its RadixAttention prefix caching. It delivers significantly faster inference, especially for tasks with repeated prefixes, making it ideal for complex, structured outputs and multi-turn conversations. Choose SGLang over alternatives like vLLM when you need constrained decoding or are building applications with extensive prefix sharing.

View skill

evaluating-llms-harness

Testing

This Claude Skill runs the lm-evaluation-harness to benchmark LLMs across 60+ standardized academic tasks like MMLU and GSM8K. It's designed for developers to compare model quality, track training progress, or report academic results. The tool supports various backends including HuggingFace and vLLM models.

View skill

llamaguard

Other

LlamaGuard is Meta's 7-8B parameter model for moderating LLM inputs and outputs across six safety categories like violence and hate speech. It offers 94-95% accuracy and can be deployed using vLLM, Hugging Face, or Amazon SageMaker. Use this skill to easily integrate content filtering and safety guardrails into your AI applications.

View skill

langchain

Meta

LangChain is a framework for building LLM applications using agents, chains, and RAG pipelines. It supports multiple LLM providers, offers 500+ integrations, and includes features like tool calling and memory management. Use it for rapid prototyping and deploying production systems like chatbots, autonomous agents, and question-answering services.

View skill