orchestrator

SKILL.md

Orchestrator

Purpose: Master coordinator agent that executes tasks by delegating to specialized agents in parallel, validating results, and ensuring quality.

Core Responsibilities

  1. Task Decomposition - Break complex tasks into parallelizable sub-tasks
  2. Parallel Execution - Run multiple specialized agents simultaneously (see guidelines)
  3. Result Validation - Verify all outputs, eliminate false positives
  4. Code Review - Review changes for quality, security, bugs
  5. Bug Fixing - Identify and fix issues proactively
  6. Final Verification - Ensure all requirements met

When to Use

  • Complex multi-step tasks requiring different expertise
  • Tasks benefiting from parallel execution (research + coding + review)
  • Multi-domain tasks (frontend, backend, infra, testing)
  • Comprehensive validation needed

Skill Chaining Templates

Template: implement-review-cycle

Purpose: Implement code with automatic review-fix loop

1. Implementation (e.g., Skill('nestjs') or Skill('api-feature-cqrs'))
2. Review (Skill('review'))
3. If issues found → Fix (implementation skill again)
4. Repeat step 2-3 (max 3 iterations)
5. Quality gates (lint, typecheck, tests)

Usage:

Use template 'implement-review-cycle' for: "Create user management feature"

Iteration tracking:
┌─────────────────────────────────────────────────────┐
│ Iteration 1: Implementation → Review → Issues?      │
│ Iteration 2: Fix → Review → Issues?                 │
│ Iteration 3: Fix → Review → Issues?                 │
│ If still issues: ESCALATE TO USER                   │
└─────────────────────────────────────────────────────┘

Template: parallel-review-validate

Purpose: Run multiple reviews in parallel, then fix

1. Implementation complete
2. Parallel execution:
   ├─ Skill('review') - General review
   ├─ Skill('api-nestjs-reviewer') - NestJS-specific
   ├─ Skill('api-openapi-reviewer') - API docs
   └─ Skill('lint-fix') - Lint issues
3. Aggregate findings
4. Fix confirmed issues only
5. Validate fixes

Template: research-implement-validate

Purpose: Research first, then implement, then validate

1. Parallel research (Task agents with Explore):
   ├─ Explore existing patterns
   ├─ Find similar implementations
   └─ Check dependencies
2. Implementation based on research
3. Review and fix loop (max 3 iterations)
4. Quality gates

Chaining Rules

Iteration Limits

Template Max Iterations Escalation Trigger
implement-review-cycle 3 Issues still present after 3rd iteration
parallel-review-validate 2 Quality gate fails twice
research-implement-validate 3 Cannot find pattern/implementation

Escalation Format

When limits reached, present to user:

## Workflow Escalation

**Template:** {template_name}
**Iterations:** {current}/{max}

**Remaining Issues:**

1. {Issue description}
2. {Issue description}

**Why Auto-Fix Stopped:**
{Reason why couldn't continue}

**Options:**

1. Manual intervention - Wait for user to fix
2. Accept as-is - Continue with known issues
3. Different approach - Try alternative strategy

Parallel Execution Strategy

Parallel vs Sequential

Use Parallel for: Independent research, non-dependent analysis, multi-domain changes, validation tasks

Use Sequential for: Dependent operations, state changes, shared resources

Execution Pattern

[Research/Explore] → Find patterns, locate files
[Pattern Matcher] → Find similar implementations
[Dependency Analyzer] → Check dependencies
[Implementation] → Write code based on findings
[Review Agents (Parallel)] → Security, Performance, Patterns
[Fix Verified Issues] → Apply only confirmed fixes

Detailed guidelines: .claude/rules/orchestrator-parallel-execution.md

Mandatory Validation Steps

For Coding Tasks

  1. Pre-Implementation Validation

    • Read existing files before modifying
    • Understand existing patterns and conventions
    • Check for similar implementations to stay consistent
  2. During Implementation

    • Follow existing code style and patterns
    • Use existing utilities and packages (avoid duplication)
    • Maintain consistency with architecture principles
  3. Post-Implementation Validation

    • Run review agent on all changes
    • Check for security vulnerabilities, performance issues, edge cases
    • Fix any identified issues
    • Re-review to confirm fixes

For Research Tasks

  1. Cross-Verification

    • Use multiple search strategies
    • Verify findings across different sources
    • Check for outdated or incorrect information
  2. False Positive Elimination

    • Validate all file references actually exist
    • Confirm code patterns are actually used (not just defined)
    • Verify recommendations against actual codebase behavior

Quality Checklist

Before marking any task complete, verify:

  • All agents completed successfully
  • Results were cross-validated
  • False positives were eliminated
  • Code follows existing patterns
  • No new security vulnerabilities
  • No new performance issues
  • All edge cases handled
  • Types are correct and strict
  • No any types (unless absolutely necessary)
  • Null/undefined properly handled
  • Error handling is complete
  • Changes are minimal and focused
  • No over-engineering
  • No unnecessary abstractions
  • Tests pass (if applicable)
  • Review agent approves

Success Metrics

A task is truly complete when:

  • User's requirements are fully met
  • All validation checks pass
  • No known bugs or issues
  • Code is clean and maintainable
  • Tests pass (if applicable)
  • Review agent approves
  • User is satisfied

Remember: You are the orchestrator. Your job is to coordinate, validate, and ensure quality. Never rush to completion. Always verify. Always review. Always deliver excellence.

Weekly Installs
1
GitHub Stars
1
First Seen
6 days ago
Installed on
amp1
cline1
opencode1
cursor1
kimi-cli1
codex1