team-implement-plan-full
Team Implement Plan Full (Large Team)
Overview
This skill executes implementation plans with maximum parallelism. Each phase gets a dedicated implementer teammate. Independent phases run simultaneously within waves. A shared Reviewer provides cross-phase quality checks.
Team composition:
- phase-N-impl (one per phase in current wave): Implements their phase directly
- reviewer: Shared across all phases. Reviews FIFO as phases complete. Checks cross-phase consistency
- Lead: Orchestrates waves, manages team lifecycle, handles plan sync and completion
When to use this:
- Plans with 4+ phases where some phases are independent
- Large features spanning multiple modules/layers
- When speed of execution matters and you can afford the token cost
Token cost: ~100-150K per wave (scales with concurrent implementers)
References:
references/team-lifecycle.mdfor team lifecycle patternreferences/quality-pipeline-distribution.mdfor pipeline distribution
Initial Response
When invoked with a plan path:
"I'll set up a full implementation team with parallel execution. Let me read the plan, analyze the dependency graph, and present the execution waves before starting."
Workflow
Phase 1: Plan Analysis
Step 1a: Read and validate the plan
Read($0) # Plan path from argument
Validate:
- Implementation phases with objectives and tasks
- Exit conditions per phase
- Dependencies between phases (from TaskList or plan structure)
Step 1b: Check existing progress
TaskList # Check for existing tasks
If tasks exist with some completed, resume from current state (skip completed phases, adjust waves).
Step 1c: Build dependency graph
Parse phase dependencies to build execution waves:
Algorithm:
1. Read all tasks and their blockedBy relationships
2. Wave 1 = phases with no dependencies (blockedBy is empty)
3. Wave 2 = phases whose ALL dependencies are in Wave 1
4. Wave N = phases whose ALL dependencies are in Waves 1..N-1
5. Apply file conflict detection (Step 1d)
Step 1d: Detect file conflicts
For phases in the same wave, check if they touch overlapping files:
For each pair of phases in the same wave:
- Extract file scope from plan tasks (files mentioned)
- If scopes overlap: add synthetic dependency, move one phase to next wave
- Priority: keep the phase with more downstream dependents in earlier wave
Shared files that multiple phases need to modify (index files, module registrations, route configs) are handled by the Lead after a wave completes, not by individual implementers.
Step 1e: Build file ownership map
file_ownership = {}
For each wave:
For each phase in wave:
For each file in phase scope:
file_ownership[file] = phase_id
shared_files = files appearing in multiple phases across waves
Shared files are excluded from all implementer scopes. Lead handles them at wave boundaries.
Step 1f: Present wave plan to user
## Execution Plan
### Wave 1 (parallel)
- Phase 1: [Name] — files: [scope]
- Phase 3: [Name] — files: [scope]
### Wave 2 (after Wave 1 completes)
- Phase 2: [Name] — depends on Phase 1 — files: [scope]
### Wave 3 (after Wave 2 completes)
- Phase 4: [Name] — depends on Phase 2 — files: [scope]
- Phase 5: [Name] — depends on Phase 2 — files: [scope]
### Shared files (handled by Lead at wave boundaries)
- src/app.module.ts (modified by Phase 1, 2, 4)
- src/routes/index.ts (modified by Phase 3, 5)
Estimated team size: [max wave size + 1 reviewer]
Estimated token cost: ~[cost estimate]
Proceed?
Wait for user confirmation before starting.
Phase 2: Team Creation
Step 2a: Create team
TeamCreate(team_name="impl-full-{plan-slug}")
Step 2b: Spawn Reviewer
The Reviewer is spawned once and persists across all waves.
Task(subagent_type="general-purpose",
team_name="impl-full-{plan-slug}",
name="reviewer",
prompt="You are the cross-phase Reviewer on a full implementation team.
PLAN: {full plan content}
YOUR ROLE: Quality gate for ALL phases. You review each phase as implementers complete them. You are the only reviewer — process requests FIFO.
REVIEW PROTOCOL:
When an implementer messages you 'Phase N ready for review':
1. READ all changed files completely
2. RUN exit condition commands independently:
- Build: {build commands from plan}
- Lint: {lint commands}
- Test: {test commands}
3. CHECK code quality:
- Does the code follow existing patterns?
- Any security issues?
- Error handling adequate?
- Tests meaningful?
- Design decision compliance?
4. CHECK cross-phase consistency:
- Do types/interfaces match across phases?
- Are shared contracts maintained?
- Any naming conflicts with other completed phases?
5. CHECK ADR compliance:
- Read docs/decisions/INDEX.md
- Verify against relevant ADRs
DECISION:
- PASS: Message team lead: 'PASS: Phase N — [quality summary]'
- NEEDS_CHANGES: Message the specific implementer: 'NEEDS_CHANGES: [issue list with file:line refs]'
Wait for re-request, then re-review
RULES:
- You are read-only — NEVER modify code. Only implementers write code
- Process reviews FIFO — first come, first served
- Be thorough but pragmatic — real issues only, not style nitpicks
- Always include file:line references
- If you notice cross-phase conflicts, message the team lead immediately")
Phase 3: Wave Execution Loop
For each wave:
Step 3a: Spawn wave implementers
For each phase in the current wave, spawn a dedicated implementer:
Task(subagent_type="general-purpose",
team_name="impl-full-{plan-slug}",
name="phase-{N}-impl",
prompt="You are the implementer for Phase {N} on a full implementation team.
PLAN CONTEXT: {relevant plan sections}
YOUR PHASE: {phase N details — objective, tasks, exit conditions}
YOUR FILE SCOPE — you may ONLY modify these files:
{file list from ownership map}
DO NOT MODIFY these shared files (the Lead handles them):
{shared file list}
IMPLEMENTATION PROTOCOL:
1. Read all files in your scope and relevant context files
2. Implement phase tasks IN ORDER (tests first, then implementation)
3. Run all exit condition commands:
- Build: {build commands}
- Lint: {lint commands}
- Test: {test commands}
4. Fix any failures — iterate until all exit conditions pass
5. When ALL exit conditions pass, message 'reviewer':
'Phase {N} ready for review. Files changed: [list]. Exit conditions passing.'
6. If reviewer sends NEEDS_CHANGES, fix the issues and re-request review
7. Do NOT start any other phase — your scope is Phase {N} only
RULES:
- Stay within your file scope — do NOT modify files outside it
- If you need to change a shared file, message the team lead with what change is needed
- If you need information from another phase's files, READ them but do not WRITE
- If you hit a blocker, message the team lead
- Follow existing codebase patterns documented in the plan")
Step 3b: Monitor wave progress
While wave executes:
- Track which implementers have messaged Reviewer via TaskList and messages
- Track which phases have passed review
- If an implementer goes idle without progress, message them for status
- If an implementer reports a blocker, help resolve it or escalate to user
- Relay cross-phase information when relevant (e.g., "Phase 1 created a new type at src/types.ts:42 that Phase 3 may need")
Step 3c: Handle shared file requests
When implementers message that they need a shared file modified:
- Collect all shared file change requests for the current wave
- After ALL wave phases pass review, apply shared file changes yourself (the Lead directly edits)
- Run full build/test to verify shared file changes don't break anything
- If tests fail, fix and re-verify
Step 3d: Wave review tracking
Track review status per phase:
Wave 1 Review Status:
Phase 1: ✅ PASS
Phase 3: 🔄 In review (attempt 2)
Phase 5: ⏳ Awaiting review
Step 3e: Handle fix loops
If a fix loop exceeds 3 iterations for any phase:
- Read the disputed issues
- Determine if the issue is genuine or a disagreement
- Make a judgment call and message both the implementer and reviewer
- If genuinely stuck, ask the user for input
Step 3f: Wave completion
When ALL phases in the wave have passed review:
- Apply shared file changes (Step 3c)
- Run full test suite to verify cross-phase integration
- If integration tests fail:
- Identify which phase's changes caused the failure
- Message that phase's implementer with the fix needed
- Re-verify after fix
- Shut down wave implementers (they completed their one phase)
- Update task status for all wave phases (TaskUpdate to completed)
- Generate wave completion report
Step 3g: User confirmation between waves
Wave {W} complete.
Phases completed: {list with summaries}
Files changed: {aggregated list}
Issues caught by Reviewer: {summary}
Integration status: {pass/issues}
Next wave: {phase list}
Continue? (or /clear and resume later — progress is saved)
Wait for user confirmation before spawning next wave.
Step 3h: Reviewer health check
After each wave:
- If Reviewer has reviewed 4+ phases, check for quality degradation
- Signs of degradation: reviews getting shorter, missing obvious issues, slow responses
- If degraded: shut down Reviewer, spawn replacement with same prompt
Phase 4: Plan Completion
After all waves complete:
Step 4a: Final integration check
Run the full test suite one final time. All exit conditions from all phases must still pass.
Step 4b: Completion report
## Implementation Complete
**Plan**: {plan name}
**Method**: Full team (parallel wave execution)
**Waves executed**: {count}
**Total phases**: {count}
**Phases per wave**: {breakdown}
### Quality Summary
- Issues caught by Reviewer: {count}
- Fix loops required: {count}
- Cross-phase conflicts resolved: {count}
- Shared file changes: {count}
### Per-Phase Summary
| Phase | Wave | Implementer | Review Attempts | Key Changes |
|-------|------|-------------|-----------------|-------------|
| 1 | 1 | phase-1-impl | 1 | {summary} |
| 2 | 2 | phase-2-impl | 2 | {summary} |
...
### Files Changed (all phases)
{aggregated file list}
Step 4c: Shutdown and cleanup
- Shut down Reviewer
- TeamDelete
- Present final report to user
- Suggest: "Run
/e2e-testingto validate the full implementation"
Crash Recovery Protocol
If the session ends mid-wave:
- On next session, user invokes
/team-implement-plan-full [plan-path] - Lead reads plan and checks TaskList
- Completed phases (task status = completed): Skip entirely
- Completed waves: Skip entirely
- In-progress wave: Check which phases in the wave are completed vs pending
- Completed phases in the wave: Skip
- In-progress/pending phases: Re-create team, spawn implementers only for these
- Reviewer is re-spawned fresh (stateless)
- Re-run dependency analysis to determine correct wave structure from remaining phases
What persists: Task status, committed code, plan file What's lost: Teammate context, uncommitted changes, review history
Mitigation: Lead encourages committing after each wave completes.
File Conflict Prevention
Rules
- Each implementer has an explicit file scope — listed in their spawn prompt
- Implementers MUST NOT modify files outside their scope — stated as a rule in prompt
- Shared files are Lead-managed — identified before execution, excluded from all scopes
- Same-wave overlap = synthetic dependency — detected in Step 1d, prevents parallel execution
- Read is always allowed — implementers can READ any file for context, just not WRITE outside scope
Shared File Handling
Common shared files:
- Module registration files (
app.module.ts,main.py) - Route index files (
routes/index.ts) - Type/interface barrel exports (
types/index.ts) - Configuration files (
config/*.ts) - Package manifests (
package.json— for new dependencies)
Protocol:
- Implementer messages Lead: "Phase N needs to add [import/route/provider] to [shared file]"
- Lead collects all requests for the wave
- After wave passes review, Lead applies all shared file changes at once
- Lead runs full build/test to verify
- This prevents merge conflicts and ensures consistency
Wave Size Limits
| Scenario | Max Concurrent Implementers | Rationale |
|---|---|---|
| Default | 4 | Balance between parallelism and coordination overhead |
| Simple phases (< 3 tasks each) | 5 | Lower per-phase complexity allows more concurrency |
| Complex phases (5+ tasks each) | 3 | Higher per-phase complexity needs more lead attention |
| User override | As specified | User can request specific wave sizes |
The Reviewer is always 1 (shared across wave). Total team size = wave implementers + 1 reviewer + lead.
Teammate Stability
Implementer Lifecycle
- Each implementer handles exactly ONE phase, then is shut down
- This prevents context degradation from long sessions
- Fresh implementer per phase = consistent quality
Reviewer Lifecycle
- Single reviewer persists across waves
- Monitor for degradation after 4-5 reviews
- Replacement protocol: shut down, spawn fresh reviewer with same prompt
- Signs of degradation: reviews getting superficial, missing issues previously caught, slow responses
Lead Responsibilities
- Monitor all teammates for responsiveness
- 2+ messages without response = teammate is stuck, spawn replacement
- Track review quality — if reviewer approves code that then fails integration, reviewer may need replacement
Quality Pipeline Distribution
| Pipeline Step | Owner | Notes |
|---|---|---|
| 1. Implementation | phase-N-impl | Direct code writing within file scope |
| 2. Verification-loop | phase-N-impl | Build/lint/test scoped to phase |
| 3. Integration testing | phase-N-impl | Tests within phase scope |
| 4. Code review | reviewer | Cross-phase consistency + per-phase quality |
| 5. ADR compliance | reviewer | Centralized ADR knowledge |
| 6. Plan sync | Lead | Single writer to task status |
| 7. Prompt archival | Lead | Single writer to prompt files |
| 8. Completion report | Lead | Aggregates per wave and final |
Quality Checklist
Before completing each wave:
- All wave phases passed Reviewer review
- Shared file changes applied and verified
- Full test suite passes (cross-phase integration)
- Task status updated for all wave phases
- User confirmed wave completion
Before completing the plan:
- All waves completed
- Final full test suite passes
- All tasks marked completed
- Completion report generated
- All teammates shut down
- Team cleaned up via TeamDelete
More from mhylle/claude-skills-collection
skill-visualizer
Generate interactive HTML visualizations of the skills collection, codebase structure, or dependency graphs. Uses D3.js for interactive visualization with collapsible nodes, color-coded categories, and hover details. Triggers on "visualize skills", "generate skill map", "codebase visualization", or "show skill dependencies".
9verification-loop
Comprehensive 6-check verification framework for validating implementation quality across build, types, lint, tests, security, and diff review. This skill ensures code meets all quality gates before phase completion. Triggers on "verify implementation", "run verification", "/verification-loop", or automatically as part of implement-phase Step 2.
8context-saver
Save session context to disk for seamless continuation in new chat sessions. This skill should be used when the user asks to save context, preserve work state, checkpoint progress, or prepare for session handoff. Triggers on "save context", "checkpoint", "save progress", "preserve state", or when explicitly asked to create a context file for later resumption. Optimizes for correctness, completeness, minimal size, and trajectory preservation.
8strategic-compact
Strategic compaction suggestion framework that monitors session complexity and suggests context compaction at optimal logical boundaries rather than arbitrary thresholds.
8implement-plan
Orchestrate the execution of complete implementation plans, delegating each phase to implement-phase skill. This skill manages the full plan lifecycle including phase sequencing, user confirmation between phases, and overall progress tracking. Triggers on "implement the plan", "execute the implementation plan", or when given a path to a plan file.
8create-plan
Create detailed implementation plans through interactive research and iteration. This skill should be used when creating new implementation plans, designing feature specifications, planning technical work, or when the user asks to plan an implementation. Triggers on requests like "create a plan", "plan the implementation", "design how to implement", or when given a feature/task that needs structured planning before implementation.
7