implementation-plan-review
Implementation Plan Review
This skill reviews generated implementation plans against established best practices to ensure quality, completeness, and development readiness.
Prerequisites
- Generated
milestones.yamlfile - Associated
milestone-m*.tasks.yamlfiles - Optional: Source business-requirements.yaml and technical-requirements.yaml for alignment checks
Review Process
Phase 1: Structure Validation
Load and parse all plan files:
-
File Discovery
- Locate
milestones.yaml - Find all
milestone-m*.tasks.yamlfiles - Verify all referenced task files exist
- Locate
-
Schema Validation
- Validate YAML syntax
- Check required fields present
- Verify data types and formats
-
Reference Integrity
- Validate milestone dependencies exist
- Validate task dependencies exist
- Check task file references match milestones
Phase 2: Best Practices Compliance
Review against embedded best practices:
Style Anchors
- 2-3 style anchors included per milestone task file
- Concrete file paths (not abstract descriptions)
- Specific line numbers provided
- Complete examples (code + tests + README where applicable)
- Current patterns reflecting best practices
- Clear descriptions explaining what pattern is demonstrated
Task Sizing (from task-sizing.md)
- Minimum 30 minutes per task (document rationale if smaller)
- Optimal 30-150 minutes for most tasks
- Maximum 150 minutes (2.5 hours) - split if larger
- Limited file scope (1-3 files preferred, >5 requires justification)
- Clear deliverables for each task
- Acceptance tests defined
TDD Requirements
- TDD checklist included for code tasks
- Test-first approach specified
- Validation commands provided
- Expected outputs documented
- Test modification forbidden when tests fail
Affirmative Instructions
- Permitted actions explicit (e.g., "ONLY use: X, Y, Z")
- Negative framing avoided
- Constraints clear and actionable
- File scopes explicit ("touch ONLY", "modify ONLY")
Drift Prevention (from drift-policy.md)
- Drift policy stated in task instructions
- Stop criteria clear (unexpected dependencies, file scope violations)
- Revert instructions provided
- Incident documentation process specified
- Allowed deviations understood (formatting, whitespace)
Quality Gates
- Pre-commit hooks defined
- Validation commands specified per task
- Success criteria measurable
- Failure handling documented
Prompt Positioning
- Critical specs at beginning of task instructions
- Style anchors early in task definitions
- Hard rules reiterated at end
- Requirements not buried in middle
Phase 3: Drift Policy Compliance
Review against embedded drift policy:
Stop & Revert Criteria Coverage
Check that tasks include guidance for:
- New dependencies detection - How to identify unauthorized dependencies
- File scope violations - What to do if >3 unexpected files touched
- Linting/type errors - Handling unresolvable errors
- Test failures - Policy on modifying vs fixing implementation
Immediate Actions Documentation
- Stop instruction clear - "STOP immediately" language used
- Error message format - How to summarize deviation
- Revert process - When and how to revert
- Incident note location - Where to document (docs/drift-incidents/)
Recording Learnings
- Rule update process mentioned
- Where to update (.cursor/rules/, CLAUDE.md)
Allowed Deviations
- Minor formatting - Understood as acceptable
- Whitespace edits - Allowed without review
- Single-line refactors - Within scope and type-checked
Phase 4: Task Sizing Analysis
Review against embedded task sizing guidelines:
Duration Compliance
Generate report:
task_sizing_report:
total_tasks: [count]
by_duration:
too_small: # < 30m
- task_id: [id]
estimate_minutes: [n]
rationale: [why small]
recommendation: [merge or accept]
optimal: # 30-150m
count: [n]
percentage: [%]
too_large: # > 150m
- task_id: [id]
estimate_minutes: [n]
recommendation: [split into X, Y, Z]
File Scope Analysis
- 1-3 files preferred
- >5 files justified or split
- Explicit file lists in each task
- Touch-only files clearly marked
Task Decomposition Quality
- Smallest useful deliverable defined
- Clear acceptance test per task
- Example files provided
- Commit after each task strategy
Large Task Detection
Identify tasks requiring decomposition:
decomposition_needed:
- task_id: [id]
current_estimate: [minutes]
suggested_split:
- name: [subtask 1]
estimate: [minutes]
- name: [subtask 2]
estimate: [minutes]
- name: [subtask 3]
estimate: [minutes]
Phase 5: Completeness Check
Milestone Structure
- All milestones have unique IDs
- Names are clear and descriptive
- Descriptions explain deliverables
- Dependencies correctly ordered
- No circular dependencies
- Estimated durations realistic
- Task files referenced correctly
- Success criteria testable
Task Breakdown
- All tasks have unique IDs
- Task names descriptive
- Descriptions clear and actionable
- Estimates follow sizing rules
- Types specified (code/test/docs/config)
- Dependencies explicit and correct
- File lists complete and accurate
- Instructions detailed and clear
Task Quality
- Objectives clearly stated
- Implementation steps numbered
- Constraints explicit
- TDD checklists included (code tasks)
- Validation commands provided
- Expected outputs documented
- Drift policy stated
Phase 6: Alignment Check
Business Requirements (if provided)
- All functional requirements mapped to tasks
- Priority aligns with business priorities
- MVP scope clearly defined
- Success criteria can verify business goals
Generate coverage matrix:
business_requirement_coverage:
FR-1: [task-ids]
FR-2: [task-ids]
FR-3: [gap] # Not covered
Technical Requirements (if provided)
- Architecture decisions reflected
- Technology choices enforced
- Testing strategy implemented
- Security requirements addressed
- Performance requirements considered
Generate coverage matrix:
technical_requirement_coverage:
architecture:
- decision: [description]
enforced_in: [task-ids]
technology:
- choice: [description]
enforced_in: [task-ids]
Phase 7: Dependency Analysis
Critical Path
- Critical path identified
- Duration calculated
- Parallel opportunities noted
- Unnecessary dependencies removed
- Dependencies minimal but sufficient
Generate dependency graph:
dependency_analysis:
critical_path:
- [milestone-id]
- [milestone-id]
critical_path_duration: [time]
parallel_opportunities:
- can_parallelize: [[m1, m2], [m3, m4]]
time_saved: [duration]
dependency_issues:
- task: [id]
issue: [description]
recommendation: [fix]
Circular Dependency Check
- No circular dependencies in milestones
- No circular dependencies in tasks
- Dependency graph is a DAG
Phase 8: Gap Identification
Common Gaps
- Infrastructure tasks present (setup, tooling)
- Test tasks included
- Documentation tasks included
- Error handling tasks included
- Configuration tasks included
- Deployment/release tasks included
- Assumptions documented
- Style anchors for new patterns
- Clear file boundaries
- Integration tasks between components
Generate gap report:
gaps:
missing_task_types:
- type: [infrastructure/tests/docs/etc]
importance: [high/medium/low]
recommendation: [description]
missing_requirements_coverage:
- requirement: [id]
category: [functional/non-functional]
impact: [high/medium/low]
undocumented_assumptions:
- task: [id]
assumption: [description]
recommendation: [document explicitly]
Phase 9: Style Anchor Validation
- Anchors reference existing files (if repository available)
- Line numbers accurate (if verifiable)
- Patterns demonstrated correctly
- Descriptions helpful
- Sufficient anchors for each new pattern
- Test anchors included
- README anchors included where applicable
Phase 10: Quality Gate Validation
- Linter commands specified
- Type checker commands specified
- Test commands specified
- Commands are executable
- Expected outputs defined
- Failure handling documented
Review Output
Generate comprehensive review report:
implementation_plan_review:
metadata:
plan_files:
- milestones.yaml
- milestone-m0.tasks.yaml
- milestone-m1.tasks.yaml
review_date: [timestamp]
reviewer: implementation-plan-review skill
overall_scores:
structure: [1-10]
best_practices_compliance: [1-10]
task_sizing: [1-10]
completeness: [1-10]
alignment: [1-10]
development_readiness: [1-10]
summary:
total_milestones: [count]
total_tasks: [count]
estimated_total_duration: [time]
critical_path_duration: [time]
task_sizing_summary:
too_small: [count]
optimal: [count]
too_large: [count]
compliance_summary:
style_anchors_present: [count/total]
tdd_checklists_present: [count/total]
drift_policies_present: [count/total]
quality_gates_defined: [yes/no/partial]
critical_issues:
- category: [category]
severity: critical
issue: [description]
affected_items: [list]
recommendation: [how to fix]
blocks_development: [true/false]
high_priority_issues:
- category: [category]
severity: high
issue: [description]
affected_items: [list]
recommendation: [how to fix]
medium_priority_issues:
- category: [category]
severity: medium
issue: [description]
recommendation: [how to fix]
best_practices_violations:
style_anchors:
- issue: [description]
affected_tasks: [list]
guidance: [reference to best practice]
task_sizing:
- issue: [description]
affected_tasks: [list]
guidance: [reference to sizing rules]
tdd_requirements:
- issue: [description]
affected_tasks: [list]
guidance: [reference to TDD practices]
drift_prevention:
- issue: [description]
affected_tasks: [list]
guidance: [reference to drift policy]
quality_gates:
- issue: [description]
affected_milestones: [list]
guidance: [reference to quality constraints]
gaps:
missing_coverage:
requirements: [list of uncovered requirements]
task_types: [list of missing task types]
missing_elements:
- element: [description]
importance: [high/medium/low]
recommendation: [how to add]
strong_areas:
- [what's well-planned]
- [what follows best practices well]
optimization_opportunities:
- type: [parallelization/merging/splitting]
description: [opportunity]
benefit: [time saved / quality improvement]
affected_items: [list]
specific_recommendations:
- priority: [critical/high/medium/low]
category: [category]
recommendation: [specific action]
rationale: [why this matters]
readiness_assessment:
ready_for_development: [yes/no/with-modifications]
blocking_issues: [count]
recommended_modifications: [count]
estimated_effort_to_fix:
critical: [time]
high: [time]
medium: [time]
timeline_impact:
if_fixed: [time]
if_not_fixed: [time]
detailed_findings:
structure_validation: [detailed report]
best_practices_compliance: [detailed report]
task_sizing_analysis: [detailed report]
drift_policy_compliance: [detailed report]
dependency_analysis: [detailed report]
coverage_analysis: [detailed report]
Usage
Review an implementation plan:
/implementation-plan-review path/to/milestones.yaml
The skill will:
- Load all milestone and task files
- Validate structure and schema
- Check compliance against all best practices
- Analyze task sizing and dependencies
- Identify gaps and issues
- Generate comprehensive review report
- Provide specific recommendations
Output Files
Generates:
implementation-plan-review.yaml- Comprehensive review reporttask-sizing-analysis.yaml- Detailed task sizing breakdownrequirement-coverage.yaml- Requirements to tasks mappingdependency-graph.yaml- Milestone and task dependencies
Best Practice References
This skill embeds and enforces best practices directly:
1. Implementation Plan Best Practices
Core Principle:
Models optimize locally; enforce global constraints with layered verification (prompt → IDE → commit → CI → runtime).
Key Practices:
- Style Anchors: Always include 2-3 exemplary files as templates with exact paths and line numbers. Prefer concrete repository examples (code + tests + README). Enforce anchors early to prevent architectural drift.
- Task Sizing: Split work into 30m–2.5h atomic tasks (30–150 minutes optimal). Limit scope to specific files. Commit after each small task. Revert immediately on drift. If a task is shorter than 30m, either increase estimate or split it with rationale.
- Affirmative Instructions: State permitted actions explicitly (e.g., "ONLY use: X, Y, Z"). Avoid negative framing.
- Tiered Rules: Global (user prefs), Project (persistent in CLAUDE.md or .cursor/rules/), Context-aware (per directory/file).
- TDD as Anchor: Require TDD checklist before implementation (tests → minimal code → more tests → refactor). When tests fail, return failing output with instruction: "Revise implementation to pass this test while keeping all previously passing tests. Do not modify the test. Do not add dependencies."
- Drift Handling: Stop and revert immediately on unexpected dependencies or unfamiliar patterns. Do not fix mid-stream. Document learnings and update persistent rules after each session.
- Prompt Positioning: Put critical specs, style anchors, and hard rules at the beginning and reiterate at the end of prompts (avoid burying requirements in the middle).
- External Data Validation: Never use type assertions on external data. Validate all external inputs with proper error handling and validation.
IDE/Linting:
- Enforce strict linting (go vet, golangci-lint, gofmt)
- Run tests with race detection
- Zero warnings policy
Quality Gates:
- Pre-commit: make lint and make test with zero warnings
- CI: Count violations and fail if thresholds exceeded
2. Drift Policy
Purpose: Define when to stop, revert, and document when the model's edits deviate from project rules.
Stop & Revert Criteria:
- Model introduces new dependencies not listed in allowed stack
- Edits touch files outside specified targets (>3 unexpected files)
- Linting or type errors introduced that cannot be resolved within task scope
- Tests fail and model proposes changing tests to pass
Immediate Actions:
- Stop the model session and return clear error message summarizing deviation
- Revert workspace to state before task started (use git where appropriate)
- Create incident note in
docs/drift-incidents/with: what happened, files changed, new dependencies, remediation steps
Recording Learnings:
- After resolving, update
.cursor/rules/orCLAUDE.mdwith new rules to prevent recurrence
Allowed Deviations:
- Minor formatting changes (editor config) or whitespace-only edits
- Single-line refactors within targeted files if within scope and type-checked
Review & Approval:
- Any revert or incident note must be reviewed by human maintainer before re-running
3. Task Sizing Guidance
Purpose: Keep model-directed work in predictable, reviewable chunks to reduce drift and unexpected scope growth.
Guidelines:
- Target task duration: 30 minutes to 2.5 hours
- Prefer changes limited to 1-3 files where possible
- If a requested change touches >5 files or will take >2.5 hours, split into smaller tasks
Process:
- Define smallest useful deliverable and clear acceptance test
- Provide example files and exact paths to edit
- After each completed task, run tests and commit with clear message
Breaking Down Large Tasks:
- Decompose into: (a) tests + scaffolding, (b) minimal implementation, (c) refactor & polish
- Create plan note listing sub-tasks and acceptance criteria
Examples:
- Small: fix a bug in
src/utils/parse.ts— 30-60 mins - Medium: add new API endpoint
src/server/user.tswith tests — 1-2.5 hours - Large: migrate auth system — split into design + incremental PRs
Decision Support
After review, provide recommendations:
If Critical Issues Found
"I've identified [n] critical issues that block development:
- [Critical issue 1]
- [Critical issue 2]
Recommendation: Fix these issues before starting development.
Would you like to:
- Fix issues now (I'll regenerate affected tasks)
- Review issues manually
- Proceed with awareness (not recommended)"
If High Priority Issues Found
"I've identified [n] high-priority issues that should be addressed:
- [Issue 1]
- [Issue 2]
Recommendation: Address these before development or early in milestone 0.
Would you like to:
- Address now (I'll update the plan)
- Note for milestone 0
- Proceed and handle as needed"
If Ready with Modifications
"The plan is mostly ready with [n] recommended improvements.
Strong areas:
- [What's well-planned]
Recommended improvements:
- [Improvement 1]
- [Improvement 2]
Timeline impact: [estimate]
Would you like to:
- Apply improvements now
- Note for implementation
- Start development as-is"
If Ready for Development
"The implementation plan is ready for development!
Strengths:
- [What's excellent]
- [Best practices followed]
Minor optimization opportunities:
- [Opportunity 1]
- [Opportunity 2]
Estimated timeline: [duration] Critical path: [duration]
You can begin development with confidence."
Integration
This skill integrates with:
- implementation-planner - Reviews plans generated by this skill
- business-requirements-interview - Validates alignment with business needs
- technical-requirements-interview - Validates technical specification coverage
Examples
See examples/ for sample review outputs.