test-analyze

SKILL.md

Test Analysis and Fix Planning

Analyzes test results from any testing framework, uses Zen planner to create a systematic fix strategy, and delegates fixes to appropriate subagents.

Usage

/test:analyze <results-path> [--type <test-type>] [--focus <area>]

Parameters

  • <results-path>: Path to test results directory or file (required)

    • Examples: ./test-results/, ./coverage/, pytest-report.xml
  • --type <test-type>: Type of tests (optional, auto-detected if omitted)

    • accessibility - Playwright a11y, axe-core
    • unit - Jest, pytest, cargo test
    • integration - API tests, database tests
    • e2e - Playwright, Cypress, Selenium
    • security - OWASP ZAP, Snyk, TruffleHog
    • performance - Lighthouse, k6, JMeter
  • --focus <area>: Specific area to focus on (optional)

    • Examples: authentication, api, ui-components, database

Examples

# Analyze Playwright accessibility test results
/test:analyze ./test-results/ --type accessibility

# Analyze unit test failures with focus on auth
/test:analyze ./coverage/junit.xml --type unit --focus authentication

# Auto-detect test type and analyze all issues
/test:analyze ./test-output/

# Analyze security scan results
/test:analyze ./security-report.json --type security

Command Flow

  1. Analyze Test Results

    • Parse test result files (XML, JSON, HTML, text)
    • Extract failures, errors, warnings
    • Categorize issues by type and severity
    • Identify patterns and root causes
  2. Plan Fixes with PAL Planner

    • Use mcp__pal__planner for systematic planning
    • Break down complex fixes into actionable steps
    • Identify dependencies between fixes
    • Estimate effort and priority
  3. Delegate to Subagents

    • Accessibility issuescode-review agent (WCAG compliance)
    • Security vulnerabilitiessecurity-audit agent
    • Performance problemssystem-debugging agent
    • Code quality issuescode-refactoring agent
    • Test infrastructuretest-architecture agent
    • Integration failuressystem-debugging agent
    • Documentation gapsdocumentation agent
  4. Execute Plan

    • Sequential execution based on dependencies
    • Verification after each fix
    • Re-run tests to confirm resolution

Subagent Selection Logic

The command uses this decision tree to delegate:

  • Accessibility violations (WCAG, ARIA, contrast) → code-review agent with accessibility focus

  • Security issues (XSS, SQLi, auth bypass) → security-audit agent with OWASP analysis

  • Performance bottlenecks (slow queries, memory leaks) → system-debugging agent with profiling

  • Code smells (duplicates, complexity, coupling) → code-refactoring agent with SOLID principles

  • Flaky tests (race conditions, timing issues) → test-architecture agent with stability analysis

  • Build/CI failures (pipeline errors, dependency issues) → cicd-pipelines agent with workflow optimization

Output

The command produces:

  1. Summary Report

    • Total issues found
    • Breakdown by category/severity
    • Top priorities
  2. Fix Plan (from PAL planner)

    • Step-by-step remediation strategy
    • Dependency graph
    • Effort estimates
  3. Subagent Assignments

    • Which agent handles which issues
    • Rationale for delegation
    • Execution order
  4. Actionable Next Steps

    • Commands to run
    • Files to modify
    • Verification steps

Notes

  • Works with any test framework that produces structured output
  • Auto-detects common test result formats (JUnit XML, JSON, TAP)
  • Preserves test evidence for debugging
  • Can be chained with /git:smartcommit for automated fixes
  • Respects TDD workflow (RED → GREEN → REFACTOR)

Related Commands

  • /test:run - Run tests with framework detection
  • /code:review - Manual code review for test files
  • /docs:update - Update test documentation
  • /git:smartcommit - Commit fixes with conventional messages

Prompt:

Analyze test results from {{ARG1}} and create a systematic fix plan.

{{#if ARG2}} Test type: {{ARG2}} {{else}} Auto-detect test type from file formats and content. {{/if}}

{{#if ARG3}} Focus area: {{ARG3}} {{/if}}

Step 1: Analyze Test Results

Read the test result files from {{ARG1}} and extract:

  • Failed tests with error messages
  • Warnings and deprecations
  • Performance metrics (if available)
  • Coverage gaps (if available)
  • Categorize by: severity (critical/high/medium/low), type (functional/security/performance/accessibility)

Step 2: Use PAL Planner

Call mcp__pal__planner with model "gemini-2.5-pro" to create a systematic fix plan:

  • Step 1: Summarize findings and identify root causes
  • Step 2: Prioritize issues (impact × effort matrix)
  • Step 3: Break down fixes into actionable tasks
  • Step 4: Identify dependencies between fixes
  • Step 5: Assign each fix category to appropriate subagent
  • Continue planning steps as needed for complex scenarios

Step 3: Subagent Delegation Strategy

Based on the issue categories, delegate to:

  • Accessibility violations (WCAG, ARIA, color contrast, keyboard nav) → Use Task tool with subagent_type: code-review → Focus: WCAG 2.1 compliance, semantic HTML, ARIA best practices

  • Security vulnerabilities (XSS, SQLi, CSRF, auth issues) → Use Task tool with subagent_type: security-audit → Focus: OWASP Top 10, input validation, authentication

  • Performance issues (slow tests, memory leaks, timeouts) → Use Task tool with subagent_type: system-debugging → Focus: Profiling, bottleneck identification, optimization

  • Code quality (duplicates, complexity, maintainability) → Use Task tool with subagent_type: code-refactoring → Focus: SOLID principles, DRY, code smells

  • Flaky/unreliable tests (race conditions, timing, dependencies) → Use Task tool with subagent_type: test-architecture → Focus: Test stability, isolation, determinism

  • CI/CD failures (build errors, pipeline issues) → Use Task tool with subagent_type: cicd-pipelines → Focus: GitHub Actions, dependency management, caching

  • Documentation gaps (missing docs, outdated examples) → Use Task tool with subagent_type: documentation → Focus: API docs, test documentation, migration guides

Step 4: Create Execution Plan

For each subagent assignment:

  1. Context: What files/areas need attention
  2. Objective: Specific fix goal
  3. Success Criteria: How to verify the fix
  4. Dependencies: What must be done first
  5. Verification: Commands to re-run tests

Step 5: Present Summary

Provide:

  • 📊 Issue Breakdown: Count by category and severity
  • 🎯 Priorities: Top 3-5 issues to fix first
  • 🤖 Subagent Plan: Which agents will handle what
  • Next Steps: Concrete actions to take
  • 🔍 Verification: How to confirm fixes worked

{{#if ARG3}} Additional focus on {{ARG3}}: Prioritize issues related to this area and provide extra context for relevant subagents. {{/if}}

Documentation-First Reminder: Before implementing fixes, research relevant documentation using context7 to verify:

  • Test framework best practices
  • Accessibility standards (WCAG 2.1)
  • Security patterns (OWASP)
  • Performance optimization techniques

TDD Workflow: Follow RED → GREEN → REFACTOR:

  1. Verify tests fail (RED) ✓ (already done)
  2. Implement minimal fix (GREEN)
  3. Refactor for quality
  4. Re-run tests to confirm

Do you want me to proceed with the analysis and planning, or would you like to review the plan first?

Weekly Installs
46
GitHub Stars
13
First Seen
Feb 9, 2026
Installed on
cline46
github-copilot46
codex46
kimi-cli46
gemini-cli46
cursor46