skills/javalenciacai/qaskills/test-execution-manager

test-execution-manager

SKILL.md

Test Execution Manager

Expert skill for executing tests systematically and managing the test execution lifecycle.

When to Use

Use this skill when you need to:

  • Execute test suites
  • Log defects with complete information
  • Perform exploratory testing
  • Monitor test execution progress
  • Report test results
  • Triage test failures

Test Execution Types

1. Smoke Testing

Purpose: Quick validation that critical functionality works When: After each build/deployment Scope: Core features only (15-30 minutes) Goal: Determine if further testing is viable

2. Regression Testing

Purpose: Verify existing functionality still works When: Before each release Scope: Previously working features Goal: Catch unintended side effects

3. Functional Testing

Purpose: Verify features work per requirements When: During development and QA cycles Scope: All functional requirements Goal: Validate business logic

4. Integration Testing

Purpose: Verify components work together When: After module completion Scope: Inter-system interfaces, APIs, data flow Goal: Validate system integration

5. Exploratory Testing

Purpose: Unscripted investigation to find edge cases When: Throughout testing cycle Scope: Areas prone to defects Goal: Discover unexpected behaviors

Test Execution Workflow

1. Prepare Test Environment
   - Verify environment setup
   - Load test data
   - Check system availability

2. Execute Test Suite
   - Follow test case steps exactly
   - Document actual results
   - Capture screenshots/logs for failures

3. Log Defects
   - Use defect template
   - Include reproduction steps
   - Attach evidence

4. Triage Results
   - Categorize: Pass/Fail/Blocked
   - Prioritize failures
   - Determine root cause

5. Report Status
   - Update test metrics
   - Communicate blockers
   - Provide recommendations

Defect Logging Template

Required Information:

Defect ID: BUG-XXX
Title: [Clear, specific title]
Severity: [Critical/High/Medium/Low]
Priority: [P0/P1/P2/P3]

Environment:
- OS: [Windows/Mac/Linux + version]
- Browser: [Chrome/Firefox/etc + version]
- App Version: [x.y.z]
- Test Environment: [Dev/QA/Staging]

Steps to Reproduce:
1. [Exact step]
2. [Exact step]
3. [Exact step]

Expected Result: [What should happen]
Actual Result: [What actually happened]

Attachments:
- Screenshots
- Error logs
- Network traces
- Video recording (if applicable)

Additional Info:
- Frequency: [Always/Intermittent]
- Related tickets: [Links]
- Workaround: [If any]

Severity Definitions

  • Critical: System crash, data loss, security breach
  • High: Major feature broken, no workaround
  • Medium: Feature partially broken, workaround exists
  • Low: Cosmetic issue, minor inconvenience

Priority Definitions

  • P0: Blocks release, fix immediately
  • P1: Must fix before release
  • P2: Should fix, can defer if needed
  • P3: Nice to have, can be backlogged

Test Execution Best Practices

During Execution

  • ✓ Execute tests exactly as written
  • ✓ Don't assume - verify every step
  • ✓ Capture evidence immediately
  • ✓ Note environment details
  • ✓ Report blockers immediately
  • ✓ Retest fixed defects promptly

Test Data Management

  • ✓ Use test accounts, not production
  • ✓ Reset data between test runs
  • ✓ Protect sensitive information
  • ✓ Document data dependencies

Communication

  • ✓ Daily test status updates
  • ✓ Flag risks early
  • ✓ Collaborate with developers
  • ✓ Provide clear reproduction steps

Test Execution Metrics

Track these metrics:

  • Test Pass Rate: Passed / Total executed
  • Defect Detection Rate: Defects found / Total tests
  • Test Coverage: Tests executed / Total planned
  • Execution Progress: Completed / Total
  • Blocked Tests: Tests unable to run
  • Flaky Tests: Inconsistent results

Exploratory Testing Charter Template

Mission: [What to explore]
Duration: 60-90 minutes
Areas:
- [Specific feature/flow to investigate]
- [Edge cases to try]
- [Integration points]

Notes:
- [Observations]
- [Interesting behaviors]
- [Potential defects]

Defects Found: [Count + IDs]
Coverage: [What was tested]
Risks: [Concerns identified]

Test Execution Report Template

Test Execution Summary
Date: [Date]
Sprint/Release: [Version]
Tester: [Name]

Overall Status: [On Track / At Risk / Blocked]

Test Results:
- Total Planned: XXX
- Executed: XXX
- Passed: XXX (XX%)
- Failed: XXX (XX%)
- Blocked: XXX (XX%)
- Remaining: XXX

Defects Summary:
- Critical: XX
- High: XX
- Medium: XX
- Low: XX
- Total: XX

Blockers:
1. [Description + impact]
2. [Description + impact]

Risks:
1. [Risk + mitigation]

Next Steps:
1. [Action item]
2. [Action item]

Completion Criteria

A test cycle is complete when:

  • ✓ All planned tests executed
  • ✓ Critical defects fixed and verified
  • ✓ Pass rate meets threshold (typically 95%+)
  • ✓ No blocking defects remain
  • ✓ Test evidence documented
  • ✓ Sign-off obtained from stakeholders
Weekly Installs
2
GitHub Stars
1
First Seen
9 days ago
Installed on
mcpjam2
claude-code2
replit2
junie2
windsurf2
zencoder2