skills/petrkindlmann/qa-skills/test-reliability

test-reliability

Installation
SKILL.md

Before starting: Check for .agents/qa-project-context.md in the project root. It contains known flaky areas, selector strategy, and CI environment details.


Discovery Questions

  1. What is your current flaky test rate? Check CI failure stats over the last 30 days. Below 2% is healthy. 2-5% needs attention. Above 5% is eroding team trust.

  2. Where is the pain concentrated? Is it locator breakage? Timing issues? Test data? Environment instability? If unknown, instrument first (see Flake Classification).

  3. What is your current selector strategy? data-testid everywhere? Mixed CSS and role-based? No strategy (whatever works)?

  4. How do you handle flaky tests today? Retry and hope? Skip and forget? Something structured?

  5. What CI environment runs the tests? Same machine every time or different runners? Consistent resources or variable? How does the CI runner compare to local dev machines?

  6. What is your test data strategy? Shared database? Per-test fixtures? Factory seeding? External services?


Core Principles

  1. Prevention over cure. Writing resilient tests costs 1x. Investigating a flaky test costs 10x. Losing team trust in the suite costs 100x.

  2. Healing must be observable and reviewable. Every automated repair produces evidence: what broke, what was tried, what worked, what the confidence score is. Silent fixes erode trust as fast as silent failures.

  3. Classify before fixing. The fix for a timing issue is completely different from the fix for a data dependency. Wrong diagnosis wastes effort and may make things worse.

  4. Flaky tests are bugs. They are not annoyances to tolerate. A flaky test either has a test bug (fix the test), reveals an app bug (fix the app), or exposes an environment issue (fix the environment).

  5. Track reliability as a metric, not a feeling. Measure flaky rate, mean time to heal, quarantine age, and selector stability. What gets measured gets fixed.

  6. Self-healing is a spectrum. Start with resilient locators (Level 1), add fallback strategies (Level 2), then environment-aware healing (Level 3), then confidence-scored auto-repair (Level 4). Do not jump to Level 4 before mastering Level 1.


Locator Resilience

Multi-Attribute Selectors (Beyond Fallback Chains)

A single locator strategy is a single point of failure. Multi-attribute selectors combine multiple signals for a single element lookup, creating resilience without fallback chain complexity.

The key insight: instead of "try A, then B, then C," use "find element matching A AND B AND C with tolerance for one signal missing."

// Multi-attribute locator: tries combinations from most specific to least
const submitBtn = await multiAttributeLocator(page, {
  testId: 'checkout-submit',              // most stable signal
  role: 'button',                          // semantic signal
  name: /place order/i,                    // accessible name
  nearText: 'Order Summary',              // visual context
});
// Internally: tries testId+role+name first, then testId alone, then role+name,
// then text, then nearText+role. Returns first visible match.
// Unlike fallback chains, it combines signals for higher confidence.

DOM Similarity / Neighbor Context Matching

When a locator fails, the element may still exist but with changed attributes. Use surrounding DOM context to find it through three strategies:

  1. Parent + tag + type: Find the container element (by testId), then locate by tag and type within it
  2. Preceding label: Find sibling text (label), then locate the adjacent input/button
  3. Nearby text context: Find visible text near the target, then locate the element type in the same parent

These strategies are used as repair candidates (scored by the confidence system below), not as runtime fallbacks.

Selector Stability Scoring

Rate every selector on a 0-5 scale to prioritize refactoring.

Score Strategy Survives
5 getByTestId('submit-order') CSS, text, and structural changes
4 getByRole('button', { name: 'Submit' }) CSS and structural changes
3 getByLabel('Email') CSS changes; breaks on label rewording
2 getByText('Submit Order') Breaks on any copy change
1 locator('.btn-primary.submit') Breaks on CSS or structural change
0 locator('//div[3]/button[1]') Breaks on any DOM change

Target: Average score of 3.5+ across the test suite. Audit monthly. Prioritize fixing score-0 and score-1 selectors.


Flake Classification Framework

Every flaky test has a root cause category. Classifying correctly determines the fix.

Categories

Category Signal Root Cause Fix Direction
Timing Timeout errors, passes on retry, worse in CI Race condition, animation, async operation Wait for condition, not time
Data dependency Fails with other tests, passes alone Shared state, missing cleanup Isolate per-test, fixture cleanup
Environment Fails on specific runner, correlates with load Resource contention, network latency Mock externals, increase resources
Order dependency Fails with --shard or fullyParallel Test depends on another test's side effect Self-contained setup
Time sensitivity Fails at specific times (midnight, month-end) Uses real clock, date boundary Mock clock, relative comparisons
Visual rendering Screenshot diff flickers, subpixel differences Font rendering, antialiasing, animation frame Increase threshold, mask dynamic regions
External service Correlates with third-party status Real HTTP calls in tests Mock external APIs

Classification Decision Tree

Test is flaky
├── Does it pass when run alone?
│   ├── YES → ORDER DEPENDENCY or DATA DEPENDENCY
│   │   ├── Does another test create/modify data it needs? → ORDER DEPENDENCY
│   │   └── Does it share a database/file/cache? → DATA DEPENDENCY
│   │
│   └── NO → Not order/data dependent. Continue below.
├── Does it fail more often in CI than locally?
│   ├── YES → TIMING or ENVIRONMENT
│   │   ├── Timeout errors? → TIMING (CI is slower)
│   │   ├── Connection errors? → ENVIRONMENT (network/service)
│   │   └── Resource errors (OOM, disk)? → ENVIRONMENT (resources)
│   │
│   └── NO → Same rate locally and CI. Continue below.
├── Does it fail at specific times?
│   ├── YES → TIME SENSITIVITY
│   │   ├── Near midnight? → Date boundary issue
│   │   ├── Near month/year end? → Calendar calculation
│   │   └── Specific hour? → Timezone issue
│   │
│   └── NO → Continue below.
├── Does it involve screenshots or visual comparison?
│   ├── YES → VISUAL RENDERING
│   │
│   └── NO → Continue below.
├── Does it call external HTTP APIs?
│   ├── YES → EXTERNAL SERVICE
│   │
│   └── NO → TIMING (most likely — default classification)
│       └── Investigate: what async operation is not being awaited?

Environment-Aware Healing

Not all test failures are test problems. Some are environment problems. Environment-aware healing distinguishes between the two and adapts.

Slow Backend vs True UI Failure

When an action fails, check backend health before blaming the test:

  1. Action fails -> Hit /api/health endpoint
  2. Backend unhealthy (5xx or timeout) -> Retry with exponential backoff. Diagnose as backend_down. This is not a UI bug.
  3. Backend healthy (2xx) -> This is a real UI/test failure. Do not retry.

Return a structured diagnosis: { success: boolean; diagnosis: 'backend_down' | 'ui_failure' | 'backend_slow_recovered' }. This diagnosis feeds into flake classification -- backend issues are environment issues, not test bugs.

Resource Contention Detection

Before declaring test failure in CI, check for resource contention:

  • Browser health: Load about:blank. If it takes > 2s (baseline: < 500ms), the runner is overloaded.
  • API health: Hit /api/health. If it takes > 5s (baseline: < 1s), the backend is under pressure.
  • Diagnosis: If either check fails, classify as environment issue and annotate the test result. Do not count resource contention failures toward flaky test rates.

Data Healing

Test data expires, gets cleaned up, or becomes invalid. Data healing detects and regenerates stale test data.

Common Data Failure Patterns

Pattern Signal Fix
Expired auth token 401 response during test Regenerate token in fixture
Deleted test record 404 when accessing seeded data Re-seed before test
Uniqueness violation 409 or constraint error Generate unique identifiers per run
Stale cache Wrong data returned Clear cache in setup
Exceeded quota 429 or rate limit error Reset quotas or use dedicated test account

Self-Healing Test Data Fixture Pattern

Build fixtures that verify data exists and regenerate if stale:

// Pattern: verify → heal → use → cleanup
testUser: async ({ request }, use, testInfo) => {
  // 1. Try to find existing test user by deterministic email
  // 2. Verify auth token is still valid (GET /api/me)
  // 3. If token expired → refresh it (POST /refresh-token), mark as healed
  // 4. If user missing → create new one, mark as healed
  // 5. If healed → annotate testInfo for observability
  // 6. use(user) → run the test
  // 7. Cleanup: delete test user (guaranteed by fixture, even on failure)
}

Key patterns:

  • Use testInfo.testId in email/identifiers for per-test uniqueness
  • Annotate testInfo.annotations when healing occurs for observability
  • Always clean up in the fixture's post-use block, not in afterEach

Observable Repair Workflow

Core guardrail: Healing must be observable and reviewable. Every repair follows this flow:

Failure Detected
Candidate Repair Generated
Confidence Score Computed (0.0 - 1.0)
Evidence Diff Produced (what changed, what was tried)
Approval Policy Applied
  │ ├── Score >= 0.9 → Auto-apply, log for review
  │ ├── Score 0.7-0.9 → Apply in quarantine, flag for review
  │ └── Score < 0.7 → Do NOT apply, create investigation ticket
Intent Fidelity Check (does repaired test still test the same thing?)
Rollback if intent fidelity drops

Confidence Scoring

Score each repair candidate on six dimensions (weighted sum, 0.0-1.0):

Dimension Weight Scoring
Match specificity 0.30 testId=1.0, role=0.9, text=0.7, context=0.5, CSS=0.3
Element visible 0.15 1.0 if visible, 0.0 if not
Same parent container 0.15 1.0 if same container, 0.0 if different
Same element type 0.15 1.0 if same tag+role, 0.0 if different
Text similarity 0.15 0.0-1.0 (Levenshtein ratio of accessible name)
Attribute overlap 0.10 0.0-1.0 (Jaccard of shared attributes)

Score thresholds:

  • = 0.9: Auto-apply, log for batch review

  • 0.7-0.89: Apply in quarantine, flag for individual review
  • 0.5-0.69: Do not apply, create PR with evidence
  • < 0.5: Discard, manual investigation required

Repair Evidence

Every repair produces an evidence record containing: test file, test name, failure type, original locator, candidate replacements (each with confidence, evidence string, and intent-preserved flag), which candidate was selected, timestamp, approval path, and rollback trigger.

Intent Fidelity Checking

After applying a repair, verify the test still exercises the same user intent:

  • Element type changed (e.g., button -> a) -- intent NOT preserved, rollback
  • ARIA role changed (e.g., button -> link) -- intent NOT preserved, rollback
  • Form action changed (targets different endpoint) -- intent NOT preserved, rollback
  • Same tag, role, and form action -- intent preserved, keep repair

A repair that changes WHAT the test verifies (not just HOW it finds elements) must be rolled back.


Quarantine Management

Quarantine isolates flaky tests so they run but do not block CI.

Setup

// Tag the flaky test
test('intermittent WebSocket reconnect', {
  tag: ['@quarantine'],
  annotation: {
    type: 'quarantine',
    description: 'Flaky since 2026-03-15. Race condition in WebSocket handler. Ticket: BUG-1234.',
  },
}, async ({ page }) => { /* ... */ });
// playwright.config.ts — separate projects
projects: [
  {
    name: 'stable',
    testMatch: /.*\.spec\.ts/,
    grep: /^(?!.*@quarantine)/,  // exclude quarantine
  },
  {
    name: 'quarantine',
    grep: /@quarantine/,
    retries: 3,
  },
],
# CI: quarantine runs separately, does not block
- name: Run stable tests
  run: npx playwright test --project=stable

- name: Run quarantine tests
  run: npx playwright test --project=quarantine
  continue-on-error: true

Quarantine Lifecycle

1. DETECT    — Test identified as flaky (by CI reporter or manual triage)
2. TAG       — Add @quarantine annotation with ticket link and date
3. ISOLATE   — Quarantine project runs separately, does not block
4. DIAGNOSE  — Follow flaky test runbook (references/flaky-test-runbook.md)
5. FIX       — Apply fix pattern per category
6. VERIFY    — Run 50x with --repeat-each, zero failures required
7. RELEASE   — Remove @quarantine tag, add annotation documenting fix

Quarantine Hygiene Rules

  • Maximum quarantine age: 14 days. After 14 days, either fix it or delete it. Permanent quarantine is permanent rot.
  • Every quarantine entry has a ticket link. No anonymous quarantines.
  • Weekly review. Check quarantine list every sprint. Aging quarantines get escalated.
  • Track quarantine size. More than 5% of tests in quarantine = systemic problem requiring process change, not just test fixes.

Flaky Test Runbook

See references/flaky-test-runbook.md for the complete step-by-step runbook. Summary:

1. REPRODUCE     — Run with --repeat-each=20 --workers=4 --trace=on
2. CORRELATE     — Check CI history: when did it start? What changed?
3. CLASSIFY      — Use the classification decision tree above
4. FIX           — Apply fix pattern for the specific category
5. VERIFY        — Run with --repeat-each=50, zero failures required
6. RELEASE       — Remove quarantine tag, document fix

Anti-Patterns

1. Silent Selector Replacement

Automatically replacing a broken selector without logging, review, or confidence scoring. The repaired test may now verify a different element entirely. Every repair must produce evidence.

2. "Just Retry It" as a Fix

Retries are a detection mechanism, not a fix. A test that needs retry 2 of 3 times will eventually fail 3 of 3 during your most critical release.

3. Disabling Flaky Tests Permanently

test.skip('flaky, will fix later') -- "later" never comes. Either quarantine with tracking or delete entirely. Skipped tests with no ticket are dead code.

4. Treating All Flakiness the Same

Timing issues and data dependencies need completely different fixes. Adding waitForTimeout(5000) to a data dependency problem makes the test slower and still flaky.

5. waitForTimeout as a Stability Fix

// NEVER the right fix
await page.waitForTimeout(5000);

// Wait for the actual condition
await expect(page.getByRole('table')).toBeVisible();
await page.waitForResponse(resp => resp.url().includes('/api/data') && resp.status() === 200);

6. Healing Without Observability

Auto-repair that does not produce logs, evidence, or confidence scores. You cannot improve what you cannot measure. You cannot trust what you cannot review.

7. Over-Engineering Healing Before Writing Stable Tests

Building a complex self-healing framework before adopting basic resilient locator patterns. Start with multi-attribute selectors and proper waits. Add healing infrastructure only when you have data showing where breakage occurs.

8. No Quarantine Expiry

Tests sit in quarantine for months. Quarantine is a temporary state, not a permanent home. Enforce a 14-day maximum.


Done When

  • All flaky tests identified and categorized by root cause (timing, data dependency, environment, etc.)
  • Each flaky test quarantined or fixed — no test silently retried without a documented plan and ticket reference
  • Locator stability scores documented for the test suite with a target average of 3.5+
  • Flakiness rate tracked in CI dashboard and visible to the team
  • Repair actions tracked with ticket references and quarantine expiry dates set

Related Skills

  • playwright-automation — Full Playwright setup, Page Object Model, fixtures, and CI integration.
  • ci-cd-integration — Pipeline configuration, parallel execution, test reporting, and failure handling.
  • qa-metrics — Tracking flaky test rates, selector stability scores, quarantine size, and suite health.
  • ai-bug-triage — When flake investigation reveals an app bug, use the triage pipeline to classify and report it.
  • ai-test-generation — Generate reliable tests from the start using the staged pipeline.

References

  • references/flaky-test-runbook.md — Step-by-step runbook for triaging flaky tests: root cause decision tree, fix patterns per category, confidence scoring methodology.
Weekly Installs
11
GitHub Stars
4
First Seen
Apr 1, 2026
Installed on
amp10
cline10
opencode10
cursor10
kimi-cli10
warp10