testing-agents-with-subagents

Installation
SKILL.md

Testing Agents With Subagents

Overview

This skill applies TDD methodology to agent development — RED (observe failures), GREEN (fix agent definition), REFACTOR (edge cases and robustness) — with subagent dispatch as the execution mechanism.

Test what the agent DOES, not what the prompt SAYS. Evidence-based verification only: capture exact outputs from subagent dispatch, verify every prompt change through testing. Always test via the Task tool, always test via the Task tool rather than reading prompts.

Minimum test counts vary by agent type: Reviewer agents need 6 cases (2 real issues, 2 clean, 1 edge, 1 ambiguous), Implementation agents 5 cases (2 typical, 1 complex, 1 minimal, 1 error), Analysis agents 4 cases (2 standard, 1 edge, 1 malformed), Routing/orchestration 4 cases (2 correct route, 1 ambiguous, 1 invalid). No agent is simple enough to skip testing — get human confirmation before exempting any agent.

Each test runs in a fresh subagent to avoid context pollution. After any fix, re-run ALL test cases to catch regressions. One fix at a time — you cannot determine what changed the outcome with multiple simultaneous fixes.


Instructions

Phase 0: PREPARE — Understand the Agent

Goal: Read the agent definition and understand what it claims to do before writing tests.

Step 1: Read the agent file

# Read agent definition
cat agents/{agent-name}.md

# Read any referenced skills
cat skills/{skill-name}/SKILL.md

Step 2: Identify testable claims

Extract concrete, testable behaviors from the agent definition:

  • What inputs does it accept?
  • What output structure does it produce?
  • What routing triggers should activate it?
  • What error conditions does it handle?
  • What skills does it invoke?

Step 3: Determine minimum test count

Agent Type Minimum Tests Required Coverage
Reviewer agents 6 2 real issues, 2 clean, 1 edge, 1 ambiguous
Implementation agents 5 2 typical, 1 complex, 1 minimal, 1 error
Analysis agents 4 2 standard, 1 edge, 1 malformed
Routing/orchestration 4 2 correct route, 1 ambiguous, 1 invalid

No gate — this phase is preparation. Move directly to Phase 1.

Phase 1: RED — Observe Current Behavior

Goal: Run agent with test inputs and document exact current behavior before any changes.

Step 1: Define test plan

## Test Plan: {agent-name}

**Agent Purpose:** {what the agent does}
**Agent File:** agents/{agent-name}.md
**Date:** {date}

**Test Cases:**

| ID | Input | Expected Output | Validates |
|----|-------|-----------------|-----------|
| T1 | {input} | {expected} | Happy path |
| T2 | {input} | {expected} | Error handling |
| T3 | {input} | {expected} | Edge case |

Write the test plan to a file before executing. This creates a reproducible baseline.

Step 2: Dispatch subagent with test inputs

Use the Task tool to dispatch the agent:

Task(
  prompt="""
  [Test input for the agent]

  Context: [Any required context]

  {Include the actual problem/request the agent should handle}
  """,
  subagent_type="{agent-name}"
)

Each test runs in a fresh subagent — this prevents context pollution from earlier tests affecting later ones.

Step 3: Capture results verbatim

Document exact agent outputs. Record verbatim output:

## Test T1: Happy Path

**Input:**
{exact input provided}

**Expected Output:**
{what you expected}

**Actual Output:**
{verbatim output from agent — record verbatim}

**Result:** PASS / FAIL
**Failure Reason:** {if FAIL, exactly what was wrong}

Step 4: Identify failure patterns

  • Which test categories fail (happy path, error, edge)?
  • Are failures structural (missing sections) or behavioral (wrong answers)?
  • Do failures correlate with input characteristics?

Gate: All test cases executed. Exact outputs captured verbatim. Failures documented with specific issues identified. Proceed only when gate passes.

Phase 2: GREEN — Fix Agent Definition

Goal: Update agent definition until all test cases pass. One fix at a time.

Step 1: Prioritize failures

Triage failures by severity:

Severity Description Priority
Critical Agent produces wrong answers or harmful output Fix first
High Agent missing required output sections Fix second
Medium Agent formatting or structure issues Fix third
Low Agent phrasing or style inconsistencies Fix last

Step 2: Diagnose root cause

Failure Type Fix Approach
Missing output section Add explicit instruction to include section
Wrong format Add output schema with examples
Missing context handling Add instructions for handling missing info
Incorrect classification Add calibration examples
Hallucinated content Add constraint to only use provided info
Agent asks questions instead of answering Provide required context in prompt or add default handling

Step 3: Make one fix at a time

Change one thing in the agent definition. Re-run ALL test cases. Document which tests now pass/fail.

Make one fix at a time — you cannot determine which change was effective. Same debugging principle: one variable at a time.

Step 4: Iterate until green

Repeat Step 3 until all test cases pass. If a fix causes a previously passing test to fail, revert and try a different approach.

Track fix iterations:

## Fix Log

| Iteration | Change Made | Tests Passed | Tests Failed | Action |
|-----------|------------|-------------|-------------|--------|
| 1 | Added output schema | T1, T2 | T3 | Continue |
| 2 | Added error handling instruction | T1, T2, T3 || Green |

Gate: All test cases pass. No regressions from previously passing tests. Can explain what each fix changed and why. Proceed only when gate passes.

Phase 3: REFACTOR — Edge Cases and Robustness

Goal: Verify agent handles boundary conditions and produces consistent outputs.

Step 1: Add edge case tests

Category Test Cases
Empty Input Empty string, whitespace only, no context
Large Input Very long content, deeply nested structures
Unusual Input Malformed data, unexpected formats
Ambiguous Input Cases where correct behavior is unclear

Step 2: Run consistency tests

Run the same input 3 times. Outputs should be consistent:

  • Same structure
  • Same key findings (for analysis agents)
  • Acceptable variation in phrasing only

If inconsistent: add more explicit instructions to the agent definition. Re-test.

Step 3: Run regression suite

Re-run ALL test cases (original + edge cases) to confirm nothing broke during refactoring.

Step 4: Document final test report

## Test Report: {agent-name}

| Metric | Result |
|--------|--------|
| Test Cases Run | N |
| Passed | N |
| Failed | N |
| Pass Rate | N% |

## Verdict
READY FOR DEPLOYMENT / NEEDS FIXES / REQUIRES REVIEW

Gate: Edge cases handled. Consistency verified. Full suite green. Test report documented. Fix is complete.


Error Handling

Error: "Agent type not found"

Cause: Agent not registered or name misspelled Solution:

  1. Verify agent file exists: ls agents/{agent-name}.md
  2. Check YAML frontmatter has correct name field
  3. Restart Claude Code to pick up new agents

Error: "Inconsistent outputs across runs"

Cause: Agent produces different results for same input Solution:

  1. Document the inconsistency — this is a valid finding
  2. Add more explicit instructions to agent definition
  3. Re-test consistency after fix
  4. Determine if variation is acceptable (phrasing) or problematic (structure/findings)

Error: "Subagent timeout"

Cause: Agent taking too long to respond Solution:

  1. Simplify test input to reduce processing
  2. Check agent isn't in an infinite loop or excessive tool use
  3. Increase timeout if agent legitimately needs more time

Error: "Agent asks questions instead of answering"

Cause: Agent needs clarification that test input did not provide Solution:

  1. This may be correct behavior — agent properly requesting context
  2. Update test input to provide the required context
  3. Or update agent definition to handle ambiguity with defaults
  4. Document whether questioning behavior is acceptable for this agent type

Examples

Example 1: Testing a New Reviewer Agent

User says: "Test the new reviewer-security agent" Actions:

  1. Define 6 test cases: 2 real issues, 2 clean code, 1 edge case, 1 ambiguous (RED)
  2. Dispatch subagent for each, capture verbatim outputs (RED)
  3. Fix agent definition for any failures, re-run all tests (GREEN)
  4. Add edge cases (empty input, malformed code), verify consistency (REFACTOR) Result: Agent passes all tests, report documents pass rate and verdict

Example 2: Testing After Agent Modification

User says: "I updated the golang-general-engineer, make sure it still works" Actions:

  1. Run existing test cases against modified agent (RED)
  2. Compare outputs to previous baseline (RED)
  3. Fix any regressions introduced by the modification (GREEN)
  4. Test edge cases to verify robustness not degraded (REFACTOR) Result: Agent modification validated, no regressions confirmed

Example 3: Testing Routing Logic

User says: "Verify the /do router sends Go requests to the right agent" Actions:

  1. Define test cases: "Review this Go code", "Fix this .go file", "Write a goroutine" (RED)
  2. Dispatch each through router, verify correct agent handles it (RED)
  3. Fix routing triggers if wrong agent selected (GREEN)
  4. Test ambiguous inputs like "Review this code" with mixed-language context (REFACTOR) Result: Routing validated for all trigger phrases, ambiguous cases documented

References

Integration

  • agent-comparison: A/B test agent variants
  • agent-evaluation: Structural quality checks
  • test-driven-development: TDD principles applied to agents

Reference Files

  • ${CLAUDE_SKILL_DIR}/references/testing-patterns.md: Dispatch patterns, test scenarios, eval harness integration
Related skills
Installs
7
GitHub Stars
366
First Seen
Mar 23, 2026