trace
Trace Skill
Use this skill for ambiguous, causal, evidence-heavy questions where the goal is to explain why an observed result happened, not to jump directly into fixing or rewriting code.
This is the orchestration layer on top of the built-in tracer agent. The goal is to make tracing feel like a reusable OMC operating lane: restate the observation, generate competing explanations, gather evidence in parallel, rank the explanations, and propose the next probe that would collapse uncertainty fastest.
Good entry cases
Use /oh-my-claudecode:trace when the problem is:
- ambiguous
- causal
- evidence-heavy
- best answered by exploring competing explanations in parallel
Examples:
- runtime bugs and regressions
- performance / latency / resource behavior
- architecture / premortem / postmortem analysis
- scientific or experimental result tracing
- config / routing / orchestration behavior explanation
- “given this output, trace back the likely causes”
Core tracing contract
Always preserve these distinctions:
- Observation -- what was actually observed
- Hypotheses -- competing explanations
- Evidence For -- what supports each explanation
- Evidence Against / Gaps -- what contradicts it or is still missing
- Current Best Explanation -- the leading explanation right now
- Critical Unknown -- the missing fact keeping the top explanations apart
- Discriminating Probe -- the highest-value next step to collapse uncertainty
Do not collapse into:
- a generic fix-it coding loop
- a generic debugger summary
- a raw dump of worker output
- fake certainty when evidence is incomplete
Evidence strength hierarchy
Treat evidence as ranked, not flat.
From strongest to weakest:
- Controlled reproductions / direct experiments / uniquely discriminating artifacts
- Primary source artifacts with tight provenance (trace events, logs, metrics, benchmark outputs, configs, git history, file:line behavior)
- Multiple independent sources converging on the same explanation
- Single-source code-path or behavioral inference
- Weak circumstantial clues (timing, naming, stack order, resemblance to prior bugs)
- Intuition / analogy / speculation
Explicitly down-rank hypotheses that depend mostly on lower tiers when stronger contradictory evidence exists.
Strong falsification / disconfirmation rules
Every serious /trace run must try to falsify its own favorite explanation.
For each top hypothesis:
- collect evidence for it
- collect evidence against it
- state what distinctive prediction it makes
- state what observation would be hard to reconcile with it
- identify the cheapest probe that would discriminate it from the next-best alternative
Down-rank a hypothesis when:
- direct evidence contradicts it
- it survives only by adding new unverified assumptions
- it makes no distinctive prediction compared with rivals
- a stronger alternative explains the same facts with fewer assumptions
- its support is mostly circumstantial while the rival has stronger evidence tiers
Team-mode orchestration shape
Use Claude built-in team mode for /trace.
The lead should:
- Restate the observed result or “why” question precisely
- Extract the tracing target
- Generate multiple deliberately different candidate hypotheses
- Spawn 3 tracer lanes by default in team mode
- Assign one tracer worker per lane
- Instruct each tracer worker to gather evidence for and against its lane
- Run a rebuttal round between the leading hypothesis and the strongest remaining alternative
- Detect whether the top lanes genuinely differ or actually converge on the same root cause
- Merge findings into a ranked synthesis with an explicit critical unknown and discriminating probe
Important: workers should pursue deliberately different explanations, not the same explanation in parallel.
Default hypothesis lanes for v1
Unless the prompt strongly suggests a better partition, use these 3 default lanes:
- Code-path / implementation cause
- Config / environment / orchestration cause
- Measurement / artifact / assumption mismatch cause
These defaults are intentionally broad so the first slice works across bug, performance, architecture, and experiment tracing.
Mandatory cross-check lenses
After the initial evidence pass, pressure-test the leaders with these lenses when relevant:
- Systems lens -- queues, retries, backpressure, feedback loops, upstream/downstream dependencies, boundary failures, coordination effects
- Premortem lens -- assume the current best explanation is incomplete or wrong; what failure mode would embarrass the trace later?
- Science lens -- controls, confounders, measurement bias, alternative variables, falsifiable predictions
These lenses are not filler. Use them when they can surface a missed explanation, hidden dependency, or weak inference.
Worker contract
Each worker should be a tracer lane owner, not a generic executor.
Each worker must:
- own exactly one hypothesis lane
- restate its lane hypothesis explicitly
- gather evidence for the lane
- gather evidence against the lane
- rank the evidence strength behind its case
- call out missing evidence, failed predictions, and remaining uncertainty
- name the critical unknown for the lane
- recommend the best lane-specific discriminating probe
- avoid collapsing into implementation unless explicitly told to do so
Useful evidence sources include:
- relevant code, tests, configs, docs, logs, outputs, and benchmark artifacts
- existing trace artifacts via
trace_timeline - existing aggregate trace evidence via
trace_summary
Recommended worker return structure:
- Lane
- Hypothesis
- Evidence For
- Evidence Against / Gaps
- Evidence Strength
- Critical Unknown
- Best Discriminating Probe
- Confidence
Leader synthesis contract
The final /trace answer should synthesize, not just concatenate.
Return:
- Observed Result
- Ranked Hypotheses
- Evidence Summary by Hypothesis
- Evidence Against / Missing Evidence
- Rebuttal Round
- Convergence / Separation Notes
- Most Likely Explanation
- Critical Unknown
- Recommended Discriminating Probe
- Additional Trace Lanes (optional, only if uncertainty remains high)
Preserve a ranked shortlist even if one explanation is currently dominant.
Rebuttal round and convergence detection
Before closing the trace:
- let the strongest non-leading lane present its best rebuttal to the current leader
- force the leader to answer the rebuttal with evidence, not assertion
- if the rebuttal materially weakens the leader, re-rank the table
- if two “different” hypotheses reduce to the same underlying mechanism, merge them and say so explicitly
- if two hypotheses still imply different next probes, keep them separate even if they sound similar
Do not claim convergence just because multiple workers use similar language. Convergence requires either:
- the same root causal mechanism, or
- independent evidence streams pointing to the same explanation
Explicit down-ranking guidance
The lead should explicitly say why a hypothesis moved down:
- contradicted by stronger evidence
- lacks the observation it predicted
- requires extra ad hoc assumptions
- explains fewer facts than the leader
- lost the rebuttal round
- converged into a stronger parent explanation
This is important because /trace should teach the reader why one explanation outranks another, not just present a final table.
Suggested lead prompt skeleton
Use a team-oriented orchestration prompt along these lines:
- “Restate the observation exactly.”
- “Generate 3 deliberately different hypotheses.”
- “Create one tracer lane per hypothesis using Claude built-in team mode.”
- “For each lane, gather evidence for and against, rank evidence strength, and name the critical unknown plus best discriminating probe.”
- “Apply systems, premortem, and science lenses to the leaders if useful.”
- “Run a rebuttal round between the top two explanations.”
- “Return a ranked explanation table, convergence notes, the critical unknown, and the single best discriminating probe.”
Output quality bar
Good /trace output is:
- evidence-backed
- concise but rigorous
- skeptical of premature certainty
- explicit about missing evidence
- practical about the next action
- explicit about why weaker explanations were down-ranked
Example final synthesis shape
Observed Result
[What happened]
Ranked Hypotheses
| Rank | Hypothesis | Confidence | Evidence Strength | Why it leads |
|---|---|---|---|---|
| 1 | ... | High / Medium / Low | Strong / Moderate / Weak | ... |
Evidence Summary by Hypothesis
- Hypothesis 1: ...
- Hypothesis 2: ...
- Hypothesis 3: ...
Evidence Against / Missing Evidence
- Hypothesis 1: ...
- Hypothesis 2: ...
- Hypothesis 3: ...
Rebuttal Round
- Best rebuttal to leader: ...
- Why leader held / failed: ...
Convergence / Separation Notes
- ...
Most Likely Explanation
[Current best explanation]
Critical Unknown
[Single missing fact keeping uncertainty open]
Recommended Discriminating Probe
[Single next probe]
Additional Trace Lanes
[Only if uncertainty remains high]