skills/canonical/copilot-collections/documentation-verify

documentation-verify

Installation
SKILL.md

Documentation Accuracy Verification

Scope

Accuracy verification only: cross-reference documentation claims, commands, API names, and configuration keys against source code in the same repository. Flag anything that cannot be verified.

Inputs

  • Changed documentation files (from git diff). If git diff is unavailable or empty, use files explicitly provided for review; otherwise, treat all documentation files under docs/ as changed.
  • Full codebase.
  • Test files, configuration schemas.
  • Documentation structure.

Workflow

Follow this three-stage process to verify documentation accuracy:

Stage 1: Discovery Scan

Objective: Identify documentation claims and form initial hypotheses.

  1. Run git diff to list changed documentation files.
  2. Categorize changes into claim types (behaviour, CLI, API, config, examples, error messages, etc.).
  3. Form initial hypotheses: Supported, Unsupported, Speculative, Ambiguous, or Outdated.

For detailed categorization and hypothesis formation procedures, see references/verification_procedures.md → Discovery Scan.


Stage 2: Verification Pass

Objective: Verify every hypothesis with code evidence. Code is the source of truth.

CRITICAL: Complete verification before reporting any claims.

  1. Use at least two search strategies per claim (direct search, entrypoint tracing, test evidence, schema/validation search).
  2. Apply claim-type-specific verification checklists (behaviour, CLI, API, config, examples, errors, terminology).
  3. Document evidence for each finding (file paths, line numbers, search commands).
  4. Reclassify hypotheses based on verification results.
  5. Apply false-positive prevention rules.
  6. Cross-check documentation coverage.

For comprehensive verification checklists and classification rules, see references/verification_procedures.md → Verification Pass.


Stage 3: Report Generation

Objective: Present verified findings with evidence and conservative recommendations.

  1. Group findings by final classification (unsupported, outdated, incorrect, imprecise, speculative, inconclusive, no issues).
  2. Format each finding using the standard template (doc claim, verification checklist, code evidence, assessment, recommended action).
  3. Provide conservative change suggestions aligned with code reality.
  4. Link to specific code artifacts (files, functions, structs, line numbers).
  5. Integrate with other analysis findings (e.g., Diataxis compliance).

For report formatting template and change suggestion guidelines, see references/report_format.md.


Constraints

  • Complete the verification pass (Stage 2) before reporting any claims.
  • Provide code evidence for all documentation consistency claims.
  • Code is the source of truth: flag documentation that contradicts code behaviour, not vice versa.
  • Do not claim documentation is "unsupported by code" without verification evidence (at least two search strategies with explicit code search and no-match confirmation).
  • Do not report false positives.
  • Do not prefer "unsupported" when docs are vague or imprecise; use accurate classifications.
  • Do not recommend changing code to match docs as the primary action; only documentation should be adjusted to match code reality.

Output

Only verified findings with evidence make it to the final report. Do not include intermediate hypotheses or reasoning.

Weekly Installs
1
GitHub Stars
19
First Seen
Mar 27, 2026