review
Installation
SKILL.md
Summary
Use this skill for static content quality reviews that produce clear findings and practical improvements.
When to use
- Reviewing documents for clarity, correctness, and completeness.
- Reviewing code for duplication, maintainability, and readability risks.
- Reviewing datasets for obvious consistency or redundancy issues.
- Providing a second opinion before publishing or merging content.
When not to use
- Runtime debugging or execution failures that require running systems or logs.
- Security audits or compliance checks that need specialized security workflows.
- Tasks requiring direct implementation rather than analysis.
- Reviews that depend on unavailable external business context.
Inputs
- Content to review (document, code, or data).
- Requested focus (for example: clarity, duplication, completeness, consistency).
- Optional context: style guide, acceptance criteria, constraints, or target audience.
Outputs
- Structured review report with:
- Key findings (prioritized)
- Duplications
- Clarity/completeness gaps
- Accuracy/consistency issues
- Actionable improvements
- Open questions (if context is missing)
Procedure
- Clarify scope only if review focus is ambiguous.
- Inspect content for duplication, ambiguity, inconsistency, and omission.
- Apply provided standards or default community conventions.
- Produce a concise, prioritized report with actionable fixes.
- Separate confirmed findings from context-dependent questions.
Best practices and constraints
- Keep findings specific and evidence-based.
- Distinguish objective issues from style preferences.
- Prefer actionable recommendations over generic commentary.
- Do not claim runtime behavior; this skill is static analysis only.
Gotchas and edge cases
- Missing context can make valid patterns appear inconsistent; surface as open questions.
- Very small snippets can hide dependencies; avoid overconfident conclusions.
- Domain-specific correctness may require expert validation beyond this review.
Examples
- "Review this requirements doc for missing acceptance criteria and ambiguity."
- "Review this module for duplication and maintainability risks."
- "Review this dataset description for consistency and clarity."
Acceptance criteria
- Report is structured, prioritized, and actionable.
- Findings clearly separate confirmed issues from open questions.
- Recommendations are tied to observed content, not speculation.
Notes for agents
- Ask at most one clarifying question when scope is unclear.
- Cite exact sections or lines when available.
- Keep output concise and decision-oriented.
Limitations
- Does not execute code or validate runtime behavior.
- May miss highly domain-specific issues without additional context.
- Feedback quality depends on the completeness of provided inputs.
Related skills