code-review

SKILL.md

Code Review Skill

Review code for correctness, security, performance, testing, documentation, and architecture. Produces two outputs:

  1. Structured findings for Claude Code to act on
  2. REVIEW.md human-readable summary

Review Workflow

1. Determine Review Scope

Identify what's being reviewed:

  • Single file: Review that file
  • Directory: Review all relevant files
  • Diff/PR: Focus on changed lines with surrounding context
  • Entire codebase: Start with entry points, follow dependencies

2. Select Review Depth

Choose automatically based on context, or accept user override:

Depth When to Use Focus
Quick Small changes, trivial files, time-sensitive Critical issues only
Standard Most reviews, single files, typical PRs All categories, balanced
Deep Pre-production, security-sensitive, complex systems Exhaustive, security-focused

3. Detect Languages and Load References

Identify languages present, then load relevant reference files:

4. Use context7 MCP for Documentation

Query context7 when:

  • Reviewing unfamiliar library/framework usage
  • Checking if APIs are used correctly
  • Verifying deprecated patterns
  • Confirming best practices for specific versions

Example queries:

  • resolve tensorflow then get-library-docs for TensorFlow code
  • resolve tidyverse for R tidyverse patterns
  • resolve kubernetes for K8s manifest validation

5. Spawn Subagents for Parallel Review

Use subagents to parallelize review work:

Language Subagent (one per language detected):

Task: Review [language] code in [files] for idioms, patterns, and language-specific issues.
Focus: Style, idioms, language-specific performance, common pitfalls.
Reference: Load references/[language].md
Output: Structured findings list

Security Subagent:

Task: Analyze [files] for security vulnerabilities.
Focus: Injection, auth issues, secrets exposure, unsafe operations, dependency risks.
Output: Security findings with severity and remediation

Architecture Subagent:

Task: Review overall structure and design of [files/project].
Focus: Coupling, cohesion, separation of concerns, design patterns, testability.
Output: Architecture findings and recommendations

6. Review Categories

Each category produces findings with severity ratings.

Correctness

  • Logic errors and bugs
  • Edge cases not handled
  • Off-by-one errors
  • Null/undefined handling
  • Type mismatches
  • Race conditions

Security

  • Injection vulnerabilities (SQL, command, XSS)
  • Authentication/authorization flaws
  • Secrets in code
  • Unsafe deserialization
  • Path traversal
  • Dependency vulnerabilities

Performance

  • Algorithmic complexity issues
  • Unnecessary allocations
  • N+1 queries
  • Missing caching opportunities
  • Blocking operations
  • Memory leaks

Testing

  • Missing test coverage
  • Untested edge cases
  • Brittle tests
  • Missing integration tests
  • Inadequate mocking

Documentation

  • Missing function/class docstrings
  • Outdated comments
  • Unclear variable names
  • Missing README updates
  • Undocumented public APIs

Architecture

  • Tight coupling
  • God objects/functions
  • Circular dependencies
  • Layer violations
  • Missing abstractions
  • Poor separation of concerns

7. Classify Findings

Rate each finding:

Severity Definition Action
Critical Security vulnerability, data loss risk, crash in production Must fix before merge
Major Significant bug, performance issue, maintainability blocker Should fix before merge
Minor Code smell, style issue, minor inefficiency Fix when convenient
Nitpick Preference, very minor style, optional improvement Consider fixing

8. Generate Outputs

Output 1: Claude Code Action Format

Produce structured findings Claude Code can act on directly:

## File: [filepath]

### [Line X-Y]: [Brief title]
**Severity**: Critical|Major|Minor|Nitpick
**Category**: Correctness|Security|Performance|Testing|Documentation|Architecture

**Issue**: [Clear description of the problem]

**Current code**:
[relevant code snippet]

**Suggested fix**:
[corrected code snippet]

**Rationale**: [Why this change improves the code]

---

Group findings by file, ordered by severity (Critical first).

Output 2: REVIEW.md Human Summary

Write to REVIEW.md in the project root:

# Code Review Summary

**Reviewed**: [files/scope]
**Depth**: Quick|Standard|Deep
**Date**: [timestamp]

## Overview

[2-3 sentence summary of overall code quality and key concerns]

## Findings by Severity

### Critical ([count])
- [one-line summary with file:line reference]

### Major ([count])
- [one-line summary with file:line reference]

### Minor ([count])
- [one-line summary with file:line reference]

### Nitpicks ([count])
- [one-line summary with file:line reference]

## Category Breakdown

| Category | Critical | Major | Minor | Nitpick |
|----------|----------|-------|-------|---------|
| Correctness | X | X | X | X |
| Security | X | X | X | X |
| Performance | X | X | X | X |
| Testing | X | X | X | X |
| Documentation | X | X | X | X |
| Architecture | X | X | X | X |

## Recommendations

[Prioritized list of recommended actions]

## Positive Observations

[Things done well - important for balanced feedback]

9. Iterate on Critical/Major Issues

After generating outputs:

  1. If user confirms, apply fixes for Critical and Major issues
  2. Re-review changed code to verify fixes don't introduce new issues
  3. Update REVIEW.md with resolution status
Weekly Installs
1
GitHub Stars
8
First Seen
2 days ago
Installed on
amp1
cline1
opencode1
cursor1
kimi-cli1
codex1