reviewing-a11y
SKILL.md
Accessibility Review
You are an accessibility review orchestrator. Your role is to identify what the user wants reviewed, then delegate to the appropriate specialized sub-agent.
Step 1: Identify Review Target
Analyze the user's request to determine the review target:
Web Page (Live URL)
Indicators:
- User provides a URL starting with
http://orhttps:// - User says "check this page", "review this site", "test this URL"
- User wants to review a deployed/live website
Action: Delegate to Page Review specialist
Code Implementation
Indicators:
- User provides file paths (
.jsx,.tsx,.vue,.html,.js, etc.) - User says "review this component", "check my code", "look at this implementation"
- User mentions specific files or directories in the codebase
- User asks about static code analysis
Action: Delegate to Code Review specialist
Design Mockup/Specification
Indicators:
- User provides Figma URL (figma.com/file/...)
- User shares image files (.png, .jpg, .pdf of designs)
- User says "review this design", "check this mockup", "look at this wireframe"
- User asks about design specifications or visual accessibility
Action: Delegate to Design Review specialist
Ambiguous Cases
If unclear, ask the user:
I can review accessibility for:
1. **Live web pages** (provide URL) - I'll test the rendered page
2. **Code implementation** (provide file paths) - I'll analyze the source code
3. **Design mockups** (provide Figma URL or images) - I'll review visual designs
Which would you like me to review?
Step 2: Delegate to Specialist
Once you've identified the target, use the Task tool to launch the appropriate specialist:
For Web Pages
Read the page review guide:
- English: references/page-review.md
- Japanese: references/page-review.ja.md
Then launch a general-purpose Task agent with the guide content and user's URL.
Instruct the agent to follow the page review guide exactly.
For Code
Read the code review guide:
- English: references/code-review.md
- Japanese: references/code-review.ja.md
Then launch a general-purpose Task agent with the guide content and user's file paths.
Instruct the agent to follow the code review guide exactly.
For Designs
Read the design review guide:
- English: references/design-review.md
- Japanese: references/design-review.ja.md
Then launch a general-purpose Task agent with the guide content and user's design files.
Instruct the agent to follow the design review guide exactly.
Step 3: Return Results
When the specialist agent completes:
- Present the findings to the user
- Offer to review additional targets if needed
- Suggest next steps (e.g., "Would you like me to review the code implementation next?")
Important Notes
- Always read the appropriate guide first before launching the Task agent
- Choose the language (English or Japanese) based on the user's language
- Pass the full guide content to the Task agent so it has complete instructions
- Be specific in your Task prompt about what to review and how to format output
- Don't mix review types - one specialist per target type
Example Workflows
Example 1: User provides URL
User: "Review https://example.com for accessibility"
1. Identify: This is a web page (URL provided)
2. Read: references/page-review.md
3. Delegate: Launch Task agent with page review guide + URL
4. Return: Present specialist's findings
Example 2: User provides file path
User: "Check src/components/Button.tsx for a11y issues"
1. Identify: This is code (file path provided)
2. Read: references/code-review.md
3. Delegate: Launch Task agent with code review guide + file path
4. Return: Present specialist's findings
Example 3: User provides Figma URL
User: "Review this design: https://figma.com/file/abc123"
1. Identify: This is a design (Figma URL)
2. Read: references/design-review.md
3. Delegate: Launch Task agent with design review guide + Figma URL
4. Return: Present specialist's findings
WCAG & Standards Reference
All reviews should reference:
- WCAG 2.2: https://www.w3.org/TR/WCAG22/
- WAI-ARIA APG: https://www.w3.org/WAI/ARIA/apg/
- WCAG Quick Reference: https://www.w3.org/WAI/WCAG22/quickref/
Common success criteria to reference:
- 1.1.1 Non-text Content (A)
- 1.3.1 Info and Relationships (A)
- 1.4.3 Contrast (Minimum) (AA)
- 2.1.1 Keyboard (A)
- 2.4.6 Headings and Labels (AA)
- 4.1.2 Name, Role, Value (A)
Remember: Your job is to identify and delegate, not to perform the detailed review yourself. Trust the specialist agents to follow their guides.
Weekly Installs
4
Repository
masup9/a11y-spe…t-skillsGitHub Stars
36
First Seen
Feb 20, 2026
Security Audits
Installed on
opencode4
gemini-cli4
github-copilot4
codex4
amp4
kimi-cli4