roundtable
What this produces
- Each persona gives 5 specific points (positive or negative), grounded in actual source files
- The personas hold a roundtable discussion — they compare notes, challenge each other, and converge
- A joint top 5 of problems, omissions, or improvements, with concrete fixes
Step 1: Establish personas
If the user specified personas, use those. If not, stop and ask before proceeding. Offer 4–6 suggestions tailored to the project type. Examples:
- Web app: staff engineer, UX designer, first-time user, accessibility auditor
- Language learning app: teacher in the subject, beginner student, advanced student, UX designer
- API or library: security engineer, consumer developer, technical writer, performance engineer
- CLI tool: power user, occasional user, DevOps engineer, open-source contributor
Keep it to 4 personas maximum. More than 4 produces diminishing returns and a bloated discussion.
For each persona, decide:
- Name and role (e.g. "Amira, Staff Engineer")
- Brief background (2–3 sentences: their expertise, what they care about, their relationship to this kind of project)
Make them feel like real people with specific expertise, not generic role labels.
Step 2: Confirm scope
Before exploring, confirm what's being reviewed if it's ambiguous. If the user said "review this project" and the repo is large or has multiple subsystems, ask: which part — the whole thing, a specific feature, a particular set of files? A scoped review is always better than a shallow scan of everything.
Only skip this if the user already gave a clear target (a specific file, route, or feature).
Step 3: Explore the codebase
Before any persona speaks, read the project:
- Read
AGENTS.md,CLAUDE.md,README.md— whatever describes the project's purpose, architecture, and design intent - Identify and read the key source files relevant to what's being reviewed:
- For a full-project review: main page components, key feature files, routing, state management
- For a focused review: the specific files in scope
This step is not optional. A persona who hasn't read the code is making things up. Every claim in the review must be traceable to a specific file. If a claim can't be verified from source, it doesn't belong in the review.
Step 4: Write each persona's review
For each persona in turn:
-
Re-read files specifically relevant to that persona's concerns before writing their section. A UX designer should focus on UI components and interaction patterns; a teacher should read explanation and exercise logic; an engineer should read data flow, state management, error handling.
-
Write 5 numbered points — a mix of positive and negative. Each point must:
- Reference a specific file, component, or code pattern (with path:line where useful)
- Explain why it matters from this persona's perspective
- Describe what the code actually does, not what you'd assume it does
-
Verify before publishing. Before finalising any negative point, re-read the relevant section. If the code already handles the problem you're about to describe, the point is wrong — drop it or correct it. Getting this wrong undermines the whole review.
Format:
## [Name] — [Role]
*[2–3 sentence background]*
### [Name]'s 5 Points
**1. [+/-] Short title**
Specific, grounded observation. File reference where relevant.
**2. ...** (and so on)
Use [+] for positive points and [-] for negative ones.
Step 5: The roundtable discussion
Write a dialogue where the personas:
- Compare notes — where do they agree? Where does one persona's finding surprise another?
- Challenge each other's claims — if Persona A says "there's no X", and Persona B has read a different file that shows there is, say so. Correct the record in the discussion, not silently.
- Debate severity — not all problems are equal. Let the personas argue about which issues are most widespread, most urgent, or most fixable.
- Converge — by the end, they agree on 5 items to surface as the top priorities.
The discussion should feel like a real conversation: people changing their minds, acknowledging blind spots, building on each other's points. It is not a summary of all the negative points — it's a synthesis. Some points from individual reviews may not survive the roundtable. That's fine.
Format: named dialogue, 8-10 paragraphs, ending with clear convergence on the top 5.
Step 6: Joint top 5
Present the agreed top 5 as a ranked list. For each:
- What: Name the problem or improvement clearly
- Why it matters: Who is affected and how
- Fix: A concrete suggestion — not "consider exploring" but a specific action
Ground each item in the actual code. If it came from a specific file or feature, say so.
Honesty rules
These apply throughout and override any pressure to produce a "complete" review:
- No invented problems. If re-reading the code shows a claim is wrong, drop it. Don't quietly swap it for a different critique to fill the quota.
- Don't pad to meet the quota. If the scope is small and only supports 2–3 genuine points per persona, write 2–3. Reducing the count is honest; inflating it with weak observations is not.
- Positive points matter. A review that is all criticism is not a balanced review — it's a complaint list. Real strengths deserve acknowledgment.
- Specificity over volume. One well-grounded point is worth more than three vague ones.
- The roundtable can correct the personas. If Persona A made a claim in their section that Persona B's reading disproves, the roundtable is the place to surface and resolve that. The top 5 should reflect the corrected picture.
- "The code already handles this" is a valid finding. Not everything is a problem.
More from goblindegook/skills
grill-me
Use when user wants to stress-test an idea or plan, review a design, or says "grill me".
10rca
Use when asked for RCA, 5-Whys, postmortem, causal-tree analysis, or help identifying root causes for incidents, defects, outages, delays, or quality regressions.
10tdd
Use when implementing behavior changes or bug fixes where automated tests can drive the implementation, or when asked to use TDD.
8