claims-extractor
SKILL.md
Claims Extractor (peer review)
Goal: turn a manuscript into an auditable list of claims that downstream skills can check.
Inputs
Required:
output/PAPER.md(or equivalent plain-text manuscript)
Optional:
DECISIONS.md(review scope or constraints)
Outputs
output/CLAIMS.md
Output format (recommended)
For each claim, include at minimum:
Claim: one sentenceType:empirical|conceptualScope: what the claim applies to / what it does not apply toSource: a locatable pointer intooutput/PAPER.md(section + page/figure/table + a short quote)
Workflow
- If
DECISIONS.mdexists, apply any review scope/format constraints. - Read the manuscript (
output/PAPER.md) end-to-end (at least abstract + intro + method + experiments + limitations). - Extract:
- primary contributions (what is new)
- key claims (what is asserted)
- assumptions (what must be true for claims to hold)
- Normalize each item into one sentence.
- Attach a source pointer for every item.
- Split into two sections:
- Empirical claims (must be backed by experiments/data)
- Conceptual claims (must be backed by argument/definition)
Definition of Done
-
output/CLAIMS.mdexists. - Every claim has a source pointer that can be located in
output/PAPER.md. - Empirical vs conceptual claims are clearly separated.
Troubleshooting
Issue: the paper is only a PDF or HTML
Fix:
- Convert/extract it into a plain-text
output/PAPER.mdfirst (even rough extraction is OK), then run claim extraction.
Issue: claims are vague (“significant”, “better”, “state-of-the-art”)
Fix:
- Rewrite each claim to include the measurable dimension (metric/dataset/baseline) or mark it as “underspecified” with a note.
Weekly Installs
24
Repository
willoscar/resea…e-skillsGitHub Stars
301
First Seen
Jan 23, 2026
Security Audits
Installed on
claude-code20
gemini-cli20
cursor18
opencode18
codex18
antigravity15