autoresearch
Autoresearch for Skills
Most skills work about 70% of the time. The other 30% you get garbage. The fix isn't to rewrite the skill from scratch. It's to let an agent run it dozens of times, score every output, and tighten the prompt until that 30% disappears.
This skill adapts Andrej Karpathy's autoresearch methodology (autonomous experimentation loops) to Claude Code skills. Instead of optimizing ML training code, we optimize skill prompts.
the core job
Take any existing skill, define what "good output" looks like as binary yes/no checks, then run an autonomous loop that:
- Generates outputs from the skill using test inputs
- Scores every output against the eval criteria
- Mutates the skill prompt to fix failures
- Keeps mutations that improve the score, discards the rest
- Repeats until the score ceiling is hit or the user stops it
Output: An improved SKILL.md + results.tsv log + changelog.md of every mutation attempted + a live HTML dashboard you can watch in your browser.
before starting: gather context
STOP. Do not run any experiments until all fields below are confirmed with the user. Ask for any missing fields before proceeding.
- Target skill — Which skill do you want to optimize? (need the exact path to SKILL.md)
- Test inputs — What 3-5 different prompts/scenarios should we test the skill with? (variety matters — pick inputs that cover different use cases so we don't overfit to one scenario)
- Eval criteria — What 3-6 binary yes/no checks define a good output? (these are your "test questions" — see references/eval-guide.md for how to write good evals)
- Runs per experiment — How many times should we run the skill per mutation? Default: 5. (more runs = more reliable scores, but slower and more expensive. 5 is the sweet spot for most skills.)
- Run interval — How often should experiments cycle? Default: every 2 minutes. (shorter = faster iteration, but costs more)
- Budget cap — Optional. Max number of experiment cycles before stopping. Default: no cap (runs until you stop it).
step 1: read the skill
Before changing anything, read and understand the target skill completely.
- Read the full SKILL.md file
- Read any files in
references/that the skill links to - Identify the skill's core job, process steps, and output format
- Note any existing quality checks or anti-patterns already in the skill
Do NOT skip this. You need to understand what the skill does before you can improve it.
step 2: build the eval suite
Convert the user's eval criteria into a structured test. Every check must be binary — pass or fail, no scales.
Format each eval as:
EVAL [number]: [Short name]
Question: [Yes/no question about the output]
Pass condition: [What "yes" looks like — be specific]
Fail condition: [What triggers a "no"]
Rules for good evals:
- Binary only. Yes or no. No "rate 1-7" scales. Scales compound variability and give unreliable results.
- Specific enough to be consistent. "Is the text readable?" is too vague. "Are all words spelled correctly with no truncated sentences?" is testable.
- Not so narrow that the skill games the eval. "Contains fewer than 200 words" will make the skill optimize for brevity at the expense of everything else.
- 3-6 evals is the sweet spot. More than that and the skill starts parroting eval criteria back instead of actually improving.
See references/eval-guide.md for detailed examples of good vs bad evals.
Max score calculation:
max_score = [number of evals] × [runs per experiment]
Example: 4 evals × 5 runs = max score of 20.
step 3: generate the live dashboard
Before running any experiments, create a live HTML dashboard at autoresearch-[skill-name]/dashboard.html and open it in the browser.
The dashboard must:
- Auto-refresh every 10 seconds (reads from results.tsv)
- Show a score progression line chart (experiment number on X axis, pass rate % on Y axis)
- Show a colored bar for each experiment: green = keep, red = discard, blue = baseline
- Show a table of all experiments with: experiment #, score, pass rate, status, description
- Show per-eval breakdown: which evals pass most/least across all runs
- Show current status: "Running experiment [N]..." or "Idle"
- Use clean styling with soft colors (white background, pastel accents, clean sans-serif font)
Generate the dashboard as a single self-contained HTML file with inline CSS and JavaScript. Use Chart.js loaded from CDN for the line chart. The JS should fetch results.json (which you update after each experiment alongside results.tsv) and re-render.
Open it immediately after creating it: open dashboard.html (macOS) so the user can see it in their browser.
Update results.json after every experiment so the dashboard stays current. The JSON format:
{
"skill_name": "[name]",
"status": "running",
"current_experiment": 3,
"baseline_score": 70.0,
"best_score": 90.0,
"experiments": [
{
"id": 0,
"score": 14,
"max_score": 20,
"pass_rate": 70.0,
"status": "baseline",
"description": "original skill — no changes"
}
],
"eval_breakdown": [
{"name": "Text legibility", "pass_count": 8, "total": 10},
{"name": "Pastel colors", "pass_count": 9, "total": 10}
]
}
When the run finishes (user stops it or ceiling hit), update status to "complete" so the dashboard shows a "Done" state with final summary.
step 4: establish baseline
Run the skill AS-IS before changing anything. This is experiment #0.
- Create a working directory:
autoresearch-[skill-name]/inside the skill's folder - Create
results.tsvwith the header row - Create
results.jsonanddashboard.html, then open the dashboard in the browser - Back up the original SKILL.md as
SKILL.md.baseline - Run the skill [N] times using the test inputs
- Score every output against every eval
- Record the baseline score and update both results.tsv and results.json
results.tsv format (tab-separated):
experiment score max_score pass_rate status description
0 14 20 70.0% baseline original skill — no changes
IMPORTANT: After establishing baseline, confirm the score with the user before proceeding. If baseline is already 90%+, the skill may not need optimization — ask the user if they want to continue.
step 5: run the experiment loop
This is the core autoresearch loop. Once started, run autonomously until stopped.
LOOP:
-
Analyze failures. Look at which evals are failing most. Read the actual outputs that failed. Identify the pattern — is it a formatting issue? A missing instruction? An ambiguous directive?
-
Form a hypothesis. Pick ONE thing to change. Don't change 5 things at once — you won't know what helped.
Good mutations:
- Add a specific instruction that addresses the most common failure
- Reword an ambiguous instruction to be more explicit
- Add an anti-pattern ("Do NOT do X") for a recurring mistake
- Move a buried instruction higher in the skill (priority = position)
- Add or improve an example that shows the correct behavior
- Remove an instruction that's causing the skill to over-optimize for one thing at the expense of others
Bad mutations:
- Rewriting the entire skill from scratch
- Adding 10 new rules at once
- Making the skill longer without a specific reason
- Adding vague instructions like "make it better" or "be more creative"
-
Make the change. Edit SKILL.md with ONE targeted mutation.
-
Run the experiment. Execute the skill [N] times with the same test inputs.
-
Score it. Run every output through every eval. Calculate total score.
-
Decide: keep or discard.
- Score improved → KEEP. Log it. This is the new baseline.
- Score stayed the same → DISCARD. Revert SKILL.md to previous version. The change added complexity without improvement.
- Score got worse → DISCARD. Revert SKILL.md to previous version.
-
Log the result in results.tsv.
-
Repeat. Go back to step 1 of the loop.
NEVER STOP. Once the loop starts, do not pause to ask the user if you should continue. They may be away from the computer. Run autonomously until:
- The user manually stops you
- You hit the budget cap (if one was set)
- You hit 95%+ pass rate for 3 consecutive experiments (diminishing returns)
If you run out of ideas: Re-read the failing outputs. Try combining two previous near-miss mutations. Try a completely different approach to the same problem. Try removing things instead of adding them. Simplification that maintains the score is a win.
step 6: write the changelog
After each experiment (whether kept or discarded), append to changelog.md:
## Experiment [N] — [keep/discard]
**Score:** [X]/[max] ([percent]%)
**Change:** [One sentence describing what was changed]
**Reasoning:** [Why this change was expected to help]
**Result:** [What actually happened — which evals improved/declined]
**Failing outputs:** [Brief description of what still fails, if anything]
This changelog is the most valuable artifact. It's a research log that any future agent (or smarter future model) can pick up and continue from.
step 7: deliver results
When the user returns or the loop stops, present:
- Score summary: Baseline score → Final score (percent improvement)
- Total experiments run: How many mutations were tried
- Keep rate: How many mutations were kept vs discarded
- Top 3 changes that helped most (from the changelog)
- Remaining failure patterns (what the skill still gets wrong, if anything)
- The improved SKILL.md (already saved in place)
- Location of results.tsv and changelog.md for reference
output format
The skill produces four files in autoresearch-[skill-name]/:
autoresearch-[skill-name]/
├── dashboard.html # live browser dashboard (auto-refreshes)
├── results.json # data file powering the dashboard
├── results.tsv # score log for every experiment
├── changelog.md # detailed mutation log
└── SKILL.md.baseline # original skill before optimization
Plus the improved SKILL.md saved back to its original location.
results.tsv example:
experiment score max_score pass_rate status description
0 14 20 70.0% baseline original skill — no changes
1 16 20 80.0% keep added explicit instruction to avoid numbering in diagrams
2 16 20 80.0% discard tried enforcing left-to-right layout — no improvement
3 18 20 90.0% keep added color palette hex codes instead of vague "pastel" description
4 18 20 90.0% discard added anti-pattern for neon colors — no improvement
5 19 20 95.0% keep added worked example showing correct label formatting
example: optimizing a diagram-generator skill
Context gathered:
- Target skill:
~/.claude/skills/diagram-generator/SKILL.md - Test inputs: "OAuth flow diagram", "CI/CD pipeline", "microservices architecture", "user onboarding funnel", "database schema relationships"
- Evals: (1) All text legible and spelled correctly? (2) Uses only pastel/soft colors? (3) Linear layout — left-to-right or top-to-bottom? (4) Free of numbers, ordinals, and ordering?
- Runs per experiment: 10
- Max score: 40
Baseline run (experiment 0): Generated 10 diagrams. Scored each against 4 evals. Result: 32/40 (80%). Common failures: 3 diagrams had numbered steps, 2 had bright red elements, 3 had illegible small text.
Experiment 1 — KEEP (35/40, 87.5%): Change: Added "NEVER include step numbers, ordinal numbers (1st, 2nd), or any numerical ordering in diagrams" to the anti-patterns section. Result: Numbering failures dropped from 3 to 1. Other evals held steady.
Experiment 2 — DISCARD (34/40, 85%): Change: Added "All text must be minimum 14px font size." Result: Legibility improved by 1, but color compliance dropped by 2. Reverted.
Experiment 3 — KEEP (37/40, 92.5%):
Change: Replaced vague "pastel colors" instruction with specific hex codes: #A8D8EA, #AA96DA, #FCBAD3, #FFFFD2, #B5EAD7.
Result: Color eval went from 8/10 to 10/10. Other evals held.
Experiment 4 — DISCARD (37/40, 92.5%): Change: Added anti-pattern "Do NOT use red (#FF0000), orange (#FF8C00), or neon green (#39FF14)." Result: No change. The hex codes from experiment 3 already solved the color problem. Reverted to keep skill simpler.
Experiment 5 — KEEP (39/40, 97.5%): Change: Added a worked example showing a correct diagram with properly formatted labels (no numbers, pastel fills, left-to-right flow, legible text). Result: Hit 39/40. One remaining failure: a complex diagram with overlapping labels. Diminishing returns — stopped.
Final delivery:
- Baseline: 32/40 (80%) → Final: 39/40 (97.5%)
- 5 experiments, 3 kept, 2 discarded
- Top changes: specific hex codes for colors, explicit anti-numbering rule, worked example
- Remaining issue: very complex diagrams occasionally get overlapping labels (1/40 failure rate)
how this connects to other skills
What feeds into autoresearch:
- Any existing skill that needs optimization
- User-defined eval criteria (or help them define evals using the eval guide)
What autoresearch feeds into:
- The improved skill replaces the original
- The changelog can be passed to future models for continued optimization
- The eval suite can be reused whenever the skill is updated
the test
A good autoresearch run:
- Started with a baseline — never changed anything before measuring the starting point
- Used binary evals only — no scales, no vibes, no "rate this 1-10"
- Changed one thing at a time — so you know exactly what helped
- Kept a complete log — every experiment recorded, kept or discarded
- Improved the score — measurable improvement from baseline to final
- Didn't overfit — the skill got better at the actual job, not just at passing the specific test inputs
- Ran autonomously — didn't stop to ask permission between experiments
If the skill "passes" all evals but the actual output quality hasn't improved — the evals are bad, not the skill. Go back to step 2 and write better evals.