learn

SKILL.md

/learn — Skill Extraction Workflow

Extract non-obvious discoveries into reusable skills that persist across sessions.

When to Use This Skill

Invoke /learn when you encounter:

  • Non-obvious debugging — Investigation that took significant effort, not in docs
  • Misleading errors — Error message was wrong, found the real cause
  • Workarounds — Found a limitation with a creative solution
  • Tool integration — Undocumented API usage or configuration
  • Trial-and-error — Multiple attempts before success
  • Repeatable workflows — Multi-step task you'd do again
  • User-facing automation — Reports, checks, or processes users will request

Workflow Phases

PHASE 1: Evaluate (Self-Assessment)

Before creating a skill, answer these questions:

  1. "What did I just learn that wasn't obvious before starting?"
  2. "Would future-me benefit from this being documented?"
  3. "Was the solution non-obvious from documentation alone?"
  4. "Is this a multi-step workflow I'd repeat?"

Continue only if YES to at least one question.

PHASE 2: Check Existing Skills

Search for related skills to avoid duplication:

# Check project skills
ls .claude/skills/ 2>/dev/null

# Check global skills
ls ~/.claude/skills/ 2>/dev/null

# Search for keywords
grep -r -i "KEYWORD" .claude/skills/ ~/.claude/skills/ 2>/dev/null

Outcomes:

  • Nothing related → Create new skill (continue to Phase 3)
  • Same trigger & fix → Update existing skill (bump version)
  • Partial overlap → Update with new variant

PHASE 3: Create Skill

Create the skill file at .claude/skills/[skill-name]/SKILL.md:

---
name: descriptive-kebab-case-name
description: |
  [CRITICAL: Include specific triggers in the description]
  - What the skill does
  - Specific trigger conditions (exact error messages, symptoms)
  - When to use it (contexts, scenarios)
author: Claude Code Academic Workflow
version: 1.0.0
argument-hint: "[expected arguments]"  # Optional
---

# Skill Name

## Problem
[Clear problem description — what situation triggers this skill]

## Context / Trigger Conditions
[When to use — exact error messages, symptoms, scenarios]
[Be specific enough that you'd recognize it again]

## Solution
[Step-by-step solution]
[Include commands, code snippets, or workflows]

## Verification
[How to verify it worked]
[Expected output or state]

## Example
[Concrete example of the skill in action]

## References
[Documentation links, related files, or prior discussions]

PHASE 4: Quality Gates

Before finalizing, verify:

  • Description has specific trigger conditions (not vague)
  • Solution was verified to work (tested)
  • Content is specific enough to be actionable
  • Content is general enough to be reusable
  • No sensitive information (credentials, personal data)
  • Skill name is descriptive and uses kebab-case

Output

After creating the skill, report:

Skill created: .claude/skills/[name]/SKILL.md
  Trigger: [when to use]
  Problem: [what it solves]

Example: Creating a Skill

User discovers that a specific package silently drops data:

---
name: silent-data-loss-detection
description: |
  Detect silent data loss when packages drop observations without warning.
  Use when: results seem wrong, sample size unexpectedly small, or comparing
  results between different tools.
author: Claude Code Academic Workflow
version: 1.0.0
---

# Silent Data Loss Detection

## Problem
Some data processing packages silently drop observations when inputs
have missing values, producing unexpected results.

## Context / Trigger Conditions
- Sample size is smaller than expected
- Results differ between tools processing the same data
- Model has inputs with potential missing values

## Solution
1. Check for missing value patterns before processing:
   - Count complete cases vs total rows
2. Explicitly handle missing values
3. Document the expected sample size

## Verification
Compare output count with input count — difference indicates dropped observations.
Weekly Installs
1
GitHub Stars
2
First Seen
8 days ago
Installed on
zencoder1
amp1
cline1
openclaw1
opencode1
cursor1