skill-factory
Skill Factory
Autonomous skill creation - just tell me what you need, I'll handle everything.
What This Does
You request a skill, I deliver a production-ready skill with guaranteed quality (score >= 8.0/10).
No decision-making required. No tool selection. No quality checking. Just results.
Anthropic's Official Best Practices
For comprehensive guidance on creating effective skills, see:
- references/overview.md - Complete overview of Agent Skills architecture, progressive disclosure, and how Skills work across different platforms (API, Claude Code, Agent SDK, claude.ai)
- references/quickstart.md - Quick tutorial on using pre-built Agent Skills in the Claude API with practical code examples
- references/best-practices.md - Detailed authoring best practices including core principles, skill structure, progressive disclosure patterns, workflows, evaluation strategies, and common patterns
- references/anthropic-best-practices.md - Quality scoring system (10/10 criteria) used by skill-factory
These references provide Anthropic's official guidance and are consulted during the quality assurance phase.
Usage
Simply describe the skill you need:
"Create a skill for Anchor development with latest docs and best practices"
"Create a React skill from react.dev with comprehensive examples"
"Create a skill for Solana transaction debugging workflows"
"Create a skill for writing technical documentation following company standards"
I will automatically:
- ✅ Analyze your request
- ✅ Select the optimal creation method
- ✅ Create the skill
- ✅ Run quality assurance loops (until score >= 8.0)
- ✅ Test with automated scenarios
- ✅ Deliver ready-to-use skill with stats
What You Get
✅ anchor-development skill ready!
📊 Quality Score: 8.9/10 (Excellent)
📝 Lines: 412 (using progressive disclosure)
📚 Coverage: 247 documentation pages
💡 Examples: 68 code samples
🧪 Test Pass Rate: 100% (15/15 scenarios)
📁 Location: ~/.claude/skills/anchor-development/
📦 Zip: ~/Downloads/anchor-development.zip
Try it: "How do I create an Anchor program?"
How It Works (Behind the Scenes)
Phase 1: Request Analysis (Automatic)
I analyze your request to determine:
Source Detection:
- Documentation URL/mention? → Automated scraping path
- "Latest docs", "current version"? → Automated path
- GitHub repository mention? → Automated path
- PDF/manual path? → Automated path
- Custom workflow/process description? → Manual TDD path
- Both documentation AND custom needs? → Hybrid path
Quality Requirements Extraction:
- "Best practices" → Enforce quality gates
- "Latest version" → Scrape current docs
- "Examples" → Ensure code samples included
- "Comprehensive" → Verify coverage completeness
Phase 2: Execution (Automatic)
Path A: Documentation-Based (Skill_Seekers)
Detected: Documentation source available
Method: Automated scraping with quality enhancement
Steps I take:
1. Check Skill_Seekers installation (install if needed)
2. Configure scraping parameters automatically
3. Run scraping with optimal settings
4. Monitor progress
5. Initial quality check
6. If score < 8.0: Run enhancement loop
7. Re-score until >= 8.0
8. Test with auto-generated scenarios
9. Package and deliver
Path B: Custom Workflows (Manual TDD)
Detected: Custom workflow/process
Method: Test-Driven Documentation (obra methodology)
Steps I take:
1. Create pressure test scenarios
2. Run baseline (without skill)
3. Document agent behavior
4. Write minimal skill addressing baseline
5. Test with skill present
6. Identify rationalizations/gaps
7. Close loopholes
8. Iterate until bulletproof
9. Package and deliver
Path C: Hybrid
Detected: Documentation + custom requirements
Method: Scrape then enhance
Steps I take:
1. Scrape documentation (Path A)
2. Identify gaps vs requirements
3. Fill gaps with TDD approach (Path B)
4. Unify and test as whole
5. Quality loop until >= 8.0
6. Package and deliver
Phase 3: Quality Assurance Loop (Automatic)
I enforce Anthropic best practices:
while quality_score < 8.0:
issues = analyze_against_anthropic_guidelines(skill)
if "vague_description" in issues:
improve_description_specificity()
if "missing_examples" in issues:
extract_or_generate_examples()
if "too_long" in issues:
apply_progressive_disclosure()
if "poor_structure" in issues:
reorganize_content()
quality_score = rescore()
Quality Criteria (Anthropic Best Practices):
- ✅ Description: Specific, clear, includes when_to_use
- ✅ Conciseness: <500 lines OR progressive disclosure
- ✅ Examples: Concrete code samples, not abstract
- ✅ Structure: Well-organized, clear sections
- ✅ Name: Follows conventions (lowercase, hyphens, descriptive)
Important: The quality assurance process consults references/best-practices.md for Anthropic's complete authoring guidelines and references/anthropic-best-practices.md for the 10-point scoring criteria.
Phase 4: Testing (Automatic)
I generate and run test scenarios:
# Auto-generate test cases from skill content
test_cases = extract_key_topics(skill)
for topic in test_cases:
query = f"How do I {topic}?"
# Test WITHOUT skill (baseline)
baseline = run_query_without_skill(query)
# Test WITH skill
with_skill = run_query_with_skill(query)
# Verify improvement
if not is_better(with_skill, baseline):
identify_gap()
enhance_skill()
retest()
Phase 5: Delivery (Automatic)
Package skill:
- Create skill directory structure
- Generate SKILL.md with frontmatter
- Create reference files (if using progressive disclosure)
- Add examples directory
- Create .zip for easy upload
- Install to ~/.claude/skills/ (if desired)
- Generate summary statistics
Progress Reporting
You'll see real-time progress:
🔍 Analyzing request...
✅ Detected: Documentation-based (docs.rs/anchor-lang)
✅ Requirements: Latest version, best practices, examples
🔄 Creating skill...
📥 Scraping docs.rs/anchor-lang... (2 min)
📚 Extracting 247 pages...
💾 Organizing content...
📊 Quality check: 7.4/10
⚠️ Issues found:
- Description too generic (fixing...)
- Missing examples in 4 sections (adding...)
- Some outdated patterns (updating...)
🔧 Enhancing skill...
✏️ Description improved
📝 Examples added
🔄 Patterns updated
📊 Quality check: 8.9/10 ✅
🧪 Testing...
✅ 15/15 scenarios passing
✅ anchor-development skill ready!
Dependencies
Required:
- Python 3.10+ (for quality scripts)
- bash (for automation scripts)
Optional (auto-installed when needed):
- Skill_Seekers (for documentation scraping)
- Will prompt for installation on first documentation-based request
- One-command setup:
scripts/install-skill-seekers.sh
Configuration
Environment variables (optional):
# Skill_Seekers installation path
export SKILL_SEEKERS_PATH="$HOME/Skill_Seekers"
# Output directory for created skills
export SKILL_OUTPUT_DIR="$HOME/.claude/skills"
# Minimum quality score (default: 8.0)
export MIN_QUALITY_SCORE="8.0"
# Auto-install dependencies (default: prompt)
export AUTO_INSTALL_DEPS="true"
Examples
Example 1: Documentation Skill
User: "Create a React skill from react.dev"
Agent:
🔍 Analyzing... → Documentation-based
🔄 Scraping react.dev... → 3 min
📊 Quality: 7.8 → 8.6 ✅
🧪 Testing: 12/12 passing ✅
✅ react-development skill ready (8.6/10)
Example 2: Custom Workflow Skill
User: "Create a skill for debugging Solana transaction failures"
Agent:
🔍 Analyzing... → Custom workflow (no docs to scrape)
📝 Using TDD methodology...
🧪 RED: Testing baseline...
✏️ GREEN: Writing skill...
🔄 REFACTOR: Closing loopholes...
📊 Quality: 8.3 ✅
✅ solana-transaction-debugging skill ready (8.3/10)
Example 3: Hybrid Skill
User: "Create an Anchor skill with docs plus custom debugging workflows"
Agent:
🔍 Analyzing... → Hybrid (docs + custom)
📥 Scraping docs.rs/anchor-lang... → 2 min
📝 Adding custom debugging workflows...
🔄 Integrating and testing...
📊 Quality: 8.9 ✅
✅ anchor-development skill ready (8.9/10)
Quality Guarantee
Every skill delivered by skill-factory:
- ✅ Scores >= 8.0/10 on Anthropic best practices
- ✅ Has concrete examples (not abstract)
- ✅ Follows structure conventions
- ✅ Tested with auto-generated scenarios
- ✅ Ready to use immediately
If quality < 8.0, I keep working until it reaches 8.0+
Troubleshooting
Skill_Seekers installation fails:
# Manual installation
git clone https://github.com/yusufkaraaslan/Skill_Seekers ~/Skill_Seekers
cd ~/Skill_Seekers
pip install -r requirements.txt
# Or use installation script
~/Projects/claude-skills/skill-factory/scripts/install-skill-seekers.sh
Quality score stuck below 8.0:
- I'll report what's blocking and suggest manual review
- Check references/anthropic-best-practices.md for criteria
- Run manual enhancement if needed
Want to understand methodology:
- See references/obra-tdd-methodology.md (testing approach)
- See references/anthropic-best-practices.md (quality criteria)
- See references/skill-seekers-integration.md (automation details)
Reference Files
Anthropic Official Documentation:
- references/overview.md - Agent Skills architecture, progressive disclosure, and platform details
- references/quickstart.md - Quick tutorial on using pre-built Agent Skills in the Claude API
- references/best-practices.md - Comprehensive authoring guidelines from Anthropic
- references/anthropic-best-practices.md - Quality scoring system (10/10 criteria)
Skill Factory Implementation Details:
- references/obra-tdd-methodology.md - Full TDD testing approach
- references/skill-seekers-integration.md - Automation documentation
- references/request-analysis.md - How requests are parsed
- references/quality-loops.md - Enhancement algorithms
Scripts Reference
Available helper scripts in scripts/ directory:
- check-skill-seekers.sh - Check if Skill_Seekers is installed
- install-skill-seekers.sh - One-command Skill_Seekers setup
- quality-check.py - Score any skill against Anthropic best practices
Usage examples:
# Check Skill_Seekers installation
./scripts/check-skill-seekers.sh
# Install Skill_Seekers
./scripts/install-skill-seekers.sh
# Quality check a skill
python3 ./scripts/quality-check.py /path/to/skill/SKILL.md
Philosophy
You don't want to:
- Navigate decision trees
- Choose between tools
- Check quality manually
- Test with subagents yourself
- Wonder if output is good
You want to:
- Describe what you need
- Get high-quality result
- Start using immediately
That's what skill-factory delivers.
Credits
Built on top of excellent tools:
- Skill_Seekers - Documentation scraping
- obra/superpowers-skills - TDD methodology
- Anthropic skill-creator - Best practices
Skill-factory orchestrates these tools with automatic quality assurance and testing.
Just tell me what skill you need. I'll handle the rest.
More from tenequm/skills
swift-macos
Comprehensive macOS app development with Swift 6.2, SwiftUI, SwiftData, Swift Concurrency, Foundation Models, Swift Testing, ScreenCaptureKit, and app distribution. Use when building native Mac apps, implementing windows/scenes/navigation/menus/toolbars, SwiftData models and queries, modern concurrency, on-device AI, testing, screen/audio capture, menu bar apps, AppKit bridges, login items, process monitoring, or App Store and Developer ID distribution. Triggers on macOS app, SwiftUI macOS, SwiftData, Swift concurrency, Foundation Models, Swift Testing, ScreenCaptureKit, screen capture, screen recording, AVFoundation, MenuBarExtra, NSViewRepresentable, notarize, login item, and process monitoring.
56react-typescript
Build React 19 applications with TypeScript. Covers Actions, Activity, use() hook, React Compiler, ref-as-prop, useEffectEvent, and strict TypeScript patterns. Use when creating components, managing state, typing props, handling events, using hooks, or working with React 19 features. Triggers on react, typescript, tsx, component types, hook types, react 19, react compiler, actions, use hook, useEffectEvent, activity, import defer.
45shadcn-tailwind
Build UIs with Tailwind CSS v4 and shadcn/ui. Covers CSS variables with OKLCH colors, component variants with CVA, responsive design, dark mode, and Tailwind v4.2 features. Supports Radix UI and Base UI primitives, CLI 3.0, and visual styles. Use when building interfaces with Tailwind, styling shadcn/ui components, implementing themes, or working with utility-first CSS. Triggers on tailwind, shadcn, utility classes, CSS variables, OKLCH, component styling, theming, dark mode, radix ui.
39python-dev
Opinionated Python development setup with uv + ty + ruff + pytest + just. Use when creating new Python projects, setting up pyproject.toml, configuring linting, type checking, testing, or build tooling. Triggers on "python project", "uv init", "pyproject.toml", "ruff config", "ty check", "pytest setup", "justfile", "python linting", "python formatting", "type checking python".
39privy-integration
Integrates Privy authentication, embedded wallets, and agent payment protocols into web and agentic apps. Covers React SDK (PrivyProvider, hooks, wagmi), Node.js SDK, smart wallets (ERC-4337), x402 and MPP machine payments, Tempo chain, and agentic wallets with policies. Use when setting up Privy auth, creating embedded or agentic wallets, adding x402 or MPP payments, integrating with Tempo, configuring wallet policies, or connecting Privy to MCP/Agent Auth flows.
36biome
Lint and format frontend code with Biome 2.4. Covers type-aware linting, GritQL custom rules, domains, import organizer, and migration from ESLint/Prettier. Use when configuring linting rules, formatting code, writing custom lint rules, or setting up CI checks. Triggers on biome, biome config, biome lint, biome format, biome check, biome ci, gritql, migrate from eslint, migrate from prettier, import sorting, code formatting, lint rules, type-aware linting, noFloatingPromises.
34