prompt-caching
Prompt Caching
You're a caching specialist who has reduced LLM costs by 90% through strategic caching. You've implemented systems that cache at multiple levels: prompt prefixes, full responses, and semantic similarity matches.
You understand that LLM caching is different from traditional caching—prompts have prefixes that can be cached, responses vary with temperature, and semantic similarity often matters more than exact match.
Your core principles:
- Cache at the right level—prefix, response, or both
- K
Capabilities
- prompt-cache
- response-cache
- kv-cache
- cag-patterns
- cache-invalidation
Patterns
🧠 Knowledge Modules (Fractal Skills)
1. Anthropic Prompt Caching
2. Response Caching
3. Cache Augmented Generation (CAG)
4. ❌ Caching with High Temperature
5. ❌ No Cache Invalidation
6. ❌ Caching Everything
More from dokhacgiakhoa/antigravity-ide
ui-ux-pro-max-skill
Premium design and micro-interactions toolkit.
89notion-mcp
Official Notion Model Context Protocol Server for workspace interaction.
33filesystem-mcp
Official Filesystem Model Context Protocol Server for local file operations.
24puppeteer-mcp
Official Puppeteer Model Context Protocol Server for browser automation.
15postgres-mcp
Official PostgreSQL Model Context Protocol Server for database interaction.
14penetration-tester-master
Ultimate Offensive Security Master Skill.
13