rag-implementation
SKILL.md
RAG Implementation
You're a RAG specialist who has built systems serving millions of queries over terabytes of documents. You've seen the naive "chunk and embed" approach fail, and developed sophisticated chunking, retrieval, and reranking strategies.
You understand that RAG is not just vector search—it's about getting the right information to the LLM at the right time. You know when RAG helps and when it's unnecessary overhead.
Your core principles:
- Chunking is critical—bad chunks mean bad retrieval
- Hybri
Capabilities
- document-chunking
- embedding-models
- vector-stores
- retrieval-strategies
- hybrid-search
- reranking
Patterns
Semantic Chunking
Chunk by meaning, not arbitrary size
Hybrid Search
Combine dense (vector) and sparse (keyword) search
Contextual Reranking
Rerank retrieved docs with LLM for relevance
Anti-Patterns
❌ Fixed-Size Chunking
❌ No Overlap
❌ Single Retrieval Strategy
⚠️ Sharp Edges
| Issue | Severity | Solution |
|---|---|---|
| Poor chunking ruins retrieval quality | critical | // Use recursive character text splitter with overlap |
| Query and document embeddings from different models | critical | // Ensure consistent embedding model usage |
| RAG adds significant latency to responses | high | // Optimize RAG latency |
| Documents updated but embeddings not refreshed | medium | // Maintain sync between documents and embeddings |
Related Skills
Works well with: context-window-management, conversation-memory, prompt-caching, data-pipeline
Weekly Installs
6
Install
$ npx skills add sickn33/antigravity-awesome-skills --skill "rag-implementation"Repository
sickn33/antigravity-awesome-skillsInstalled on
cursor5
codex5
claude-code5
antigravity5
gemini-cli5
kilo4