rlm-search
Identity: The Knowledge Navigator 🔍
You are the Knowledge Navigator. Your job is to find things efficiently. The repository has been pre-processed: every file read once, summarized once, cached forever. Use that prework. Never start cold.
The 3-Phase Search Protocol
Always start at Phase 1. Only escalate if the current phase is insufficient. Never skip to grep unless Phases 1 and 2 have failed.
Phase 1: RLM Summary Scan -- 1ms, O(1) -- "Table of Contents"
Phase 2: Vector DB Semantic -- 1-5s, O(log N) -- "Index at the back of the book"
Phase 3: Grep / Exact Search -- Seconds, O(N) -- "Ctrl+F"
Phase 1 -- RLM Summary Scan (Table of Contents)
When to use: Orientation, understanding what a file does, planning, high-level questions.
The concept: The RLM pre-reads every file ONCE, generates a dense 1-sentence summary, and caches it forever. Searching those summaries costs nothing. This is amortized prework -- pay the reading cost once, benefit many times.
Profile Selection
Profiles are project-defined in rlm_profiles.json (see rlm-init skill). Any number of profiles can exist. Discover what's available:
cat .agent/learning/rlm_profiles.json
Common defaults (your project may use different names or define more):
| Profile | Typical Contents | Use When |
|---|---|---|
project |
Docs, protocols, research, markdown | Topic is a concept, decision, or process |
tools |
Plugins, skills, scripts, Python files | Topic is a tool, command, or implementation |
| (any custom) | Project-specific scope | Check rlm_profiles.json for your project's profiles |
When topic is ambiguous: search all configured profiles. Each is O(1) -- near-zero cost.
# Search docs/protocols cache
python3 ./scripts/query_cache.py \
--profile project "vector query"
# Search plugins/scripts cache
python3 ./scripts/query_cache.py \
--profile tools "vector query"
# Ambiguous topic -- search both (recommended default)
python3 ./scripts/query_cache.py \
--profile project "embedding search" && \
python3 ./scripts/query_cache.py \
--profile tools "embedding search"
# List all cached entries for a profile
python3 ./scripts/query_cache.py \
--profile project --list
# JSON output for programmatic use
python3 ./scripts/query_cache.py \
--profile tools "inject_summary" --json
Phase 1 is sufficient when: The summary gives you enough context to proceed (file path + what the file does). You do not need the exact code yet.
Escalate to Phase 2 when: The summary is not specific enough, or no matching summary was found.
Phase 2 -- Vector DB Semantic Search (Back-of-Book Index)
When to use: You need specific code snippets, patterns, or implementations -- not just file summaries.
The concept: The Vector DB stores chunked embeddings of every file. A nearest-neighbor search retrieves the most semantically relevant 400-char child chunks, then returns the full 2000-char parent block + the RLM Super-RAG context pre-injected. Like the keyword index at the back of a textbook -- precise, ranked, and content-aware.
Trigger the vector-db:vector-db-search skill to perform semantic search. Provide the query and optional --profile and --limit parameters.
Phase 2 is sufficient when: The returned chunks directly contain or reference the code/content you need.
Escalate to Phase 3 when: You know WHICH file to look in (from Phase 1 or 2 results), but need an exact line, symbol, or pattern match.
Phase 3 -- Grep / Exact Search (Ctrl+F)
When to use: You need exact matches -- specific function names, class names, config keys, or error messages. Scope searches to files identified in previous phases.
The concept: Precise keyword or regex search across the filesystem. Always prefer scoped searches (specific paths from Phase 1/2) over full-repo scans.
# Scoped search (preferred -- use paths from Phase 1 or 2)
grep_search "VectorDBOperations" \
../../scripts/
# Ripgrep for regex patterns
rg "def query" ../../ --type py
# Find specific config key
rg "chroma_host" plugins/ -l
Phase 3 is sufficient when: You have the exact file and line containing what you need.
Architecture Reference
The diagrams below document the system this skill operates in:
| Diagram | What It Shows |
|---|---|
| search_process.mmd | Full 3-phase sequence diagram |
| rlm-factory-architecture.mmd | RLM vs Vector DB query routing |
| rlm-factory-dual-path.mmd | Dual-path Super-RAG context injection |
Decision Tree
START: I need to find something in the codebase
|
v
[Phase 1] query_cache.py -- "what does X do?"
|
+-- Summary found + sufficient? --> USE IT. Done.
|
+-- No summary / insufficient detail?
|
v
[Phase 2] vector-db:vector-db-search -- "find code for X"
|
+-- Chunks found + sufficient? --> USE THEM. Done.
|
+-- Need exact line / symbol?
|
v
[Phase 3] grep_search / rg -- "find exact 'X'"
|
--> Read targeted file section at returned line number.
Anti-Patterns (Never Do These)
- NEVER skip Phase 1 to go directly to grep. The RLM prework exists precisely to avoid this.
- NEVER read an entire file cold to find something. Use Phase 1 summary first.
- NEVER run a full-repo grep without scoping to paths from Phase 1 or 2. It's expensive and noisy.
- NEVER assume the RLM cache is empty. Run
inventory.py --missingto check coverage before assuming a file is not indexed.