openclaw-memory-enhancer
SKILL.md
🧠 OpenClaw Memory Enhancer
Give OpenClaw long-term memory - remember important information across sessions and automatically recall relevant context for conversations.
Core Capabilities
| Capability | Description |
|---|---|
| 🔍 Semantic Search | Vector similarity search, understanding intent not just keywords |
| 📂 Auto Load | Automatically reads all files from memory/ directory |
| 💡 Smart Recall | Finds relevant historical memory during conversations |
| 🔗 Memory Graph | Builds connections between related memories |
| 💾 Local Storage | 100% local, no cloud, complete privacy |
| 🚀 Edge Optimized | <10MB memory, runs on Jetson/Raspberry Pi |
Quick Reference
| Task | Command (Edge Version) | Command (Standard Version) |
|---|---|---|
| Load memories | python3 memory_enhancer_edge.py --load |
python3 memory_enhancer.py --load |
| Search | --search "query" |
--search "query" |
| Add memory | --add "content" |
--add "content" |
| Export | --export |
--export |
| Stats | --stats |
--stats |
When to Use
Use this skill when:
- You want OpenClaw to remember things across sessions
- You need to build a knowledge base from chat history
- You're working on long-term projects that need context
- You want automatic FAQ generation from conversations
- You're running on edge devices with limited memory
Don't use when:
- Simple note-taking apps are sufficient
- You don't need cross-session memory
- You have plenty of memory and want maximum accuracy (use standard version)
Versions
Edge Version ⭐ Recommended
Best for: Jetson, Raspberry Pi, embedded devices
python3 memory_enhancer_edge.py --load
Features:
- Zero dependencies (Python stdlib only)
- Memory usage < 10MB
- Lightweight keyword + vector matching
- Perfect for resource-constrained devices
Standard Version
Best for: Desktop/server, maximum accuracy
pip install sentence-transformers numpy
python3 memory_enhancer.py --load
Features:
- Uses sentence-transformers for high-quality embeddings
- Better semantic understanding
- Memory usage 50-100MB
- Requires model download (~50MB)
Installation
Via ClawHub (Recommended)
clawhub install openclaw-memory-enhancer
Via Git
git clone https://github.com/henryfcb/openclaw-memory-enhancer.git \
~/.openclaw/skills/openclaw-memory-enhancer
Usage Examples
Command Line
# Load existing OpenClaw memories
cd ~/.openclaw/skills/openclaw-memory-enhancer
python3 memory_enhancer_edge.py --load
# Search for memories
python3 memory_enhancer_edge.py --search "voice-call plugin setup"
# Add a new memory
python3 memory_enhancer_edge.py --add "User prefers dark mode"
# Show statistics
python3 memory_enhancer_edge.py --stats
# Export to Markdown
python3 memory_enhancer_edge.py --export
Python API
from memory_enhancer_edge import MemoryEnhancerEdge
# Initialize
memory = MemoryEnhancerEdge()
# Load existing memories
memory.load_openclaw_memory()
# Search for relevant memories
results = memory.search_memory("AI trends report", top_k=3)
for r in results:
print(f"[{r['similarity']:.2f}] {r['content'][:100]}...")
# Recall context for a conversation
context = memory.recall_for_prompt("Help me check billing")
# Returns formatted memory context
# Add new memory
memory.add_memory(
content="User prefers direct results",
source="chat",
memory_type="preference"
)
OpenClaw Integration
# In your OpenClaw agent
from skills.openclaw_memory_enhancer.memory_enhancer_edge import MemoryEnhancerEdge
class EnhancedAgent:
def __init__(self):
self.memory = MemoryEnhancerEdge()
self.memory.load_openclaw_memory()
def process(self, user_input: str) -> str:
# 1. Recall relevant memories
memory_context = self.memory.recall_for_prompt(user_input)
# 2. Enhance prompt with context
enhanced_prompt = f"""
{memory_context}
User: {user_input}
"""
# 3. Call LLM with enhanced context
response = call_llm(enhanced_prompt)
return response
Memory Types
| Type | Description | Example |
|---|---|---|
daily_log |
Daily memory files | memory/2026-02-22.md |
capability |
Capability records | Skills, tools |
core_memory |
Core conventions | Important rules |
qa |
Question & Answer | Q: How to... A: You should... |
instruction |
Direct instructions | "Remember: always do X" |
solution |
Technical solutions | Step-by-step guides |
preference |
User preferences | "User likes dark mode" |
How It Works
Memory Encoding (Edge Version)
- Keyword Extraction: Extract important words from text
- Hash Vector: Map keywords to vector positions
- Normalization: L2 normalize the vector
- Storage: Save to local JSON file
Memory Retrieval
- Query Encoding: Convert query to same vector format
- Keyword Pre-filter: Fast filter by common keywords
- Similarity Calculation: Cosine similarity between vectors
- Ranking: Return top-k most similar memories
Privacy Protection
- All data stored locally in
~/.openclaw/workspace/knowledge-base/ - No network requests
- No external API calls
- No data leaves your device
Technical Specifications
Edge Version
Vector Dimensions: 128
Memory Usage: < 10MB
Dependencies: None (Python stdlib)
Storage Format: JSON
Max Memories: 1000 (configurable)
Query Latency: < 100ms
Standard Version
Vector Dimensions: 384
Memory Usage: 50-100MB
Dependencies: sentence-transformers, numpy
Storage Format: NumPy + JSON
Model Size: ~50MB download
Query Latency: < 50ms
Configuration
Edit these parameters in the code:
self.config = {
"vector_dim": 128, # Vector dimensions
"max_memory_size": 1000, # Max number of memories
"chunk_size": 500, # Content chunk size
"min_keyword_len": 2, # Minimum keyword length
}
Troubleshooting
No results found
# Lower the threshold
results = memory.search_memory(query, threshold=0.2) # Default 0.3
# Increase top_k
results = memory.search_memory(query, top_k=10) # Default 5
Memory limit reached
The system automatically removes oldest memories when limit is reached.
To increase limit:
self.config["max_memory_size"] = 5000 # Increase from 1000
Slow performance
- Use Edge version instead of Standard
- Reduce
max_memory_size - Use keyword pre-filtering (automatic)
Contributing
- Fork the repository
- Create a feature branch
- Make your changes
- Submit a Pull Request
License
MIT License - See LICENSE file for details.
Acknowledgments
- Built for the OpenClaw ecosystem
- Optimized for edge computing devices
- Inspired by long-term memory systems in AI
Not an official OpenClaw or Moonshot AI product.
Users must provide their own OpenClaw workspace and API keys.
Weekly Installs
1
Repository
aaaaqwq/agi-super-skillsGitHub Stars
13
First Seen
4 days ago
Security Audits
Installed on
zencoder1
amp1
cline1
openclaw1
opencode1
cursor1