ollama-launch

SKILL.md

Ollama Launch

Ollama provides local LLM inference for RLM distillation (seal phase summarization) and embeddings.

When You Need This

  • Seal fails with Connection refused to 127.0.0.1:11434
  • RLM distillation shows [DISTILLATION FAILED] for new files
  • Any tool that calls the Ollama API locally

Pre-Flight Check

# Check if Ollama is already running
curl -sf http://127.0.0.1:11434/api/tags > /dev/null && echo "✅ Ollama running" || echo "❌ Ollama not running"

If running, you're done. If not, proceed.

Start Ollama

# Start Ollama in the background
ollama serve &>/dev/null &

# Wait and verify (2-3 seconds)
sleep 3
curl -sf http://127.0.0.1:11434/api/tags > /dev/null && echo "✅ Ollama ready" || echo "❌ Ollama failed to start"

Verify Model Available

For RLM distillation, the project uses model you define in .env

# List available models
ollama list

# If the model is missing, pull it
ollama pull qwen2:7b

Troubleshooting

Symptom Fix
Connection refused after start Wait longer (sleep 5), model may be loading
ollama: command not found Ollama not installed — ask user to install from https://ollama.com
Port 11434 already in use Another process on that port — lsof -i :11434 to identify

Integration Points

  • Learning Loop Seal Phase: RLM synthesis calls Ollama for distillation
  • RLM Factory: /rlm-factory:distill requires Ollama for batch summarization
  • Embeddings: Any tool that needs local vector embeddings
Weekly Installs
14
GitHub Stars
1
First Seen
10 days ago
Installed on
opencode14
github-copilot14
codex14
kimi-cli14
amp14
cline14