setup
Setup GBrain
Set up GBrain from scratch. Target: working brain in under 5 minutes.
Contract
- Setup completes with a working brain verified by
gbrain doctor --json(all checks OK). - The brain-first lookup protocol is injected into the project's AGENTS.md or equivalent.
- Live sync is configured and verified (a test change pushed and found via search).
- Schema state is tracked in
~/.gbrain/update-state.jsonso future upgrades know what the user adopted or declined. - No Supabase anon key is requested; GBrain uses only the database connection string.
Install (if not already installed)
bun add github:garrytan/gbrain
How GBrain connects
GBrain connects directly to Postgres over the wire protocol. NOT through the
Supabase REST API. You need the database connection string (a postgresql:// URI),
not the project URL or anon key. The password is embedded in the connection string.
Use the Shared Pooler connection string (port 6543), not the direct connection (port 5432). The direct hostname resolves to IPv6 only, which many environments can't reach. Find it: go to the project, click Get Connected next to the project URL, then Direct Connection String > Session Pooler, and copy the Shared Pooler connection string.
Do NOT ask for the Supabase anon key. GBrain doesn't use it.
Why Supabase
Supabase gives you managed Postgres + pgvector (vector search built in) for $25/mo:
- 8GB database + 100GB storage on Pro tier
- No server to manage, automatic backups, dashboard for debugging
- pgvector pre-installed, just works
- Alternative: any Postgres with pgvector extension (self-hosted, Neon, Railway, etc.)
Prerequisites
- A Supabase account (Pro tier recommended, $25/mo) OR any Postgres with pgvector
- An OpenAI API key (for semantic search embeddings, ~$4-5 for 7,500 pages)
- A git-backed markdown knowledge base (or start fresh)
Available init options
gbrain init --supabase-- interactive wizard (prompts for connection string)gbrain init --url <connection_string>-- direct, no promptsgbrain init --non-interactive --url <connection_string>-- for scripts/agentsgbrain doctor --json-- health check after init
There is no --local, --sqlite, or offline mode. GBrain requires Postgres + pgvector.
Phase A: Supabase Setup (recommended)
Guide the user through creating a Supabase project:
- "Go to https://supabase.com and sign up or log in."
- "Click 'New Project' in the top left."
- Name:
gbrain - Region: pick the one closest to you
- Database password: generate a strong one and save it
- Name:
- "Wait about 2 minutes for the project to initialize."
- "Find the connection string: go to your project, click Get Connected next to the project URL, then Direct Connection String > Session Pooler, and copy the Shared Pooler connection string (port 6543)."
- Initialize gbrain:
gbrain init --non-interactive --url "postgresql://postgres.[ref]:[password]@aws-0-[region].pooler.supabase.com:6543/postgres" - Verify:
gbrain doctor --json
OpenClaw/Hermes agent instructions: Store the Supabase access token in your persistent
env as SUPABASE_ACCESS_TOKEN. gbrain doesn't store it, you need it for future
gbrain doctor runs. Generate at: https://supabase.com/dashboard/account/tokens
Phase B: BYO Postgres (alternative)
If the user already has Postgres with pgvector:
- Get the connection string from the user.
- Run:
gbrain init --non-interactive --url "<connection_string>" - Verify:
gbrain doctor --json
If the connection fails with ECONNREFUSED and the URL contains supabase.co,
the user probably pasted the direct connection (IPv6 only). Guide them to the
Session pooler string instead (see Phase A step 4).
Phase C: First Import
- Discover markdown repos. Scan the environment for git repos with markdown content.
echo "=== GBrain Environment Discovery ==="
for dir in /data/* ~/git/* ~/Documents/* 2>/dev/null; do
if [ -d "$dir/.git" ]; then
md_count=$(find "$dir" -name "*.md" -not -path "*/node_modules/*" -not -path "*/.git/*" 2>/dev/null | wc -l | tr -d ' ')
if [ "$md_count" -gt 10 ]; then
total_size=$(du -sh "$dir" 2>/dev/null | cut -f1)
echo " $dir ($total_size, $md_count .md files)"
fi
fi
done
echo "=== Discovery Complete ==="
-
Import the best candidate. For large imports (>1000 files), use nohup to survive session timeouts:
nohup gbrain import <dir> --no-embed --workers 4 > /tmp/gbrain-import.log 2>&1 &Then check progress:
tail -1 /tmp/gbrain-import.logFor smaller imports, run directly:
gbrain import <dir> --no-embed -
Prove search works. Pick a semantic query based on what you imported:
gbrain search "<topic from the imported data>"This is the magical moment: the user sees search finding things grep couldn't.
-
Start embeddings. Refresh stale embeddings (runs in background). Keyword search works NOW, semantic search improves as embeddings complete.
-
Backfill the knowledge graph. Populate typed links and structured timeline from the imported pages. Auto-link maintains both going forward, but historical pages need a one-time backfill.
gbrain extract links --source db --dry-run | head -20 # preview gbrain extract links --source db # commit gbrain extract timeline --source db # dated events gbrain stats # verify links > 0After this,
gbrain graph-query <slug> --depth 2works and search ranks well-connected entities higher. Idempotent — safe to re-run anytime. Supports--since YYYY-MM-DDfor incremental runs on huge brains.Skip if Phase C imported zero pages (auto-link handles new writes).
-
Offer file migration. If the repo has binary files (.raw/ directories with images, PDFs, audio):
"You have N binary files (X GB) in your brain repo. Want to move them to cloud storage? Your git repo will drop from X GB to Y MB. All links keep working."
If the user agrees, configure storage and run migration:
# Configure storage backend (Supabase Storage recommended) gbrain config set storage.backend supabase gbrain config set storage.bucket brain-files gbrain config set storage.projectUrl <supabase-url> gbrain config set storage.serviceRoleKey <service-role-key> # Migrate binary files to cloud (3-step lifecycle) gbrain files mirror <brain-dir> # Upload to cloud, keep local gbrain files redirect <brain-dir> # Replace local with .redirect.yaml pointers # (optional) gbrain files clean <brain-dir> --yes # Remove pointers tooAfter migration,
gbrain files upload-rawhandles new files automatically: small text/PDFs stay in git, large/media files go to cloud with.redirect.yamlpointers. Files >= 100 MB use TUS resumable upload for reliability.
If no markdown repos are found, create a starter brain with a few template pages (a person page, a company page, a concept page) from docs/GBRAIN_RECOMMENDED_SCHEMA.md.
Phase C.5: One-step autopilot + Minions install (v0.11.1+)
Run the migration runner once, then install autopilot. Two commands, done:
gbrain apply-migrations --yes # applies any pending migrations; idempotent on healthy installs
gbrain autopilot --install # supervises itself + forks the Minions worker; env-aware
What gbrain autopilot --install does:
- On macOS: writes a launchd plist at
~/Library/LaunchAgents/com.gbrain.autopilot.plist. - On Linux with systemd: writes
~/.config/systemd/user/gbrain-autopilot.servicewithRestart=on-failure. - On ephemeral containers (Render / Railway / Fly / Docker): writes
~/.gbrain/start-autopilot.shand prints the one-line your agent's bootstrap should source to launch autopilot on every container start. Auto-injects into OpenClaw'shooks/bootstrap/ensure-services.shif detected (use--no-injectto opt out). - On Linux without systemd: installs a crontab entry (every 5 min).
Autopilot then supervises the Minions worker as a child process. Users get
sync + extract + embed + backlinks + durable Postgres-backed job processing
from ONE install step. No separate gbrain jobs work daemon to manage.
On PGLite, autopilot runs inline (PGLite's exclusive file lock blocks a separate worker process). Everything else still works.
If apply-migrations prints "N host-specific items need your agent's
attention," read ~/.gbrain/migrations/pending-host-work.jsonl + walk
skills/migrations/v0.11.0.md + docs/guides/plugin-handlers.md to
register host-specific handlers. Re-run apply-migrations after each
batch.
Phase D: Brain-First Lookup Protocol
Inject the brain-first lookup protocol into the project's AGENTS.md (or equivalent). This replaces grep-based knowledge lookups with structured gbrain queries.
BEFORE (grep) vs AFTER (gbrain)
| Task | Before (grep) | After (gbrain) |
|---|---|---|
| Find a person | grep -r "Pedro" brain/ |
gbrain search "Pedro" |
| Understand a topic | grep -rl "deal" brain/ | head -5 && cat ... |
gbrain query "what's the status of the deal" |
| Read a known page | cat brain/people/pedro.md |
gbrain get people/pedro |
| Find connections | grep -rl "Brex" brain/ | xargs grep "Pedro" |
gbrain query "Pedro Brex relationship" |
Lookup sequence (MANDATORY for every entity question)
gbrain search "name"-- keyword match, fast, works without embeddingsgbrain query "what do we know about name"-- hybrid search, needs embeddingsgbrain get <slug>-- direct page read when you know the slug from steps 1-2grepfallback -- only if gbrain returns zero results AND the file may exist outside the indexed brain
Stop at the first step that gives you what you need. Most lookups resolve at step 1.
Sync-after-write rule
After creating or updating any brain page in the repo, sync immediately so the index stays current:
gbrain sync --no-pull --no-embed
This indexes new/changed files without pulling from git or regenerating embeddings.
Embeddings can be refreshed later in batch (gbrain embed --stale).
gbrain vs memory_search
| Layer | What it stores | When to use |
|---|---|---|
| gbrain | World knowledge: people, companies, deals, meetings, concepts, media | "Who is Pedro?", "What happened at the board meeting?" |
| memory_search | Agent operational state: preferences, decisions, session context | "How does the user like formatting?", "What did we decide about X?" |
Both should be checked. gbrain for facts about the world. memory_search for how the agent should behave.
Phase E: Load the Production Agent Guide
Read docs/GBRAIN_SKILLPACK.md. This is the reference architecture for how a
production agent uses gbrain: the brain-agent loop, entity detection, enrichment
pipeline, meeting ingestion, cron schedules, and the five operational disciplines.
Inject the key patterns into the agent's system context or AGENTS.md:
- Brain-agent loop (Section 2): read before responding, write after learning
- Entity detection (Section 3): spawn on every message, capture people/companies/ideas
- Source attribution (Section 7): every fact needs
[Source: ...]
Convention: See
skills/conventions/quality.mdfor Iron Law back-linking.
Tell the user: "The production agent guide is at docs/GBRAIN_SKILLPACK.md. It covers the brain-agent loop, entity detection, enrichment, meeting ingestion, and cron schedules. Read it when you're ready to go from 'search works' to 'the brain maintains itself.'"
Phase F: Health Check
Run gbrain doctor --json and report the results. Every check should be OK.
If any check fails, the doctor output tells you exactly what's wrong and how to fix it.
Error Recovery
If any gbrain command fails, run gbrain doctor --json first. Report the full
output. It checks connection, pgvector, RLS, schema version, and embeddings.
| What You See | Why | Fix |
|---|---|---|
| Connection refused | Supabase project paused, IPv6, or wrong URL | Use Session pooler (port 6543), or supabase.com/dashboard > Restore |
| Password authentication failed | Wrong password | Project Settings > Database > Reset password |
| pgvector not available | Extension not enabled | Run CREATE EXTENSION vector; in SQL Editor |
| OpenAI key invalid | Expired or wrong key | platform.openai.com/api-keys > Create new |
| No pages found | Query before import | Import files into gbrain first |
| RLS not enabled | Security gap | Run gbrain init again (auto-enables RLS) |
Phase G: Auto-Update Check (if not already configured)
If the user's install did NOT include setting up auto-update checks (e.g., they used the manual install path or an older version of the OpenClaw/Hermes paste), offer it:
"Would you like daily GBrain update checks? I'll let you know when there's a new version worth upgrading to — including new skills and schema recommendations. You'll always be asked before anything is installed."
If they agree:
- Test:
gbrain check-update --json - Register daily cron (see GBRAIN_SKILLPACK.md Section 17)
If already configured or user declines, skip.
Phase H: Live Sync Setup (MUST ADD)
The brain repo is the source of truth. If sync doesn't run automatically, the vector DB falls behind and gbrain returns stale answers. This phase is not optional.
Read docs/GBRAIN_SKILLPACK.md Section 18 for the full reference. Key points:
-
Check the connection pooler first. Sync uses transactions on every import. If
DATABASE_URLuses Supabase's Transaction mode pooler, sync will throw.begin() is not a functionand silently skip most pages. Verify the connection string uses Session mode (port 6543, Session mode) or direct (port 5432). -
Set up automatic sync. Choose the approach that fits your environment:
- Cron (recommended for agents): register a cron every 5-30 minutes:
gbrain sync --repo /data/brain && gbrain embed --stale - Watch mode:
gbrain sync --watch --repo /data/brainunder a process manager. Pair with a cron fallback (watch exits after 5 consecutive failures). - Webhook or git hook: if available in your environment.
- Cron (recommended for agents): register a cron every 5-30 minutes:
-
Verify sync works. Don't just check that the command ran. Check that it worked:
gbrain statsshould show page count close to syncable file count in the repo.- If page count is way too low, the pooler bug is silently skipping pages.
- Push a test change and confirm it appears in
gbrain search.
-
Chain sync + embed. Always run both:
gbrain sync --repo <path> && gbrain embed --stale. For small syncs, embeddings are generated inline. Theembed --staleis a safety net for any stale chunks.
Tell the user: "Live sync is configured. The brain will stay current automatically. I'll verify it's working in the next phase."
Phase I: Full Verification
Run the full verification runbook to confirm the entire installation is working.
- Read
docs/GBRAIN_VERIFY.md - Execute each check in order
- Report results to the user
- Fix any failures before declaring setup complete
Every check in the runbook should pass. The most important one is check 4 (live sync actually works): push a change, wait for sync, search for the corrected text. "Sync ran" is not the same as "sync worked."
Tell the user: "I've verified the full GBrain installation. Here's the status of each check: [list results]. Everything is working / [specific item] needs attention."
If already configured or user declines, skip.
Schema State Tracking
After presenting the recommended directories (Phase C/E) and the user selects which
ones to create, write ~/.gbrain/update-state.json recording:
schema_version_applied: current gbrain versionskillpack_version_applied: current gbrain versionschema_choices.adopted: directories the user createdschema_choices.declined: directories the user explicitly skippedschema_choices.custom: directories the user added that aren't in the recommended schema
This file enables future upgrades to suggest new schema additions without re-suggesting things the user already declined.
Anti-Patterns
- Asking for the Supabase anon key. GBrain connects directly to Postgres over the wire protocol, not through the REST API. Only the database connection string is needed.
- Skipping live sync setup. If sync doesn't run automatically, the vector DB falls behind and search returns stale answers. Phase H is not optional.
- Declaring setup complete without verification. "The command ran" is not the same as "it worked." Push a test change, wait for sync, search for the corrected text.
- Using Transaction mode pooler. Sync uses transactions on every import. Transaction mode pooler causes
.begin() is not a functionerrors and silently skips pages. Always use Session mode (port 6543). - Importing without proving search. The magical moment is the user seeing search find things grep couldn't. Don't skip it.
Output Format
GBRAIN SETUP COMPLETE
=====================
Engine: [PGLite / Supabase Postgres]
Connection: [verified / pooler mode confirmed]
Pages imported: N
Embeddings: N/N (keyword search active, semantic improving)
Live sync: [configured / method]
Health check: all OK / [specific failures]
Verification: [GBRAIN_VERIFY.md results]
Next steps:
- Read docs/GBRAIN_SKILLPACK.md for production agent patterns
- [any pending items]
Tools Used
gbrain init --non-interactive --url ...-- create braingbrain import <dir> --no-embed [--workers N]-- import filesgbrain search <query>-- search braingbrain doctor --json-- health checkgbrain check-update --json-- check for updatesgbrain embed refresh-- generate embeddingsgbrain embed --stale-- backfill missing embeddingsgbrain sync --repo <path>-- one-shot sync from brain repogbrain sync --watch --repo <path>-- continuous sync pollinggbrain config get sync.last_run-- check last sync timestampgbrain stats-- page count + embed coverage