entity-memory
Skill: Entity Memory
This is your persistent memory for everything you know about a record — contacts, companies, employees, members, locations, devices, and more. Store and retrieve: customer data, interaction history, inputs, actions you've taken, reports, notes, and observations. Always consult Memory before acting on a record. Always update Memory after meaningful interactions.
Internal principle: Bad data in = bad personalization out. Memory is the foundation. Get it right, and every downstream feature (emails, notifications, dashboards, agents) gets better automatically.
When This Skill is Activated
This skill gives you the ability to store and retrieve data using the Personize SDK's memory system.
If the developer hasn't given a specific instruction yet, introduce yourself:
"I have access to the Memory skill. I can help you store data into Personize memory (memorize) and retrieve it (recall) — including batch syncs, semantic search, entity digests, and data export. What data are you working with?"
If the developer says something about storing, syncing, importing, or ingesting data, jump to MEMORIZE.
If the developer says something about retrieving, querying, searching, or assembling context, jump to RECALL.
When NOT to Use This Skill
- For CRM sync with deploy templates → see the CRM / Database Sync section below
- Need no-code visual workflows → use no-code-pipelines
- Need durable scheduled pipelines with retries → use code-pipelines
- Need to manage organizational rules, not entity data → use governance
- Need multi-agent coordination state → use collaboration
Works With Both SDK and MCP — One Skill, Two Interfaces
This skill works identically whether the LLM accesses memory via the SDK (code, scripts, IDE agents) or via MCP (Claude Desktop, ChatGPT, Cursor MCP connection).
| Interface | How it works | Best for |
|---|---|---|
SDK (@personize/sdk) |
client.memory.memorize(), client.memory.recall(), etc. |
Scripts, CI/CD, IDE agents, recipes |
| MCP (Model Context Protocol) | memory_store_pro, memory_recall_pro, ai_smart_guidelines tools |
Claude Desktop, ChatGPT, Cursor, any MCP-compatible client |
MCP tools map to SDK methods:
| SDK Method | MCP Tool | Purpose |
|---|---|---|
client.memory.memorize(opts) |
memory_store_pro(content, email, ...) |
Store data with AI extraction |
client.memory.smartRecall(opts) |
memory_recall_pro(query, email, ...) |
Semantic search (recommended) |
client.memory.recall(opts) |
(SDK only) | Direct DynamoDB lookup — properties + freeform memories (type required, no AI) |
client.memory.smartDigest(opts) |
(SDK only) | Compiled entity context (properties + memories) |
client.memory.search(opts) |
(SDK only) | Filter and export records |
client.memory.memorizeBatch(opts) |
(SDK only) | Batch sync with per-property control |
client.ai.smartGuidelines(opts) |
ai_smart_guidelines(message) |
Fetch guidelines by topic |
MCP-Only Feature: Self-Memory (about='self')
MCP tools support an about parameter that the SDK does not expose directly:
about='lead'(default) — store/recall about a contact or company. Requiresemail,website_url, orrecord_id.about='self'— store/recall about the current user (preferences, working style, goals). No identifier needed — identity is resolved automatically.
// MCP: Store user preferences
memory_store_pro(content="I prefer formal communication. My timezone is PST.", about="self")
// MCP: Recall user preferences
memory_recall_pro(query="What are my preferences and working style?", about="self", generate_answer=true)
When reading this skill document:
- If you're connected via MCP, use the MCP tool names (
memory_store_pro,memory_recall_pro, etc.) - If you're running via SDK, use the
client.memory.*methods - All workflows, rules, and best practices apply equally to both interfaces
Actions
You have 2 actions. Use whichever matches what the developer needs.
| Action | When to Use | Reference |
|---|---|---|
| MEMORIZE | Developer needs to store data — single items, batch sync, CRM import, webhook data, generated outputs | reference/memorize.md |
| RECALL | Developer needs to retrieve data — semantic search, entity context, filtered exports, context assembly | reference/recall.md |
Before each action: Read the reference file for full method signatures, decision trees, code examples, and common mistakes.
Action: MEMORIZE
Store data into Personize memory. The right method depends on what you're storing and how much of it.
Which Method to Use
| Scenario | Method | Why |
|---|---|---|
| One item, with AI extraction | memory.memorize() |
Rich text (notes, transcripts, emails) → AI extracts facts and creates vectors |
| Batch sync from CRM/DB | memory.memorizeBatch() |
Multiple records with per-property extractMemories control |
| Structured data, no AI needed | memory.memorizeBatch() with extractMemories: false |
Store exact key-value pairs (email, plan_tier, login_count) without AI overhead |
The extractMemories Decision
extractMemories defaults to false. You must set extractMemories: true on rich text fields to enable AI extraction and semantic search. Without it, batch-synced data is stored as structured properties only — no memories, no vector embeddings, no semantic recall.
| Data Type | extractMemories |
Reasoning |
|---|---|---|
| Rich text (notes, transcripts, emails, descriptions) | true (must set explicitly) |
AI extracts facts, creates vector embeddings for semantic search |
| Generated content (AI outputs you want to remember) | true (must set explicitly) |
Enables the feedback loop — AI knows what it already said |
| ML outputs with explanations (churn reason, lead score rationale) | true (must set explicitly) |
The explanation text benefits from extraction |
| Structured facts (email, name, plan, dates, counts) | false (default) |
Already structured — AI extraction wastes tokens and adds latency |
| Binary flags, IDs, URLs | false (default) |
No semantic content to extract |
Rule of thumb: Always set
extractMemories: trueon any field containing free-form text. If you skip it, those fields get stored as properties but produce zero memories —smartRecall()andsmartDigest()won't find them.
Quick Example
// Single item — AI extraction with identity hints
await client.memory.memorize({
content: 'Also extract First Name, Last Name, Company Name, and Job Title if mentioned.\n\nCall with Sarah Chen (VP Eng, Initech). She mentioned they are evaluating SOC2 compliance tools. Main pain point: manual audit prep taking 2 weeks per quarter. Budget approved for Q2.',
speaker: 'Sales Team',
email: 'sarah.chen@initech.com',
enhanced: true,
tags: ['call-notes', 'sales', 'source:manual'],
});
// Batch sync — per-property control
await client.memory.memorizeBatch({
source: 'Hubspot',
mapping: {
entityType: 'contact',
email: 'email',
runName: 'hubspot-contact-sync',
properties: {
full_name: { sourceField: 'firstname', collectionId: 'col_xxx', collectionName: 'Contacts', extractMemories: false },
job_title: { sourceField: 'jobtitle', collectionId: 'col_xxx', collectionName: 'Contacts', extractMemories: false },
last_notes: { sourceField: 'notes', collectionId: 'col_xxx', collectionName: 'Contacts', extractMemories: true },
},
},
rows: crmContacts, // array of objects from your CRM
});
// ⚠️ memorizeBatch() is async — records land in ~1-2 minutes (EventBridge → Lambda).
// Verify with search() or smartDigest() after processing completes.
Intelligence Tiers (memorize + batch-memorize)
memorize() and memorizeBatch() accept a tier param that selects the LLM pipeline and credit rate. Defaults to pro — no breaking change for existing callers.
| Tier | Rate | Best For |
|---|---|---|
basic |
1 credit/1K tokens | High-volume, cost-first |
pro |
2.5 credits/1K tokens | Default — best quality/cost balance |
pro_fast |
3.5 credits/1K tokens | Speed-critical (~8s latency) |
ultra |
7 credits/1K tokens | Maximum extraction depth, 50 properties |
Tiers also control maxProperties (15–50), chunkMaxWords (2000–4000), and minPropertyScore (0.2–0.4). Model fallback is automatic — if the primary model fails, the system routes to the fallback. 1 credit = $0.01.
await client.memory.memorize({
content: '...',
email: 'user@co.com',
enhanced: true,
tier: 'basic', // or 'pro' (default), 'pro_fast', 'ultra'
});
await client.memory.memorizeBatch({
source: 'HubSpot',
mapping: { ... },
rows: contacts,
tier: 'pro_fast', // fast LLMs, lower latency
});
Note: These are memorize intelligence tiers — they control the extraction pipeline. For content generation via
client.ai.prompt(), see the separate generate tiers below.
Generate Tiers (prompt)
client.ai.prompt() has its own tier system for content generation, separate from memorize tiers:
| Tier | Input | Output | Best For |
|---|---|---|---|
basic |
0.2 cr/1K tokens | 0.4 cr/1K tokens | High-volume, cost-first |
pro |
0.5 cr/1K tokens | 1.0 cr/1K tokens | Default — balanced |
ultra |
1.0 cr/1K tokens | 2.5 cr/1K tokens | Highest capability |
Pass tier to select a curated model (default). Custom model and provider require BYOK (openrouterApiKey) — without it, the API returns 400.
Constraints
Keywords follow RFC 2119: MUST = non-negotiable, SHOULD = strong default (override with stated reasoning), MAY = agent discretion.
- MUST include at least one tag on every
memorize()call (e.g.tags: ['source:hubspot', 'type:interaction', 'team:sales']) -- because tags enable filtering, attribution, and workspace scoping; untagged memories are unsearchable by category. - SHOULD include a timestamp in the
contentor use thetimestampparameter -- because temporal ordering lets recall distinguish recent facts from stale ones. - MUST NOT pre-process content with an LLM before calling
memorize()withenhanced: true-- because double-processing wastes tokens and the extraction pipeline is optimized for raw input. - MUST NOT manually deduplicate before memorizing -- because the platform deduplicates at cosine 0.92 similarity and runs background consolidation; client-side dedup adds complexity with no benefit.
- SHOULD memorize generated outputs (emails, notifications, reports) after delivery -- because the feedback loop lets future recalls see what was already sent, preventing repetition.
- SHOULD use
client.collections.create/update/delete()or the web app for schema changes -- because collections define the extraction schema and ad-hoc creation risks inconsistency. - MUST call
client.me()before batch operations to read plan rate limits -- because exceeding limits causes 429 errors and partial syncs with no automatic resume. - SHOULD prepend extraction hints for identity/demographic fields (name, company, title, location) when those fields may be empty for the record -- because the property selector uses embedding similarity, and generic identity fields score low against specific content; hints ensure they are selected alongside the content-relevant properties without limiting the selector. See
reference/memorize.md→ "Extraction Hints" for the full pattern.
Full guide: Read
reference/memorize.mdfor complete method signatures, data mapping patterns, all source-specific recipes (CRM, database, webhook, CSV), batch strategies, error handling, and the feedback loop.
CRM / Database Sync
For production-grade data sync from CRMs and databases (Salesforce, HubSpot, Postgres), this skill includes source-specific connector templates and deployment configs:
- Source templates:
templates/salesforce.md,templates/hubspot.md,templates/postgres.md— fetch patterns, auth setup, field mapping for each source - Deployment:
deploy/Dockerfile,deploy/render.yaml,deploy/github-action.yml— scheduled sync on Render, GitHub Actions, or any container platform - Advanced patterns:
reference/sync-advanced-patterns.md— incremental sync with state tracking, multi-source architecture, batch export with pagination, complete end-to-end example
The integration pattern: initialize project → client.me() for auth + limits → fetch rows from source → client.collections.list() for collection IDs → build property mapping → memorizeBatch() in chunks with 429 retry → verify with search() or smartDigest(). See recipes/data-sync.ts for a runnable example.
Action: RECALL
Retrieve data from Personize memory. The right method depends on what kind of answer you need.
Which Method to Use
| Need | Method | Returns |
|---|---|---|
| "What do we know about X topic?" | memory.smartRecall() |
Semantic search results with optional reflection/answers (recommended) |
| "Quick deterministic lookup, no AI" | memory.recall() |
Direct DynamoDB lookup (type required, no vector search, no reflection) |
| "Give me everything about this person/company" | memory.smartDigest() |
Compiled markdown context — all properties + memories for one entity |
| "List all contacts matching criteria X" | memory.search() |
Filtered records with property values |
| "What are our guidelines for X?" | ai.smartGuidelines() |
Governance variables matching a topic |
smartRecall()vsrecall(): UsesmartRecall()for most use cases — it supports reflection, answer generation,fast_mode, and inferstypefrom email/website_url. Userecall()only for simple direct lookups —typeis required (e.g.type: 'Contact').
Identifier behavior — how
websiteUrl,recordId,type-only, and no identifier affect each endpoint (error vs empty vs org-wide search) → readreference/identifier-scenarios.md.
When to Use What
Need specific facts about a topic? → smartRecall()
Need full context about ONE entity? → smartDigest()
Need to filter/segment a list of records? → search()
Need organizational rules/guidelines? → smartGuidelines()
Building a generation prompt? → smartGuidelines() + smartDigest() + smartRecall()
(governance + entity + task-specific facts)
Quick Example
// Semantic search — find specific facts (recommended)
const results = await client.memory.smartRecall({
query: 'what pain points did this contact mention?',
email: 'sarah.chen@initech.com',
type: 'Contact',
limit: 10,
minScore: 0.4,
include_property_values: true,
});
// Fast recall — skip reflection, ~500ms response
const fast = await client.memory.smartRecall({
query: 'what do we know about this contact?',
email: 'sarah.chen@initech.com',
type: 'Contact',
fast_mode: true,
});
// Entity digest — compiled context for one person
const digest = await client.memory.smartDigest({
email: 'sarah.chen@initech.com',
type: 'Contact',
token_budget: 2000,
include_properties: true,
include_memories: true,
});
// digest.data.compiledContext → ready-to-inject markdown
// Filtered export — find all enterprise contacts
const exported = await client.memory.search({
type: 'Contact',
returnRecords: true,
pageSize: 50,
groups: [{
conditions: [
{ property: 'plan_tier', operator: 'EQ', value: 'enterprise' },
{ property: 'email', operator: 'IS_SET' },
],
}],
});
The Three-Layer Agent Operating Model
Memory is one of three layers every agent should assemble before acting: Guidelines (organizational rules via smartGuidelines()), Memory (entity knowledge via smartDigest()/recall()), and Workspace (coordination state via workspace-tagged recall()/memorize()). All three together: the agent acts within governance, with full context, in coordination with others.
Full architecture guide: See the
collaborationskill'sreference/architecture.mdfor the complete three-layer model, composition patterns, and adoption path.
Cross-Entity Context
Memory gives you everything about ONE entity. But agents often need context from related entities — the company a contact works at, other contacts at the same account, related deals or projects.
Pattern: Multi-entity context assembly
// When working on a contact, also pull their company context
const [contactDigest, companyDigest] = await Promise.all([
client.memory.smartDigest({ email: 'sarah@acme.com', type: 'Contact', token_budget: 1500 }),
client.memory.smartDigest({ website_url: 'https://acme.com', type: 'Company', token_budget: 1000 }),
]);
// Now you know Sarah AND you know Acme — funding stage, tech stack, team size, etc.
When to pull cross-entity context:
- Working on a contact → also pull their company
- Working on a deal → also pull the contact AND the company
- Generating account-level content → pull all contacts at that company
- Detecting patterns → export across entity types and cross-reference
The Context Assembly Pattern
Most generation pipelines combine multiple recall methods:
async function assembleContext(email: string, task: string): Promise<string> {
const sections: string[] = [];
// 1. Governance — rules and guidelines
// Use mode: 'fast' for real-time agents (~200ms), 'full' for deep analysis (~3s)
const governance = await client.ai.smartGuidelines({
message: `${task} — guidelines, tone, constraints`,
mode: 'fast', // embedding-only routing, no LLM overhead
});
if (governance.data?.compiledContext) {
sections.push('## Guidelines\n' + governance.data.compiledContext);
}
// 2. Entity context — everything about this person
const digest = await client.memory.smartDigest({
email,
type: 'Contact',
token_budget: 2000,
include_properties: true,
include_memories: true,
});
if (digest.data?.compiledContext) {
sections.push('## Recipient Context\n' + digest.data.compiledContext);
}
// 3. Task-specific facts — semantic search
const recalled = await client.memory.smartRecall({
query: task,
email,
type: 'Contact',
fast_mode: true,
limit: 10,
minScore: 0.3,
});
if (recalled.data?.results && Array.isArray(recalled.data.results) && recalled.data.results.length > 0) {
sections.push('## Relevant Facts\n' + recalled.data.results.map((m: any) =>
`- ${m.text || m.content || JSON.stringify(m)}`
).join('\n'));
}
return sections.join('\n\n---\n\n');
}
Recall Pricing
| Method | Mode | Cost |
|---|---|---|
recall() |
fast_mode: false (default, with reflection) |
0.2 credits/call |
recall() |
fast_mode: true (no reflection, ~500ms) |
0.1 credits/call |
smartDigest() |
— | Free (reads cached data) |
1 credit = $0.01. Use fast_mode: true in loops and batch pipelines to minimize cost.
Constraints
Keywords follow RFC 2119: MUST = non-negotiable, SHOULD = strong default (override with stated reasoning), MAY = agent discretion.
- MUST set an explicit
token_budgeton everysmartDigest()call -- because the default (1000) may truncate critical context for deep personalization or waste tokens for simple lookups. - SHOULD set
minScoreonsmartRecall()(0.3 for broad context, 0.5+ for precision) -- because omitting it returns low-relevance noise that dilutes the context window. - SHOULD use
fast_mode: truefor context injection, real-time UIs, and batch processing -- because it cuts recall latency from ~10-20s to ~500ms; override for exploratory queries where reflection adds value. - SHOULD assemble context from all three layers (
smartGuidelines+smartDigest+smartRecall) before generating -- because single-source context produces governance-blind, entity-ignorant, or task-irrelevant output. Usemode: 'fast'for real-time agent flows (~200ms),mode: 'deep'for first-call or complex planning tasks (~3s). (Note:'full'was renamed to'deep'in SDK types and API.) - MAY set
include_property_values: trueonsmartRecall()-- because it returns structured properties alongside semantic results, useful when the caller needs both. - MUST paginate
export()calls usingpageandpageSize-- because unbounded exports can time out or exceed memory limits on large datasets. Default pageSize is 50. - MAY cache
smartDigest()results within a single pipeline run when the same entity is referenced multiple times -- because redundant API calls waste tokens and add latency.
Full guide: Read
reference/recall.mdfor complete method signatures, query writing strategies, token budget tuning, scoring thresholds, all context assembly patterns, export filtering, and performance optimization.
SDK Method Reference
import { Personize } from '@personize/sdk';
const client = new Personize({ secretKey: process.env.PERSONIZE_SECRET_KEY! });
Memorize Methods
| Method | Endpoint | Purpose |
|---|---|---|
memory.memorize(opts) |
POST /api/v1/memorize |
Store single item with AI extraction |
memory.memorizeBatch(opts) |
POST /api/v1/batch-memorize |
Batch sync with per-property extractMemories control |
Recall Methods
| Method | Endpoint | Purpose |
|---|---|---|
memory.smartRecall(opts) |
POST /api/v1/smart-recall |
Semantic search with reflection + answer gen (recommended) |
memory.recall(opts) |
POST /api/v1/recall |
Direct DynamoDB lookup — properties + freeform memories (type required, no AI) |
memory.smartDigest(opts) |
POST /api/v1/smart-memory-digest |
Compiled entity context (properties + memories) |
memory.search(opts) |
POST /api/v1/search |
Filter and export records |
ai.smartGuidelines(opts) |
POST /api/v1/ai/smart-guidelines |
Fetch governance variables by topic |
Key Type Signatures
// memorize() — single item
interface MemorizeProOptions {
content: string; // The text to memorize
speaker?: string; // Who said/wrote it
timestamp?: string; // When it happened
email?: string; // Match to contact by email
website_url?: string; // Match to company by website
record_id?: string; // Match to record by ID
enhanced?: boolean; // Enable AI extraction (default: false)
tags?: string[]; // Categorization tags
max_properties?: number; // Max properties to extract
schema?: Record<string, unknown>; // Extraction schema hint
actionId?: string; // Target collection ID
}
// memorizeBatch() — batch sync
interface BatchMemorizeOptions {
source: string; // Source system label ('Hubspot', 'Salesforce')
mapping: {
entityType: string; // 'contact', 'company'
email?: string; // Source field name for email
website?: string; // Source field name for website
runName?: string; // Tracking label
properties: Record<string, {
sourceField: string; // Source field name in row data
collectionId: string; // Target collection ID
collectionName: string; // Target collection name
extractMemories?: boolean; // AI extraction for this property
}>;
};
rows: Record<string, unknown>[]; // Source data rows
dryRun?: boolean; // Validate without writing
chunkSize?: number; // Rows per chunk (default: 1)
}
// smartRecall() — semantic search (recommended)
interface SmartRecallOptions {
query: string; // Natural language query
limit?: number; // Max results (default: 10)
minScore?: number; // Minimum relevance score (0-1)
email?: string; // Scope to one contact
website_url?: string; // Scope to one company
record_id?: string; // Scope to one record
type?: string; // Entity type filter (optional — inferred from email/website_url)
include_property_values?: boolean; // Include structured properties
enable_reflection?: boolean; // AI reflects on results
generate_answer?: boolean; // AI generates a direct answer
fast_mode?: boolean; // Skip reflection + answer gen, ~500ms (default: false)
min_score?: number; // Server-side score filter (in fast_mode, defaults to 0.3)
}
// recall() — direct lookup (simpler, type required)
interface RecallOptions {
query: string; // Natural language query
type: string; // Entity type — REQUIRED (e.g. 'Contact', 'Company')
record_id?: string; // Scope to one record
email?: string; // Scope to one contact
website_url?: string; // Scope to one company
filters?: Record<string, unknown>; // Additional filters
}
// smartDigest() — entity context
interface SmartDigestOptions {
email?: string; // Contact email
website_url?: string; // Company website
record_id?: string; // Record ID
type?: string; // Entity type ('Contact', 'Company')
token_budget?: number; // Max tokens for output (default: 1000)
max_memories?: number; // Max memories to include (default: 20)
include_properties?: boolean; // Include structured properties (default: true)
include_memories?: boolean; // Include free-form memories (default: true)
}
The Data Model
┌─────────────────────────────────────────────────────────────┐
│ PERSONIZE MEMORY │
│ │
│ ┌───────────────────┐ ┌──────────────────────────────┐ │
│ │ STRUCTURED DATA │ │ SEMANTIC MEMORIES │ │
│ │ (DynamoDB) │ │ (LanceDB + Vectors) │ │
│ │ │ │ │ │
│ │ Records: │ │ AI-extracted facts from: │ │
│ │ ├─ email: "..." │ │ ├─ Call notes │ │
│ │ ├─ plan: "pro" │ │ ├─ Support tickets │ │
│ │ ├─ title: "VP" │ │ ├─ Email threads │ │
│ │ └─ login_count:5 │ │ ├─ Meeting transcripts │ │
│ │ │ │ └─ Generated outputs │ │
│ └───────────────────┘ └──────────────────────────────┘ │
│ │ │ │
│ ▼ ▼ │
│ search() filters smartRecall() searches │
│ memorize() writes memorizeBatch() writes │
│ smartDigest() reads both ──────────┘ │
└─────────────────────────────────────────────────────────────┘
- Structured data = exact key-value pairs. Queryable by field, filterable, paginated.
- Semantic memories = AI-extracted facts with vector embeddings. Searchable by meaning.
- smartDigest combines both into a single, token-budgeted markdown block.
Available Resources
| Resource | Contents |
|---|---|
reference/memorize.md |
Full memorize guide: method signatures, data mapping, extractMemories decision tree, source recipes, batch strategies, error handling, feedback loop |
reference/recall.md |
Full recall guide: method signatures, query strategies, token budgets, scoring, context assembly, export filtering, performance tips |
reference/identifier-scenarios.md |
How each endpoint (memorize, recall, smartRecall, smartDigest) behaves with email, websiteUrl, recordId, type-only, or no identifier — scenarios A–G with error vs empty vs success table |
recipes/data-sync.ts |
Batch sync from CRM/database with validation and error handling |
recipes/context-assembly.ts |
Complete context assembly pattern combining all recall methods |
Signal Memorization Patterns
@personize/signal uses entity memory for its feedback loop and deferred notification pipeline. Understanding these patterns helps when debugging Signal behavior or building custom integrations.
Tag Conventions
| Tag | Written by | Purpose |
|---|---|---|
signal:sent |
Engine (step 8) | Tracks delivered notifications — recalled during context assembly to prevent repetition |
signal:deferred |
Engine (step 5) | Marks notifications scored 40-60 for later digest compilation |
signal:pending-digest |
Engine (step 5) | Paired with signal:deferred — digest builder queries these |
signal:digest |
DigestBuilder | Marks compiled digest notifications |
workspace:updates |
Engine (step 7) | Workspace entries created on SEND |
workspace:tasks |
Engine (step 7) | Workspace entries created on DEFER |
Feedback Loop
After every SEND decision, Signal memorizes what was sent:
await client.memory.memorize({
content: `[SIGNAL] Sent "${subject}" via ${channel} (score: ${score}). ${reasoning}`,
email,
enhanced: true,
tags: ['signal:sent', `signal:channel:${channel}`, `signal:type:${eventType}`],
});
On the next evaluation for the same entity, the engine recalls recent signal:sent memories (step 3, 4th parallel call). The AI sees what was recently sent and can SKIP to avoid repetition — even if the pre-check dedup window has expired.
Deferred → Digest Pipeline
- Defer (score 40-60):
memorize()with tags['signal:deferred', 'signal:pending-digest', eventType] - Digest build:
smartRecall({ query: 'deferred notifications', tags: ['signal:deferred'] })retrieves pending items - Compile:
prompt()generates a personalized digest from all deferred items + entity context - Deliver: Channel sends the compiled digest
- Mark processed:
memorize()with tagsignal:digest— future digest builds skip already-compiled items
Querying Signal History
// What notifications has Signal sent to this contact?
const sent = await client.memory.smartRecall({
query: 'notifications sent by signal',
email: 'jane@acme.com',
type: 'Contact',
fast_mode: true,
limit: 10,
});
// What's pending in the digest queue?
const pending = await client.memory.smartRecall({
query: 'deferred notifications pending digest',
email: 'jane@acme.com',
type: 'Contact',
fast_mode: true,
limit: 20,
});