atypica-research

SKILL.md

atypica Research

Access atypica.ai's multi-agent research framework for understanding consumer emotions, market perceptions, and decision preferences.

Prerequisites

IMPORTANT: This skill provides two ways to access atypica.ai research capabilities:

Option 1: MCP Server (Recommended for AI assistants)

If tools starting with atypica_ are already available, the MCP server is configured. Otherwise, guide the user to configure it.

Configuration parameters:

Example: Claude Desktop - Edit config file at:

  • macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
  • Windows: %APPDATA%\Claude\claude_desktop_config.json
{
  "mcpServers": {
    "atypica-research": {
      "transport": "http",
      "url": "https://atypica.ai/mcp/study",
      "headers": {
        "Authorization": "Bearer atypica_xxx"
      }
    }
  }
}

Restart Claude Desktop to load. For other MCP clients, configuration syntax may differ.

Option 2: Direct Bash Script (Works anywhere)

If MCP server is not available or for simpler use cases, use the bundled bash script:

scripts/mcp-call.sh <tool_name> <json_args> [options]

Setup:

export ATYPICA_TOKEN="atypica_xxx"

Examples:

# Create research session
scripts/mcp-call.sh atypica_study_create '{"content":"Research coffee preferences"}'

# Get messages with tail parameter (3-5 parts, increase if more context needed)
scripts/mcp-call.sh atypica_study_get_messages '{"userChatToken":"abc123","tail":5}'

Options:

  • -t, --token - API token (overrides ATYPICA_TOKEN)
  • -o, --output - Output format: text|json|structured|auto
  • -f, --file - Write output to file instead of stdout
  • -v, --verbose - Enable verbose output
  • -h, --help - Show help message

See scripts/mcp-call.sh for full documentation.

Quick Start

Once the MCP server is installed:

// 1. Create research session
const result = await callTool("atypica_study_create", {
  content: "Research young people's coffee preferences"
});
const userChatToken = result.structuredContent.token;

// 2. Send message (triggers AI - may take 10-120s)
await callTool("atypica_study_send_message", {
  userChatToken,
  message: {
    role: "user",
    lastPart: { type: "text", text: "Start research" }
  }
});

// 3. Poll for research progress
let result;
let pollInterval = 30000; // 30 seconds before plan confirmation
let tailSize = 3; // Start with 3-5 parts, increase if needed
do {
  await wait(pollInterval);
  result = await callTool("atypica_study_get_messages", {
    userChatToken,
    tail: tailSize  // Get last few parts for efficiency
  });

  // After plan confirmation, use longer interval
  const hasPlanConfirmed = result.structuredContent.messages.some(m =>
    m.parts.some(p => p.type === "tool-makeStudyPlan" && p.state === "output-available")
  );
  if (hasPlanConfirmed) {
    pollInterval = 300000; // 5 minutes after plan confirmation
  }

  // If needed more context, increase tail size (up to 10-15)
  // or remove tail parameter to get all messages
} while (result.structuredContent.isRunning);

// 4. Handle pending interactions or get final report
const { messages } = result.structuredContent;
const reportTool = messages
  .flatMap(m => m.parts)
  .find(p => p.type === "tool-generateReport" && p.output);

if (reportTool?.output?.reportToken) {
  const report = await callTool("atypica_study_get_report", {
    token: reportTool.output.reportToken
  });
  // Access report details
  console.log(report.structuredContent.title);
  console.log(report.structuredContent.shareUrl);  // Public share URL
  console.log(report.structuredContent.content);   // HTML content
} else {
  // If stopped without report, you can continue:
  await callTool("atypica_study_send_message", {
    userChatToken,
    message: {
      role: "user",
      lastPart: { type: "text", text: "Please continue the research" }
    }
  });
  // Then poll again...
}

Core Workflow

  1. Create research session with initial query
  2. Send messages to drive research forward (AI executes synchronously)
  3. Poll for pending interactions that require user input
  4. Handle interactions by submitting tool results
  5. Monitor progress and retrieve artifacts (reports/podcasts)

Available Tools

Session Management

atypica_study_create - Create research session

  • Input: { content: string }
  • Returns:
    {
      token: string;          // Study session token for subsequent operations
      studyId: number;        // Internal study ID
      status: "created";      // Always "created" on success
    }
    

atypica_study_send_message - Send message and execute AI (10-120s)

  • Two input types:
    • User text: { userChatToken, message: { role: "user", lastPart: { type: "text", text } } }
    • Tool result: See "User Interactions" section
  • Returns:
    {
      messageId: string;           // Message identifier
      role: "user" | "assistant";  // Message role
      status: "completed" | "saved_no_ai" | "ai_failed";
      attachmentCount?: number;    // Number of attachments (if any)
      error?: string;              // Error message (if status is "ai_failed")
      reason?: string;             // Reason (if status is "saved_no_ai")
    }
    

atypica_study_get_messages - Retrieve conversation history and execution status

  • Input: { userChatToken: string, tail?: number }
  • Returns:
    {
      isRunning: boolean;  // true = AI executing, false = can interact
      messages: Array<{
        messageId: string;
        role: "user" | "assistant";
        parts: Array<MessagePart>;  // Text, tool calls, tool results
        createdAt: string;           // ISO timestamp
      }>;
    }
    
  • Critical:
    • isRunning: true → AI is executing, wait and poll again later
    • isRunning: false → Can interact, check for pending tool calls in parts
    • tail (optional): Limit to last N parts across all messages (3-5 recommended)

atypica_study_list - List historical research sessions

  • Input:
    {
      kind?: "testing" | "insights" | "creation" | "planning" | "misc" | "productRnD" | "fastInsight";
      page?: number;      // Default: 1
      pageSize?: number;  // Default: 20, max: 100
    }
    
  • Returns:
    {
      data: Array<{
        studyId: number;
        token: string;           // Study session token
        title: string;           // Auto-generated title
        kind: string | null;     // Research type
        topic: string;           // Research topic
        hasReport: boolean;      // Has generated report
        hasPodcast: boolean;     // Has generated podcast
        replayUrl: string;       // URL to replay the study: https://atypica.ai/study/{token}/share?replay=1
        createdAt: string;       // ISO timestamp
        updatedAt: string;       // ISO timestamp
      }>;
      pagination: {
        page: number;
        pageSize: number;
        totalCount: number;
        totalPages: number;
      };
    }
    

Artifacts

atypica_study_get_report - Get research report

  • Input: { token: string }
  • Returns:
    {
      token: string;          // Report token
      instruction: string;    // Generation instruction
      title: string;          // Report title
      description: string;    // Report description
      content: string;        // HTML content (one-page format)
      coverUrl?: string;      // Signed CDN URL for cover image
      shareUrl: string;       // Public share URL: https://atypica.ai/artifacts/report/{token}/share
      generatedAt: string;    // ISO timestamp when generated
      createdAt: string;      // ISO timestamp when created
      updatedAt: string;      // ISO timestamp when last updated
    }
    

atypica_study_get_podcast - Get podcast content

  • Input: { token: string }
  • Returns:
    {
      token: string;          // Podcast token
      instruction: string;    // Generation instruction
      script: string;         // Full podcast script/transcript
      audioUrl: string;       // Signed CDN URL for audio file
      coverUrl?: string;      // Signed CDN URL for cover image
      metadata: {             // Podcast metadata
        title: string;
        duration?: number;    // Duration in seconds
        coverObjectUrl?: string;
        // ... other metadata fields
      };
      shareUrl: string;       // Public share URL: https://atypica.ai/artifacts/podcast/{token}/share
      generatedAt: string;    // ISO timestamp when generated
      createdAt: string;      // ISO timestamp when created
      updatedAt: string;      // ISO timestamp when last updated
    }
    

Personas

atypica_persona_search - Semantic search for AI personas

  • Input:
    {
      query?: string;    // Search query (uses embedding similarity if provided)
      tier?: number;     // Filter by tier (0-3)
      limit?: number;    // Max results (default: 10, max: 50)
    }
    
  • Uses embedding similarity for semantic matching
  • Returns:
    {
      data: Array<{
        personaId: number;
        token: string;       // Persona token
        name: string;        // Persona name
        source: string;      // Persona source/origin
        tier: number;        // Access tier (0-3)
        tags: string[];      // Associated tags
        createdAt: string;   // ISO timestamp
      }>;
    }
    

atypica_persona_get - Get persona details

  • Input: { personaId: number }
  • Returns:
    {
      personaId: number;
      token: string;       // Persona token
      name: string;        // Persona name
      source: string;      // Persona source/origin
      prompt: string;      // Full persona prompt (system prompt for AI)
      tier: number;        // Access tier (0-3)
      tags: string[];      // Associated tags
      locale: string;      // Persona language locale
      createdAt: string;   // ISO timestamp
      updatedAt: string;   // ISO timestamp
    }
    

Understanding Research State from Messages

All research state is in the messages - you don't need a separate status API. After calling getMessages, follow this pattern:

1. Check if AI is executing

const { isRunning, messages } = result.structuredContent;

if (isRunning) {
  // AI is working in background, cannot interact now
  // Polling strategy:
  // - Before plan confirmation: 30 seconds
  // - After plan confirmation: 5 minutes
  return "Research is running, please wait...";
}

2. Check for pending interactions

Scan the last assistant message for tool calls needing user input:

const lastMsg = messages[messages.length - 1];
if (lastMsg.role === "assistant") {
  for (const part of lastMsg.parts) {
    if (part.type.startsWith("tool-") && part.state === "input-available") {
      // Handle this pending tool call (see User Interactions section)
    }
  }
}

3. Understand research progress

Look at recent tool calls to see what's happening:

Tool Call Meaning
makeStudyPlan In Plan Mode, AI is clarifying intent
interviewChat, discussionChat Conducting interviews/discussions
webSearch, webFetch Gathering information
reasoningThinking Deep analysis in progress
generateReport Generating final report
generatePodcast Generating podcast

4. Check if research is complete

Look for generateReport or generatePodcast tool call with output:

const reportTool = messages
  .flatMap(m => m.parts)
  .find(p => p.type === "tool-generateReport" && p.state === "output-available");

if (reportTool?.output?.reportToken) {
  // Research complete! Get the report
  const report = await callTool("atypica_study_get_report", {
    token: reportTool.output.reportToken
  });
  console.log(report.structuredContent.title);
  console.log(report.structuredContent.shareUrl);  // https://atypica.ai/artifacts/report/{token}/share
}

// Similarly for podcasts
const podcastTool = messages
  .flatMap(m => m.parts)
  .find(p => p.type === "tool-generatePodcast" && p.state === "output-available");

if (podcastTool?.output?.podcastToken) {
  const podcast = await callTool("atypica_study_get_podcast", {
    token: podcastTool.output.podcastToken
  });
  console.log(podcast.structuredContent.audioUrl);
  console.log(podcast.structuredContent.shareUrl);  // https://atypica.ai/artifacts/podcast/{token}/share
}

5. Continue a stopped research

When to use: If research stops (isRunning: false) but no report/podcast was generated, you can continue it.

How to continue:

// Check if research stopped without completing
const { isRunning, messages } = result.structuredContent;
const hasReport = messages.some(m =>
  m.parts.some(p =>
    (p.type === "tool-generateReport" || p.type === "tool-generatePodcast") &&
    p.state === "output-available"
  )
);

if (!isRunning && !hasReport) {
  // Research stopped but not complete - you can continue it
  await callTool("atypica_study_send_message", {
    userChatToken,
    message: {
      role: "user",
      lastPart: {
        type: "text",
        text: "Please continue the research"  // or "[CONTINUE ASSISTANT STEPS]"
      }
    }
  });

  // Then poll again to check progress
}

Important:

  • The AI will resume from where it stopped, not restart from scratch
  • It may retry the last interrupted tool or continue with the next step
  • After sending continue message, poll getMessages again to track progress

User Interactions

Two tools require user interaction. Check getMessages response for pending calls with state === "input-available".

requestInteraction - User Choice/Input

Detect:

{
  "type": "tool-requestInteraction",
  "state": "input-available",
  "toolCallId": "call_abc",
  "input": {
    "question": "Which age group to focus on?",
    "options": ["18-22", "23-28", "29-35"],
    "maxSelect": 1  // 1=single, 2+=multi with limit, undefined=unlimited
  }
}

Submit via sendMessage:

{
  "userChatToken": "...",
  "message": {
    "id": "msg_2",  // Original message ID
    "role": "assistant",
    "lastPart": {
      "type": "tool-requestInteraction",
      "toolCallId": "call_abc",
      "state": "output-available",
      "input": { /* copy from above */ },
      "output": {
        "answer": "23-28",  // string for single, string[] for multi
        "plainText": "User selected: 23-28"
      }
    }
  }
}

makeStudyPlan - Plan Confirmation

Detect:

{
  "type": "tool-makeStudyPlan",
  "state": "input-available",
  "toolCallId": "call_xyz",
  "input": {
    "locale": "zh-CN",
    "kind": "insights",
    "role": "Market Researcher",
    "topic": "Coffee preferences study",
    "planContent": "# Research Plan\n\n## Goals\n...\n## Methods\n..."
  }
}

Submit via sendMessage:

{
  "userChatToken": "...",
  "message": {
    "id": "msg_2",
    "role": "assistant",
    "lastPart": {
      "type": "tool-makeStudyPlan",
      "toolCallId": "call_xyz",
      "state": "output-available",
      "input": { /* copy from above */ },
      "output": {
        "confirmed": true,  // boolean: true to confirm, false to cancel
        "plainText": "User confirmed research plan"
      }
    }
  }
}

Best Practices

  1. Async execution: Call sendMessage asynchronously (10-120s duration)
  2. Poll for interactions: After each sendMessage, check getMessages for pending tool calls
  3. Handle timeouts: If sendMessage times out, use getStatus to check progress
  4. Error handling: Check status field: "completed" | "saved_no_ai" | "ai_failed"
  5. Semantic search: Use natural language queries in persona_search (embedding-based)

Research Types (kind)

  • productRnD - Product research & development
  • fastInsight - Quick insights with podcast generation
  • insights - Consumer insights research
  • testing - Product testing and validation
  • creation - Creative content generation
  • planning - Strategic planning
  • misc - Miscellaneous research

Omit kind in create to enter Plan Mode where AI auto-determines research type.

Performance Expectations

  • Plan Mode: 5-10 seconds
  • Fast Insight: 20-40 seconds
  • Full Study/Product R&D: 30-120 seconds
  • Message/persona retrieval: < 2 seconds

Reference Documentation

See references/api-reference.md for complete API documentation including:

  • Detailed input/output schemas for all tools
  • Error codes and troubleshooting
  • Complete workflow examples
  • Security and limitations
Weekly Installs
10
GitHub Stars
20
First Seen
6 days ago
Installed on
opencode10
gemini-cli10
github-copilot10
codex10
kimi-cli10
amp10