arize-instrumentation
Arize Instrumentation Skill
Use this skill when the user wants to add Arize AX tracing to their application. Follow the two-phase, agent-assisted flow from the Agent-Assisted Tracing Setup and the Arize AX Tracing — Agent Setup Prompt.
Quick start (for the user)
If the user asks you to "set up tracing" or "instrument my app with Arize", you can start with:
Follow the instructions from https://arize.com/docs/PROMPT.md and ask me questions as needed.
Then execute the two phases below.
Core principles
- Prefer inspection over mutation — understand the codebase before changing it.
- Do not change business logic — tracing is purely additive.
- Use auto-instrumentation where available — add manual spans only for custom logic not covered by integrations.
- Follow existing code style and project conventions.
- Keep output concise and production-focused — do not generate extra documentation or summary files.
Phase 1: Analysis (read-only)
Do not write any code or create any files during this phase.
Steps
-
Check dependency manifests to detect stack:
- Python:
pyproject.toml,requirements.txt,setup.py,Pipfile - TypeScript/JavaScript:
package.json - Java:
pom.xml,build.gradle,build.gradle.kts
- Python:
-
Scan import statements in source files to confirm what is actually used.
-
Check for existing tracing/OTel — look for
TracerProvider,register(),opentelemetryimports,ARIZE_*,OTEL_*,OTLP_*env vars, or other observability config (Datadog, Honeycomb, etc.). -
Identify scope — for monorepos or multi-service projects, ask which service(s) to instrument.
What to identify
| Item | Examples |
|---|---|
| Language | Python, TypeScript/JavaScript, Java |
| Package manager | pip/poetry/uv, npm/pnpm/yarn, maven/gradle |
| LLM providers | OpenAI, Anthropic, LiteLLM, Bedrock, etc. |
| Frameworks | LangChain, LangGraph, LlamaIndex, Vercel AI SDK, Mastra, etc. |
| Existing tracing | Any OTel or vendor setup |
| Tool/function use | LLM tool use, function calling, or custom tools the app executes (e.g. in an agent loop) |
Key rule: When a framework is detected alongside an LLM provider, instrument both. Provider/framework instrumentors do not create spans for tool execution — only for LLM API calls. If the app runs tools (e.g. check_loan_eligibility, run_fraud_detection), add manual TOOL spans so each invocation appears with input/output (see Enriching traces below). The framework instrumentor does not trace the underlying LLM calls — you need the provider instrumentor too.
Phase 1 output
Return a concise summary:
- Detected language, package manager, providers, frameworks
- Proposed integration list (from the routing table in the docs)
- Any existing OTel/tracing that needs consideration
- If monorepo: which service(s) you propose to instrument
- If the app uses LLM tool use / function calling: note that you will add manual CHAIN + TOOL spans so each tool call appears in the trace with input/output (avoids sparse traces).
STOP. Present your analysis and wait for user confirmation before proceeding to Phase 2.
Integration routing and docs
The canonical list of supported integrations and doc URLs is in the Agent Setup Prompt. Use it to map detected signals to implementation docs.
- LLM providers: OpenAI, Anthropic, LiteLLM, Google Gen AI, Bedrock, Ollama, Groq, MistralAI, OpenRouter, VertexAI.
- Python frameworks: LangChain, LangGraph, LlamaIndex, CrewAI, DSPy, AutoGen, Semantic Kernel, Pydantic AI, Haystack, Guardrails AI, Hugging Face Smolagents, Instructor, Agno, Google ADK, MCP, Portkey, Together AI, BeeAI, AWS Bedrock Agents.
- TypeScript/JavaScript: LangChain JS, Mastra, Vercel AI SDK, BeeAI JS.
- Java: LangChain4j, Spring AI, Arconia.
- Platforms (UI-based): LangFlow, Flowise, Dify, Prompt flow.
- Fallback: Manual instrumentation, All integrations.
Fetch the matched doc pages from the full routing table in PROMPT.md for exact installation and code snippets. Use llms.txt as a fallback for doc discovery if needed.
Phase 2: Implementation
Proceed only after the user confirms the Phase 1 analysis.
Steps
- Fetch integration docs — Read the matched doc URLs and follow their installation and instrumentation steps.
- Install packages using the detected package manager before writing code:
- Python:
pip install arize-otelplusopeninference-instrumentation-{name}(hyphens in package name; underscores in import, e.g.openinference.instrumentation.llama_index). - TypeScript/JavaScript:
@opentelemetry/sdk-trace-nodeplus the relevant@arizeai/openinference-*package. - Java: OpenTelemetry SDK plus
openinference-instrumentation-*in pom.xml or build.gradle.
- Python:
- Credentials — User needs Arize Space ID and API Key from Space API Keys. Set as
ARIZE_SPACE_IDandARIZE_API_KEY. - Centralized instrumentation — Create a single module (e.g.
instrumentation.py,instrumentation.ts) and initialize tracing before any LLM client is created. - Existing OTel — If there is already a TracerProvider, add Arize as an additional exporter (e.g. BatchSpanProcessor with Arize OTLP). Do not replace existing setup unless the user asks.
Implementation rules
- Use auto-instrumentation first; manual spans only when needed.
- Fail gracefully if env vars are missing (warn, do not crash).
- Import order: register tracer → attach instrumentors → then create LLM clients.
- Project name attribute (required): Arize rejects spans with HTTP 500 if the project name is missing —
service.namealone is not accepted. Set it as a resource attribute on the TracerProvider (recommended — one place, applies to all spans): Python:register(project_name="my-app")handles it automatically (sets"openinference.project.name"on the resource); TypeScript: Arize accepts both"model_id"(shown in the official TS quickstart) and"openinference.project.name"viaSEMRESATTRS_PROJECT_NAMEfrom@arizeai/openinference-semantic-conventions(shown in the manual instrumentation docs) — both work. For routing spans to different projects in Python, useset_routing_context(space_id=..., project_name=...)fromarize.otel. - CLI/script apps — flush before exit:
provider.shutdown()(TS) /provider.force_flush()thenprovider.shutdown()(Python) must be called before the process exits, otherwise async OTLP exports are dropped and no traces appear. - When the app has tool/function execution: add manual CHAIN + TOOL spans (see Enriching traces below) so the trace tree shows each tool call and its result — otherwise traces will look sparse (only LLM API spans, no tool input/output).
Enriching traces: manual spans for tool use and agent loops
Why doesn't the auto-instrumentor do this?
Provider instrumentors (Anthropic, OpenAI, etc.) only wrap the LLM client — the code that sends HTTP requests and receives responses. They see:
- One span per API call: request (messages, system prompt, tools) and response (text, tool_use blocks, etc.).
They cannot see what happens inside your application after the response:
- Tool execution — Your code parses the response, calls
run_tool("check_loan_eligibility", {...}), and gets a result. That runs in your process; the instrumentor has no hook into yourrun_tool()or the actual tool output. The next API call (sending the tool result back) is just anothermessages.createspan — the instrumentor doesn't know that the message content is a tool result or what the tool returned. - Agent/chain boundary — The idea of "one user turn → multiple LLM calls + tool calls" is an application-level concept. The instrumentor only sees separate API calls; it doesn't know they belong to the same logical "run_agent" run.
So TOOL and CHAIN spans have to be added manually (or by a framework instrumentor like LangChain/LangGraph that knows about tools and chains). Once you add them, they appear in the same trace as the LLM spans because they use the same TracerProvider.
To avoid sparse traces where tool inputs/outputs are missing:
- Detect agent/tool patterns: a loop that calls the LLM, then runs one or more tools (by name + arguments), then calls the LLM again with tool results.
- Add manual spans using the same TracerProvider (e.g.
opentelemetry.trace.get_tracer(...)afterregister()):- CHAIN span — Wrap the full agent run (e.g.
run_agent): setopeninference.span.kind="CHAIN",input.value= user message,output.value= final reply. - TOOL span — Wrap each tool invocation: set
openinference.span.kind="TOOL",input.value= JSON of arguments,output.value= JSON of result. Use the tool name as the span name (e.g.check_loan_eligibility).
- CHAIN span — Wrap the full agent run (e.g.
OpenInference attributes (use these so Arize shows spans correctly):
| Attribute | Use |
|---|---|
openinference.span.kind |
"CHAIN" or "TOOL" |
input.value |
string (e.g. user message or JSON of tool args) |
output.value |
string (e.g. final reply or JSON of tool result) |
Python pattern: Get the global tracer (same provider as Arize), then use context managers so tool spans are children of the CHAIN span and appear in the same trace as the LLM spans:
from opentelemetry.trace import get_tracer
tracer = get_tracer("my-app", "1.0.0")
# In your agent entrypoint:
with tracer.start_as_current_span("run_agent") as chain_span:
chain_span.set_attribute("openinference.span.kind", "CHAIN")
chain_span.set_attribute("input.value", user_message)
# ... LLM call ...
for tool_use in tool_uses:
with tracer.start_as_current_span(tool_use["name"]) as tool_span:
tool_span.set_attribute("openinference.span.kind", "TOOL")
tool_span.set_attribute("input.value", json.dumps(tool_use["input"]))
result = run_tool(tool_use["name"], tool_use["input"])
tool_span.set_attribute("output.value", result)
# ... append tool result to messages, call LLM again ...
chain_span.set_attribute("output.value", final_reply)
See Manual instrumentation for more span kinds and attributes.
Verification
After implementation:
- Run the application and trigger at least one LLM call.
- Use the
arize-traceskill to confirm traces arrived. If empty, retry shortly. Verify spans have expectedopeninference.span.kind,input.value/output.value, and parent-child relationships. - If no traces: verify
ARIZE_SPACE_IDandARIZE_API_KEY, ensure tracer is initialized before instrumentors and clients, check connectivity tootlp.arize.com:443; for debug setGRPC_VERBOSITY=debugor passlog_to_console=Truetoregister(). Common gotchas: (a) missing project name resource attribute causes HTTP 500 rejections —service.namealone is not enough; Python: passproject_nametoregister(); TypeScript: set"model_id"orSEMRESATTRS_PROJECT_NAMEon the resource; (b) CLI/script processes exit before OTLP exports flush — callprovider.force_flush()thenprovider.shutdown()before exit. - If the app uses tools: confirm CHAIN and TOOL spans appear with
input.value/output.valueso tool calls and results are visible.
Leveraging the Tracing Assistant (MCP)
For deeper instrumentation guidance inside the IDE, the user can enable:
- Arize AX Tracing Assistant MCP — instrumentation guides, framework examples, and support. In Cursor: Settings → MCP → Add and use:
"arize-tracing-assistant": { "command": "uvx", "args": ["arize-tracing-assistant@latest"] } - Arize AX Docs MCP — searchable docs. In Cursor:
"arize-ax-docs": { "url": "https://arize.com/docs/mcp" }
Then the user can ask things like: "Instrument this app using Arize AX", "Can you use manual instrumentation so I have more control over my traces?", "How can I redact sensitive information from my spans?"
See the full setup at Agent-Assisted Tracing Setup.
Reference links
| Resource | URL |
|---|---|
| Agent-Assisted Tracing Setup | https://arize.com/docs/ax/alyx/tracing-assistant |
| Agent Setup Prompt (full routing + phases) | https://arize.com/docs/PROMPT.md |
| Arize AX Docs | https://arize.com/docs/ax |
| Full integration list | https://arize.com/docs/ax/integrations |
| Doc index (llms.txt) | https://arize.com/docs/llms.txt |